code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( 'files' , [ ['full:README.md', 'dataset_infos.json'], ['empty:README.md', 'dataset_infos.json'], ['dataset_infos.json'], ['full:README.md'], ] , ) def A ( __UpperCamelCase , __UpperCamelCase ) -> List[Any]: A__ = tmp_path_factory.mktemp('dset_infos_dir' ) if "full:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('---\ndataset_info:\n dataset_size: 42\n---' ) if "empty:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / 'dataset_infos.json' , 'w' ) as f: f.write('{"default": {"dataset_size": 42}}' ) A__ = DatasetInfosDict.from_directory(A_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( 'dataset_info' , [ DatasetInfo(), DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ), ] , ) def A ( __UpperCamelCase , __UpperCamelCase ) -> Optional[Any]: A__ = str(A_ ) dataset_info.write_to_directory(A_ ) A__ = DatasetInfo.from_directory(A_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(A_ , 'dataset_info.json' ) ) def A ( ) -> str: A__ = DatasetInfo( description='foo' , citation='bar' , homepage='https://foo.bar' , license='CC0' , features=Features({'a': Value('int32' )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train', 'num_examples': 42}] , download_checksums={} , download_size=1_337 , post_processing_size=442 , dataset_size=1_234 , size_in_bytes=1_337 + 442 + 1_234 , ) A__ = dataset_info._to_yaml_dict() assert sorted(A_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) A__ = yaml.safe_dump(A_ ) A__ = yaml.safe_load(A_ ) assert dataset_info_yaml_dict == reloaded def A ( ) -> int: A__ = DatasetInfo() A__ = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( 'dataset_infos_dict' , [ DatasetInfosDict(), DatasetInfosDict({'default': DatasetInfo()} ), DatasetInfosDict({'my_config_name': DatasetInfo()} ), DatasetInfosDict( { 'default': DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ) } ), DatasetInfosDict( { 'v1': DatasetInfo(dataset_size=42 ), 'v2': DatasetInfo(dataset_size=1_337 ), } ), ] , ) def A ( __UpperCamelCase , __UpperCamelCase ) -> int: A__ = str(A_ ) dataset_infos_dict.write_to_directory(A_ ) A__ = DatasetInfosDict.from_directory(A_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): A__ = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml A__ = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(A_ , 'README.md' ) )
9
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case_ ( A_ : Tuple, A_ : List[str], A_ : Optional[Any], A_ : Dict, A_ : Dict=True, A_ : int="pt" ): '''simple docstring''' _lowerCamelCase : str = {'''add_prefix_space''': True} if isinstance(A_, A_ ) and not line.startswith(''' ''' ) else {} _lowerCamelCase : Union[str, Any] = padding_side return tokenizer( [line], max_length=A_, padding='''max_length''' if pad_to_max_length else None, truncation=A_, return_tensors=A_, add_special_tokens=A_, **A_, ) def snake_case_ ( A_ : Any, A_ : Optional[int], A_ : List[Any]=None, ): '''simple docstring''' _lowerCamelCase : Optional[int] = input_ids.ne(A_ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __snake_case ( _lowercase): def __init__( self : Tuple , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Tuple="train" , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[int]=None , __lowerCAmelCase : Any=None , __lowerCAmelCase : Union[str, Any]="" , ): """simple docstring""" super().__init__() _lowerCamelCase : Optional[int] = Path(__lowerCAmelCase ).joinpath(type_path + '''.source''' ) _lowerCamelCase : List[str] = Path(__lowerCAmelCase ).joinpath(type_path + '''.target''' ) _lowerCamelCase : List[Any] = self.get_char_lens(self.src_file ) _lowerCamelCase : Optional[int] = max_source_length _lowerCamelCase : Optional[Any] = max_target_length assert min(self.src_lens ) > 0, f'''found empty line in {self.src_file}''' _lowerCamelCase : List[Any] = tokenizer _lowerCamelCase : List[Any] = prefix if n_obs is not None: _lowerCamelCase : List[str] = self.src_lens[:n_obs] _lowerCamelCase : int = src_lang _lowerCamelCase : Union[str, Any] = tgt_lang def __len__( self : int ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : Dict , __lowerCAmelCase : Optional[Any] ): """simple docstring""" _lowerCamelCase : str = index + 1 # linecache starts at 1 _lowerCamelCase : Union[str, Any] = self.prefix + linecache.getline(str(self.src_file ) , __lowerCAmelCase ).rstrip('''\n''' ) _lowerCamelCase : Optional[Any] = linecache.getline(str(self.tgt_file ) , __lowerCAmelCase ).rstrip('''\n''' ) assert source_line, f'''empty source line for index {index}''' assert tgt_line, f'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer , __lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right _lowerCamelCase : Optional[int] = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer ) _lowerCamelCase : Union[str, Any] = self.tokenizer.generator if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_source_length , '''right''' ) _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_target_length , '''right''' ) _lowerCamelCase : Optional[Any] = source_inputs['''input_ids'''].squeeze() _lowerCamelCase : Union[str, Any] = target_inputs['''input_ids'''].squeeze() _lowerCamelCase : Any = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : str ): """simple docstring""" return [len(__lowerCAmelCase ) for x in Path(__lowerCAmelCase ).open().readlines()] def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[Any] = torch.stack([x['''input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = torch.stack([x['''attention_mask'''] for x in batch] ) _lowerCamelCase : Union[str, Any] = torch.stack([x['''decoder_input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Tuple = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Union[str, Any] = trim_batch(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : List[str] = trim_batch(__lowerCAmelCase , __lowerCAmelCase , attention_mask=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch lowerCAmelCase__ = getLogger(__name__) def snake_case_ ( A_ : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(A_ ) ) def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : Dict = get_git_info() save_json(A_, os.path.join(A_, '''git_log.json''' ) ) def snake_case_ ( A_ : str, A_ : Union[str, Any], A_ : int=4, **A_ : Optional[int] ): '''simple docstring''' with open(A_, '''w''' ) as f: json.dump(A_, A_, indent=A_, **A_ ) def snake_case_ ( A_ : Any ): '''simple docstring''' with open(A_ ) as f: return json.load(A_ ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = git.Repo(search_parent_directories=A_ ) _lowerCamelCase : str = { '''repo_id''': str(A_ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case_ ( A_ : Callable, A_ : Iterable ): '''simple docstring''' return list(map(A_, A_ ) ) def snake_case_ ( A_ : str, A_ : Tuple ): '''simple docstring''' with open(A_, '''wb''' ) as f: return pickle.dump(A_, A_ ) def snake_case_ ( A_ : List[str] ): '''simple docstring''' def remove_articles(A_ : str ): return re.sub(R'''\b(a|an|the)\b''', ''' ''', A_ ) def white_space_fix(A_ : Any ): return " ".join(text.split() ) def remove_punc(A_ : List[Any] ): _lowerCamelCase : Any = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(A_ : Optional[int] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(A_ ) ) ) ) def snake_case_ ( A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : str = normalize_answer(A_ ).split() _lowerCamelCase : int = normalize_answer(A_ ).split() _lowerCamelCase : str = Counter(A_ ) & Counter(A_ ) _lowerCamelCase : Any = sum(common.values() ) if num_same == 0: return 0 _lowerCamelCase : int = 1.0 * num_same / len(A_ ) _lowerCamelCase : str = 1.0 * num_same / len(A_ ) _lowerCamelCase : List[Any] = (2 * precision * recall) / (precision + recall) return fa def snake_case_ ( A_ : Dict, A_ : str ): '''simple docstring''' return normalize_answer(A_ ) == normalize_answer(A_ ) def snake_case_ ( A_ : List[str], A_ : List[str] ): '''simple docstring''' assert len(A_ ) == len(A_ ) _lowerCamelCase : Optional[Any] = 0 for hypo, pred in zip(A_, A_ ): em += exact_match_score(A_, A_ ) if len(A_ ) > 0: em /= len(A_ ) return {"em": em} def snake_case_ ( A_ : Optional[int] ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def snake_case_ ( A_ : Dict, A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : Dict = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead _lowerCamelCase : Tuple = '''dropout_rate''' for p in extra_params: if getattr(A_, A_, A_ ): if not hasattr(A_, A_ ) and not hasattr(A_, equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(A_ ) ) delattr(A_, A_ ) continue _lowerCamelCase : Union[str, Any] = p if hasattr(A_, A_ ) else equivalent_param[p] setattr(A_, A_, getattr(A_, A_ ) ) delattr(A_, A_ ) return hparams, config
83
0
import argparse import re import requests import torch # git clone https://github.com/salesforce/BLIP.git from models.blip import blip_decoder from models.blip_itm import blip_itm from models.blip_vqa import blip_vqa from PIL import Image from torchvision import transforms from torchvision.transforms.functional import InterpolationMode from transformers import ( BertTokenizer, BlipConfig, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, ) def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> List[Any]: snake_case__ = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/demo.jpg''' snake_case__ = Image.open(requests.get(A_ , stream=A_ ).raw ).convert('''RGB''' ) snake_case__ = transforms.Compose( [ transforms.Resize((image_size, image_size) , interpolation=InterpolationMode.BICUBIC ), transforms.ToTensor(), transforms.Normalize((0.4814_5466, 0.457_8275, 0.4082_1073) , (0.2686_2954, 0.2613_0258, 0.2757_7711) ), ] ) snake_case__ = transform(A_ ).unsqueeze(0 ).to(A_ ) return image def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> str: if "visual_encoder" in key: snake_case__ = re.sub('''visual_encoder*''' , '''vision_model.encoder''' , A_ ) if "blocks" in key: snake_case__ = re.sub(r'''blocks''' , '''layers''' , A_ ) if "attn" in key: snake_case__ = re.sub(r'''attn''' , '''self_attn''' , A_ ) if "norm1" in key: snake_case__ = re.sub(r'''norm1''' , '''layer_norm1''' , A_ ) if "norm2" in key: snake_case__ = re.sub(r'''norm2''' , '''layer_norm2''' , A_ ) if "encoder.norm" in key: snake_case__ = re.sub(r'''encoder.norm''' , '''post_layernorm''' , A_ ) if "encoder.patch_embed.proj" in key: snake_case__ = re.sub(r'''encoder.patch_embed.proj''' , '''embeddings.patch_embedding''' , A_ ) if "encoder.pos_embed" in key: snake_case__ = re.sub(r'''encoder.pos_embed''' , '''embeddings.position_embedding''' , A_ ) if "encoder.cls_token" in key: snake_case__ = re.sub(r'''encoder.cls_token''' , '''embeddings.class_embedding''' , A_ ) if "self_attn" in key: snake_case__ = re.sub(r'''self_attn.proj''' , '''self_attn.projection''' , A_ ) return key @torch.no_grad() def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase=None ) -> Any: if config_path is not None: snake_case__ = BlipConfig.from_pretrained(A_ ) else: snake_case__ = BlipConfig(projection_dim=512 , text_config={} , vision_config={} ) snake_case__ = BlipForConditionalGeneration(A_ ).eval() snake_case__ = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_capfilt_large.pth''' snake_case__ = blip_decoder(pretrained=A_ , image_size=384 , vit='''base''' ) snake_case__ = pt_model.eval() snake_case__ = pt_model.state_dict() for key in modified_state_dict.copy(): snake_case__ = modified_state_dict.pop(A_ ) snake_case__ = rename_key(A_ ) snake_case__ = value hf_model.load_state_dict(A_ ) snake_case__ = 384 snake_case__ = load_demo_image(image_size=A_ , device='''cpu''' ) snake_case__ = BertTokenizer.from_pretrained('''bert-base-uncased''' ) snake_case__ = tokenizer(['''a picture of'''] ).input_ids snake_case__ = hf_model.generate(A_ , A_ ) assert out[0].tolist() == [3_0522, 1037, 3861, 1997, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] snake_case__ = hf_model.generate(A_ ) assert out[0].tolist() == [3_0522, 1037, 2450, 3564, 2006, 1996, 3509, 2007, 2014, 3899, 102] if pytorch_dump_folder_path is not None: hf_model.save_pretrained(A_ ) # model_url = 'https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_vqa.pth' snake_case__ = ( '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_vqa_capfilt_large.pth''' ) snake_case__ = blip_vqa(pretrained=A_ , image_size=A_ , vit='''base''' ) vqa_model.eval() snake_case__ = vqa_model.state_dict() for key in modified_state_dict.copy(): snake_case__ = modified_state_dict.pop(A_ ) snake_case__ = rename_key(A_ ) snake_case__ = value snake_case__ = BlipForQuestionAnswering(A_ ) hf_vqa_model.load_state_dict(A_ ) snake_case__ = ['''How many dogs are in this image?'''] snake_case__ = tokenizer(A_ , return_tensors='''pt''' ).input_ids snake_case__ = hf_vqa_model.generate(A_ , A_ ) print(tokenizer.decode(answer[0] ) ) assert tokenizer.decode(answer[0] ) == "[UNK] 1 [SEP]" if pytorch_dump_folder_path is not None: hf_vqa_model.save_pretrained(pytorch_dump_folder_path + '''_vqa''' ) snake_case__ = '''https://storage.googleapis.com/sfr-vision-language-research/BLIP/models/model_base_retrieval_coco.pth''' snake_case__ = blip_itm(pretrained=A_ , image_size=A_ , vit='''base''' ) itm_model.eval() snake_case__ = itm_model.state_dict() for key in modified_state_dict.copy(): snake_case__ = modified_state_dict.pop(A_ ) snake_case__ = rename_key(A_ ) snake_case__ = value snake_case__ = BlipForImageTextRetrieval(A_ ) snake_case__ = ['''A picture of a woman with a dog sitting in a beach'''] snake_case__ = tokenizer( A_ , return_tensors='''pt''' , padding='''max_length''' , truncation=A_ , max_length=35 , ).input_ids hf_itm_model.load_state_dict(A_ ) hf_itm_model.eval() snake_case__ = hf_itm_model(A_ , A_ , use_itm_head=A_ ) snake_case__ = hf_itm_model(A_ , A_ , use_itm_head=A_ ) assert out[0].item() == 0.2110_6874_9427_7954 assert torch.nn.functional.softmax(out_itm[0] , dim=1 )[:, 1].item() == 0.4_5698_8453_8650_5127 if pytorch_dump_folder_path is not None: hf_itm_model.save_pretrained(pytorch_dump_folder_path + '''_itm''' ) if __name__ == "__main__": lowerCamelCase__ : Dict = argparse.ArgumentParser() parser.add_argument("""--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model.""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") lowerCamelCase__ : Optional[int] = parser.parse_args() convert_blip_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
33
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/config.json''', '''umberto-commoncrawl-cased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json''' ), '''umberto-wikipedia-uncased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json''' ), } class __snake_case ( _lowercase): snake_case__ : Optional[Any] = "camembert" def __init__( self : Optional[Any] , __lowerCAmelCase : Any=3_0_5_2_2 , __lowerCAmelCase : List[str]=7_6_8 , __lowerCAmelCase : List[str]=1_2 , __lowerCAmelCase : Optional[int]=1_2 , __lowerCAmelCase : List[Any]=3_0_7_2 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Union[str, Any]=0.1 , __lowerCAmelCase : Optional[int]=0.1 , __lowerCAmelCase : Optional[int]=5_1_2 , __lowerCAmelCase : str=2 , __lowerCAmelCase : int=0.02 , __lowerCAmelCase : List[Any]=1E-12 , __lowerCAmelCase : Union[str, Any]=1 , __lowerCAmelCase : Optional[Any]=0 , __lowerCAmelCase : List[Any]=2 , __lowerCAmelCase : str="absolute" , __lowerCAmelCase : Any=True , __lowerCAmelCase : Optional[int]=None , **__lowerCAmelCase : Optional[int] , ): """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) _lowerCamelCase : Tuple = vocab_size _lowerCamelCase : str = hidden_size _lowerCamelCase : Union[str, Any] = num_hidden_layers _lowerCamelCase : Any = num_attention_heads _lowerCamelCase : Optional[Any] = hidden_act _lowerCamelCase : List[str] = intermediate_size _lowerCamelCase : Optional[Any] = hidden_dropout_prob _lowerCamelCase : List[Any] = attention_probs_dropout_prob _lowerCamelCase : Optional[Any] = max_position_embeddings _lowerCamelCase : Tuple = type_vocab_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Dict = layer_norm_eps _lowerCamelCase : List[Any] = position_embedding_type _lowerCamelCase : int = use_cache _lowerCamelCase : List[str] = classifier_dropout class __snake_case ( _lowercase): @property def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" if self.task == "multiple-choice": _lowerCamelCase : Optional[Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase : Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
83
0
from typing import List import numpy as np def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ) -> Tuple: SCREAMING_SNAKE_CASE_ : List[Any] = {key: len(A_ ) for key, value in gen_kwargs.items() if isinstance(A_ , A_ )} if len(set(lists_lengths.values() ) ) > 1: raise RuntimeError( ( 'Sharding is ambiguous for this dataset: ' + 'we found several data sources lists of different lengths, and we don\'t know over which list we should parallelize:\n' + '\n'.join(f'\t- key {key} has length {length}' for key, length in lists_lengths.items() ) + '\nTo fix this, check the \'gen_kwargs\' and make sure to use lists only for data sources, ' + 'and use tuples otherwise. In the end there should only be one single list, or several lists with the same length.' ) ) SCREAMING_SNAKE_CASE_ : List[str] = max(lists_lengths.values() , default=0 ) return max(1 , A_ ) def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ : int = [] for group_idx in range(A_ ): SCREAMING_SNAKE_CASE_ : Tuple = num_shards // max_num_jobs + (group_idx < (num_shards % max_num_jobs)) if num_shards_to_add == 0: break SCREAMING_SNAKE_CASE_ : Any = shards_indices_per_group[-1].stop if shards_indices_per_group else 0 SCREAMING_SNAKE_CASE_ : List[str] = range(A_ , start + num_shards_to_add ) shards_indices_per_group.append(A_ ) return shards_indices_per_group def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[Any]: SCREAMING_SNAKE_CASE_ : List[Any] = _number_of_shards_in_gen_kwargs(A_ ) if num_shards == 1: return [dict(A_ )] else: SCREAMING_SNAKE_CASE_ : Union[str, Any] = _distribute_shards(num_shards=A_ , max_num_jobs=A_ ) return [ { key: [value[shard_idx] for shard_idx in shard_indices_per_group[group_idx]] if isinstance(A_ , A_ ) else value for key, value in gen_kwargs.items() } for group_idx in range(len(A_ ) ) ] def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ) -> Optional[Any]: return { key: [value for gen_kwargs in gen_kwargs_list for value in gen_kwargs[key]] if isinstance(gen_kwargs_list[0][key] , A_ ) else gen_kwargs_list[0][key] for key in gen_kwargs_list[0] } def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ : List[str] = {len(A_ ) for value in gen_kwargs.values() if isinstance(A_ , A_ )} SCREAMING_SNAKE_CASE_ : Tuple = {} for size in list_sizes: SCREAMING_SNAKE_CASE_ : str = list(range(A_ ) ) rng.shuffle(indices_per_size[size] ) # Now let's copy the gen_kwargs and shuffle the lists based on their sizes SCREAMING_SNAKE_CASE_ : Optional[int] = dict(A_ ) for key, value in shuffled_kwargs.items(): if isinstance(A_ , A_ ): SCREAMING_SNAKE_CASE_ : int = [value[i] for i in indices_per_size[len(A_ )]] return shuffled_kwargs
345
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase__ = '''<<<<<<< This should probably be modified because it mentions: ''' lowerCAmelCase__ = '''======= >>>>>>> ''' lowerCAmelCase__ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] lowerCAmelCase__ = [ # (pattern, replacement) # Order is important here for some replacements (R'''tfds\.core''', R'''datasets'''), (R'''tf\.io\.gfile\.GFile''', R'''open'''), (R'''tf\.([\w\d]+)''', R'''datasets.Value(\'\1\')'''), (R'''tfds\.features\.Text\(\)''', R'''datasets.Value(\'string\')'''), (R'''tfds\.features\.Text\(''', R'''datasets.Value(\'string\'),'''), (R'''features\s*=\s*tfds.features.FeaturesDict\(''', R'''features=datasets.Features('''), (R'''tfds\.features\.FeaturesDict\(''', R'''dict('''), (R'''The TensorFlow Datasets Authors''', R'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (R'''tfds\.''', R'''datasets.'''), (R'''dl_manager\.manual_dir''', R'''self.config.data_dir'''), (R'''self\.builder_config''', R'''self.config'''), ] def snake_case_ ( A_ : Namespace ): '''simple docstring''' return ConvertCommand(args.tfds_path, args.datasets_directory ) class __snake_case ( _lowercase): @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : ArgumentParser ): """simple docstring""" _lowerCamelCase : List[str] = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=__lowerCAmelCase ) def __init__( self : str , __lowerCAmelCase : str , __lowerCAmelCase : str , *__lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[str] = get_logger('''datasets-cli/converting''' ) _lowerCamelCase : int = tfds_path _lowerCamelCase : Dict = datasets_directory def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" if os.path.isdir(self._tfds_path ): _lowerCamelCase : Union[str, Any] = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): _lowerCamelCase : Dict = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) _lowerCamelCase : int = os.path.abspath(self._datasets_directory ) self._logger.info(f'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) _lowerCamelCase : str = [] _lowerCamelCase : Union[str, Any] = [] _lowerCamelCase : Union[str, Any] = {} if os.path.isdir(self._tfds_path ): _lowerCamelCase : List[str] = os.listdir(__lowerCAmelCase ) else: _lowerCamelCase : Optional[Any] = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f'''Looking at file {f_name}''' ) _lowerCamelCase : Union[str, Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : Optional[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) if not os.path.isfile(__lowerCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(__lowerCAmelCase , encoding='''utf-8''' ) as f: _lowerCamelCase : Tuple = f.readlines() _lowerCamelCase : Optional[int] = [] _lowerCamelCase : Union[str, Any] = False _lowerCamelCase : int = False _lowerCamelCase : Tuple = [] for line in lines: _lowerCamelCase : Optional[int] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: _lowerCamelCase : Union[str, Any] = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here _lowerCamelCase : List[str] = '''''' continue elif "from absl import logging" in out_line: _lowerCamelCase : str = '''from datasets import logging\n''' elif "getLogger" in out_line: _lowerCamelCase : Union[str, Any] = out_line.replace('''getLogger''' , '''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): _lowerCamelCase : Dict = True _lowerCamelCase : Optional[int] = list(filter(lambda __lowerCAmelCase : e in out_line , __lowerCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowerCAmelCase ) + '''\n''' ) out_lines.append(__lowerCAmelCase ) out_lines.append(__lowerCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: _lowerCamelCase : str = re.sub(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: _lowerCamelCase : Dict = re.match(R'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , __lowerCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) _lowerCamelCase : Union[str, Any] = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: _lowerCamelCase : Any = True out_lines.append(__lowerCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset _lowerCamelCase : Union[str, Any] = f_name.replace('''.py''' , '''''' ) _lowerCamelCase : List[str] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) self._logger.info(f'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowerCAmelCase ) if needs_manual_update: with_manual_update.append(__lowerCAmelCase ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.writelines(__lowerCAmelCase ) self._logger.info(f'''Converted in {output_file}''' ) for utils_file in utils_files: try: _lowerCamelCase : Optional[int] = os.path.basename(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = imports_to_builder_map[f_name.replace('''.py''' , '''''' )] self._logger.info(f'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowerCAmelCase , __lowerCAmelCase ) except KeyError: self._logger.error(f'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
83
0
import numpy as np import qiskit def _UpperCAmelCase ( A = 8 , A = None ): '''simple docstring''' UpperCAmelCase__ =np.random.default_rng(seed=A_ ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. UpperCAmelCase__ =6 * key_len # Measurement basis for Alice's qubits. UpperCAmelCase__ =rng.integers(2 , size=A_ ) # The set of states Alice will prepare. UpperCAmelCase__ =rng.integers(2 , size=A_ ) # Measurement basis for Bob's qubits. UpperCAmelCase__ =rng.integers(2 , size=A_ ) # Quantum Circuit to simulate BB84 UpperCAmelCase__ =qiskit.QuantumCircuit(A_ , name="BB84" ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(A_ ): if alice_state[index] == 1: bbaa_circ.x(A_ ) if alice_basis[index] == 1: bbaa_circ.h(A_ ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(A_ ): if bob_basis[index] == 1: bbaa_circ.h(A_ ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. UpperCAmelCase__ =qiskit.Aer.get_backend("aer_simulator" ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. UpperCAmelCase__ =qiskit.execute(A_ , A_ , shots=1 , seed_simulator=A_ ) # Returns the result of measurement. UpperCAmelCase__ =job.result().get_counts(A_ ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. UpperCAmelCase__ =''''''.join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( A_ , A_ , A_ ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. UpperCAmelCase__ =gen_key[:key_len] if len(A_ ) >= key_len else gen_key.ljust(A_ , "0" ) return key if __name__ == "__main__": print(f"""The generated key is : {bbaa(8, seed=0)}""") from doctest import testmod testmod()
625
"""simple docstring""" def snake_case_ ( A_ : list, A_ : list, A_ : int, A_ : int, A_ : int ): '''simple docstring''' if index == number_of_items: return 0 _lowerCamelCase : int = 0 _lowerCamelCase : str = 0 _lowerCamelCase : Dict = knapsack(A_, A_, A_, A_, index + 1 ) if weights[index] <= max_weight: _lowerCamelCase : Tuple = values[index] + knapsack( A_, A_, A_, max_weight - weights[index], index + 1 ) return max(A_, A_ ) if __name__ == "__main__": import doctest doctest.testmod()
83
0
import copy import os import cva import numpy as np from matplotlib import pyplot as plt class A : def __init__( self : Optional[int] ) -> int: """simple docstring""" _lowerCamelCase : Optional[int] ='''''' _lowerCamelCase : str ='''''' _lowerCamelCase : Any =[] _lowerCamelCase : Optional[Any] =0 _lowerCamelCase : Tuple =256 _lowerCamelCase : Optional[int] =0 _lowerCamelCase : str =0 _lowerCamelCase : Any =0 _lowerCamelCase : Dict =0 def lowerCamelCase ( self : int , lowercase_ : List[str] ) -> Optional[int]: """simple docstring""" _lowerCamelCase : List[Any] =cva.imread(__lowerCAmelCase , 0 ) _lowerCamelCase : Tuple =copy.deepcopy(self.img ) _lowerCamelCase : Tuple =plt.hist(self.img.ravel() , 256 , [0, 256] , label='x' ) _lowerCamelCase : Dict =np.sum(__lowerCAmelCase ) for i in range(len(__lowerCAmelCase ) ): _lowerCamelCase : Tuple =x[i] / self.k self.sk += prk _lowerCamelCase : str =(self.L - 1) * self.sk if self.rem != 0: _lowerCamelCase : List[Any] =int(last % last ) _lowerCamelCase : Optional[Any] =int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(__lowerCAmelCase ) _lowerCamelCase : Optional[Any] =int(np.ma.count(self.img ) / self.img[1].size ) _lowerCamelCase : Optional[int] =self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): _lowerCamelCase : int =self.img[j][i] if num != self.last_list[num]: _lowerCamelCase : Optional[int] =self.last_list[num] cva.imwrite('output_data/output.jpg' , self.img ) def lowerCamelCase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" plt.hist(self.img.ravel() , 256 , [0, 256] ) def lowerCamelCase ( self : List[str] ) -> Tuple: """simple docstring""" cva.imshow('Output-Image' , self.img ) cva.imshow('Input-Image' , self.original_image ) cva.waitKey(5000 ) cva.destroyAllWindows() if __name__ == "__main__": lowerCamelCase = os.path.join(os.path.basename(__file__), 'image_data/input.jpg') lowerCamelCase = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
464
"""simple docstring""" from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=_lowercase): snake_case__ : Optional[Any] = ["transformers", "torch", "note_seq"] def __init__( self : Union[str, Any] , *__lowerCAmelCase : Optional[Any] , **__lowerCAmelCase : List[Any] ): """simple docstring""" requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : List[Any] , *__lowerCAmelCase : str , **__lowerCAmelCase : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : int , *__lowerCAmelCase : List[str] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
83
0
'''simple docstring''' import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def _UpperCamelCase ( UpperCamelCase__ ): UpperCAmelCase__ : Any = filter(lambda UpperCamelCase__ : p.requires_grad , model.parameters() ) UpperCAmelCase__ : Union[str, Any] = sum([np.prod(p.size() ) for p in model_parameters] ) return params __A =logging.getLogger(__name__) def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): if metric == "rouge2": UpperCAmelCase__ : Union[str, Any] = '''{val_avg_rouge2:.4f}-{step_count}''' elif metric == "bleu": UpperCAmelCase__ : Dict = '''{val_avg_bleu:.4f}-{step_count}''' elif metric == "em": UpperCAmelCase__ : Optional[int] = '''{val_avg_em:.4f}-{step_count}''' elif metric == "loss": UpperCAmelCase__ : List[str] = '''{val_avg_loss:.4f}-{step_count}''' else: raise NotImplementedError( f'''seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this''' """ function.""" ) UpperCAmelCase__ : str = ModelCheckpoint( dirpath=A_ , filename=A_ , monitor=f'''val_{metric}''' , mode="""max""" , save_top_k=1 , every_n_epochs=1 , ) return checkpoint_callback def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): return EarlyStopping( monitor=f'''val_{metric}''' , mode="""min""" if """loss""" in metric else """max""" , patience=A_ , verbose=A_ , ) class _snake_case ( pl.Callback ): def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : str = {f'''lr_group_{i}''': param['''lr'''] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups)} pl_module.logger.log_metrics(__lowerCAmelCase) @rank_zero_only def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase=True): logger.info(f'''***** {type_path} results at step {trainer.global_step:05d} *****''') UpperCAmelCase__ : Any = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["""log""", """progress_bar""", """preds"""]}) # Log results UpperCAmelCase__ : Optional[Any] = Path(pl_module.hparams.output_dir) if type_path == "test": UpperCAmelCase__ : List[str] = od / '''test_results.txt''' UpperCAmelCase__ : Dict = od / '''test_generations.txt''' else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. UpperCAmelCase__ : Tuple = od / f'''{type_path}_results/{trainer.global_step:05d}.txt''' UpperCAmelCase__ : str = od / f'''{type_path}_generations/{trainer.global_step:05d}.txt''' results_file.parent.mkdir(exist_ok=__lowerCAmelCase) generations_file.parent.mkdir(exist_ok=__lowerCAmelCase) with open(__lowerCAmelCase , """a+""") as writer: for key in sorted(__lowerCAmelCase): if key in ["log", "progress_bar", "preds"]: continue UpperCAmelCase__ : int = metrics[key] if isinstance(__lowerCAmelCase , torch.Tensor): UpperCAmelCase__ : Dict = val.item() UpperCAmelCase__ : Any = f'''{key}: {val:.6f}\n''' writer.write(__lowerCAmelCase) if not save_generations: return if "preds" in metrics: UpperCAmelCase__ : str = '''\n'''.join(metrics["""preds"""]) generations_file.open("""w+""").write(__lowerCAmelCase) @rank_zero_only def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): try: UpperCAmelCase__ : int = pl_module.model.model.num_parameters() except AttributeError: UpperCAmelCase__ : List[str] = pl_module.model.num_parameters() UpperCAmelCase__ : str = count_trainable_parameters(__lowerCAmelCase) # mp stands for million parameters trainer.logger.log_metrics({"""n_params""": npars, """mp""": npars / 1e6, """grad_mp""": n_trainable_pars / 1e6}) @rank_zero_only def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): save_json(pl_module.metrics , pl_module.metrics_save_path) return self._write_logs(__lowerCAmelCase , __lowerCAmelCase , """test""") @rank_zero_only def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase): save_json(pl_module.metrics , pl_module.metrics_save_path) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
407
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __snake_case ( unittest.TestCase): @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : Any = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Any = AutoTokenizer.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Union[str, Any] = tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids _lowerCamelCase : Optional[int] = tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids _lowerCamelCase : List[Any] = shift_tokens_right(__lowerCAmelCase , model.config.pad_token_id , model.config.decoder_start_token_id ) _lowerCamelCase : int = model(__lowerCAmelCase , decoder_input_ids=__lowerCAmelCase ).logits _lowerCamelCase : Optional[Any] = optax.softmax_cross_entropy(__lowerCAmelCase , onehot(__lowerCAmelCase , logits.shape[-1] ) ).mean() _lowerCamelCase : Dict = -(labels.shape[-1] * loss.item()) _lowerCamelCase : Dict = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
83
0
'''simple docstring''' from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class UpperCAmelCase : """simple docstring""" lowerCAmelCase_ = 42 # setable values lowerCAmelCase_ = 42 lowerCAmelCase_ = 42 lowerCAmelCase_ = None @classmethod def UpperCamelCase__ ( cls : List[str] , UpperCamelCase__ : CommonSchedulerState , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : jnp.ndarray ) -> Any: return cls(common=__lowerCAmelCase , init_noise_sigma=__lowerCAmelCase , timesteps=__lowerCAmelCase ) @dataclass class UpperCAmelCase ( _lowercase): """simple docstring""" lowerCAmelCase_ = 42 class UpperCAmelCase ( _lowercase , _lowercase): """simple docstring""" lowerCAmelCase_ = [e.name for e in FlaxKarrasDiffusionSchedulers] lowerCAmelCase_ = 42 @property def UpperCamelCase__ ( self : Tuple ) -> Optional[Any]: return True @register_to_config def __init__( self : List[str] , UpperCamelCase__ : int = 1000 , UpperCamelCase__ : float = 0.0001 , UpperCamelCase__ : float = 0.02 , UpperCamelCase__ : str = "linear" , UpperCamelCase__ : Optional[jnp.ndarray] = None , UpperCamelCase__ : str = "fixed_small" , UpperCamelCase__ : bool = True , UpperCamelCase__ : str = "epsilon" , UpperCamelCase__ : jnp.dtype = jnp.floataa , ) -> List[str]: _UpperCamelCase =dtype def UpperCamelCase__ ( self : List[str] , UpperCamelCase__ : Optional[CommonSchedulerState] = None ) -> Optional[int]: if common is None: _UpperCamelCase =CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution _UpperCamelCase =jnp.array(1.0 , dtype=self.dtype ) _UpperCamelCase =jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=__lowerCAmelCase , init_noise_sigma=__lowerCAmelCase , timesteps=__lowerCAmelCase , ) def UpperCamelCase__ ( self : Optional[int] , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : Optional[int] = None ) -> Tuple: return sample def UpperCamelCase__ ( self : Any , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : int , UpperCamelCase__ : Tuple = () ) -> Union[str, Any]: _UpperCamelCase =self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 _UpperCamelCase =(jnp.arange(0 , __lowerCAmelCase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=__lowerCAmelCase , timesteps=__lowerCAmelCase , ) def UpperCamelCase__ ( self : List[Any] , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Tuple=None , UpperCamelCase__ : int=None ) -> Union[str, Any]: _UpperCamelCase =state.common.alphas_cumprod[t] _UpperCamelCase =jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample _UpperCamelCase =(1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: _UpperCamelCase =self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": _UpperCamelCase =jnp.clip(__lowerCAmelCase , a_min=1E-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": _UpperCamelCase =jnp.log(jnp.clip(__lowerCAmelCase , a_min=1E-20 ) ) elif variance_type == "fixed_large": _UpperCamelCase =state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log _UpperCamelCase =jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": _UpperCamelCase =variance _UpperCamelCase =state.common.betas[t] _UpperCamelCase =(predicted_variance + 1) / 2 _UpperCamelCase =frac * max_log + (1 - frac) * min_log return variance def UpperCamelCase__ ( self : Optional[Any] , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : int , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : Optional[jax.random.KeyArray] = None , UpperCamelCase__ : bool = True , ) -> Any: _UpperCamelCase =timestep if key is None: _UpperCamelCase =jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: _UpperCamelCase =jnp.split(__lowerCAmelCase , sample.shape[1] , axis=1 ) else: _UpperCamelCase =None # 1. compute alphas, betas _UpperCamelCase =state.common.alphas_cumprod[t] _UpperCamelCase =jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) _UpperCamelCase =1 - alpha_prod_t _UpperCamelCase =1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": _UpperCamelCase =(sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": _UpperCamelCase =model_output elif self.config.prediction_type == "v_prediction": _UpperCamelCase =(alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` ''' ''' for the FlaxDDPMScheduler.''' ) # 3. Clip "predicted x_0" if self.config.clip_sample: _UpperCamelCase =jnp.clip(__lowerCAmelCase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _UpperCamelCase =(alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t _UpperCamelCase =state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf _UpperCamelCase =pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): _UpperCamelCase =jax.random.split(__lowerCAmelCase , num=1 ) _UpperCamelCase =jax.random.normal(__lowerCAmelCase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(__lowerCAmelCase , __lowerCAmelCase , predicted_variance=__lowerCAmelCase ) ** 0.5) * noise _UpperCamelCase =jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) _UpperCamelCase =pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=__lowerCAmelCase , state=__lowerCAmelCase ) def UpperCamelCase__ ( self : Union[str, Any] , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : jnp.ndarray , ) -> Union[str, Any]: return add_noise_common(state.common , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase__ ( self : int , UpperCamelCase__ : DDPMSchedulerState , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : jnp.ndarray , UpperCamelCase__ : jnp.ndarray , ) -> Any: return get_velocity_common(state.common , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def __len__( self : List[Any] ) -> Optional[Any]: return self.config.num_train_timesteps
404
"""simple docstring""" import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( '''files''', [ ['''full:README.md''', '''dataset_infos.json'''], ['''empty:README.md''', '''dataset_infos.json'''], ['''dataset_infos.json'''], ['''full:README.md'''], ], ) def snake_case_ ( A_ : Dict, A_ : List[str] ): '''simple docstring''' _lowerCamelCase : int = tmp_path_factory.mktemp('''dset_infos_dir''' ) if "full:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''---\ndataset_info:\n dataset_size: 42\n---''' ) if "empty:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''''' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / '''dataset_infos.json''', '''w''' ) as f: f.write('''{"default": {"dataset_size": 42}}''' ) _lowerCamelCase : str = DatasetInfosDict.from_directory(A_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( '''dataset_info''', [ DatasetInfo(), DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ), ], ) def snake_case_ ( A_ : str, A_ : DatasetInfo ): '''simple docstring''' _lowerCamelCase : Optional[Any] = str(A_ ) dataset_info.write_to_directory(A_ ) _lowerCamelCase : str = DatasetInfo.from_directory(A_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(A_, '''dataset_info.json''' ) ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = DatasetInfo( description='''foo''', citation='''bar''', homepage='''https://foo.bar''', license='''CC0''', features=Features({'''a''': Value('''int32''' )} ), post_processed={}, supervised_keys=(), task_templates=[], builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train''', '''num_examples''': 42}], download_checksums={}, download_size=13_37, post_processing_size=4_42, dataset_size=12_34, size_in_bytes=13_37 + 4_42 + 12_34, ) _lowerCamelCase : Optional[Any] = dataset_info._to_yaml_dict() assert sorted(A_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key], (list, dict, int, str) ) _lowerCamelCase : str = yaml.safe_dump(A_ ) _lowerCamelCase : Tuple = yaml.safe_load(A_ ) assert dataset_info_yaml_dict == reloaded def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : int = DatasetInfo() _lowerCamelCase : Dict = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( '''dataset_infos_dict''', [ DatasetInfosDict(), DatasetInfosDict({'''default''': DatasetInfo()} ), DatasetInfosDict({'''my_config_name''': DatasetInfo()} ), DatasetInfosDict( { '''default''': DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ) } ), DatasetInfosDict( { '''v1''': DatasetInfo(dataset_size=42 ), '''v2''': DatasetInfo(dataset_size=13_37 ), } ), ], ) def snake_case_ ( A_ : Optional[Any], A_ : DatasetInfosDict ): '''simple docstring''' _lowerCamelCase : List[str] = str(A_ ) dataset_infos_dict.write_to_directory(A_ ) _lowerCamelCase : List[Any] = DatasetInfosDict.from_directory(A_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): _lowerCamelCase : str = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml _lowerCamelCase : Any = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(A_, '''README.md''' ) )
83
0
import os import socket from contextlib import contextmanager import torch from ..commands.config.default import write_basic_config # noqa: F401 from ..state import PartialState from .dataclasses import DistributedType from .imports import is_deepspeed_available, is_tpu_available from .transformer_engine import convert_model from .versions import is_torch_version if is_deepspeed_available(): from deepspeed import DeepSpeedEngine if is_tpu_available(check_device=False): import torch_xla.core.xla_model as xm def A(__a: List[Any] ): if is_torch_version("<" , "2.0.0" ) or not hasattr(A_ , "_dynamo" ): return False return isinstance(A_ , torch._dynamo.eval_frame.OptimizedModule ) def A(__a: List[str] , __a: bool = True ): lowerCAmelCase_ = (torch.nn.parallel.DistributedDataParallel, torch.nn.DataParallel) lowerCAmelCase_ = is_compiled_module(A_ ) if is_compiled: lowerCAmelCase_ = model lowerCAmelCase_ = model._orig_mod if is_deepspeed_available(): options += (DeepSpeedEngine,) while isinstance(A_ , A_ ): lowerCAmelCase_ = model.module if not keep_fpaa_wrapper: lowerCAmelCase_ = getattr(A_ , "forward" ) lowerCAmelCase_ = model.__dict__.pop("_original_forward" , A_ ) if original_forward is not None: while hasattr(A_ , "__wrapped__" ): lowerCAmelCase_ = forward.__wrapped__ if forward == original_forward: break lowerCAmelCase_ = forward if getattr(A_ , "_converted_to_transformer_engine" , A_ ): convert_model(A_ , to_transformer_engine=A_ ) if is_compiled: lowerCAmelCase_ = model lowerCAmelCase_ = compiled_model return model def A(): PartialState().wait_for_everyone() def A(__a: Optional[int] , __a: str ): if PartialState().distributed_type == DistributedType.TPU: xm.save(A_ , A_ ) elif PartialState().local_process_index == 0: torch.save(A_ , A_ ) @contextmanager def A(**__a: Any ): for key, value in kwargs.items(): lowerCAmelCase_ = str(A_ ) yield for key in kwargs: if key.upper() in os.environ: del os.environ[key.upper()] def A(__a: Any ): if not hasattr(A_ , "__qualname__" ) and not hasattr(A_ , "__name__" ): lowerCAmelCase_ = getattr(A_ , "__class__" , A_ ) if hasattr(A_ , "__qualname__" ): return obj.__qualname__ if hasattr(A_ , "__name__" ): return obj.__name__ return str(A_ ) def A(__a: Any , __a: Dict ): for key, value in source.items(): if isinstance(A_ , A_ ): lowerCAmelCase_ = destination.setdefault(A_ , {} ) merge_dicts(A_ , A_ ) else: lowerCAmelCase_ = value return destination def A(__a: int = None ): if port is None: lowerCAmelCase_ = 2_9500 with socket.socket(socket.AF_INET , socket.SOCK_STREAM ) as s: return s.connect_ex(("localhost", port) ) == 0
122
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __snake_case : def __init__( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple=1_3 , __lowerCAmelCase : Any=2 , __lowerCAmelCase : List[str]=2_4 , __lowerCAmelCase : str=1_6 , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : Union[str, Any]=True , __lowerCAmelCase : Optional[Any]=3_2 , __lowerCAmelCase : List[Any]=5 , __lowerCAmelCase : int=4 , __lowerCAmelCase : int=3_7 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Dict=0.1 , __lowerCAmelCase : str=0.1 , __lowerCAmelCase : int=1_0 , __lowerCAmelCase : List[Any]=0.02 , __lowerCAmelCase : str=None , __lowerCAmelCase : List[str]=2 , __lowerCAmelCase : Union[str, Any]=2 , ): """simple docstring""" _lowerCamelCase : List[str] = parent _lowerCamelCase : str = batch_size _lowerCamelCase : Tuple = patch_size _lowerCamelCase : Optional[int] = max_length _lowerCamelCase : List[Any] = num_mel_bins _lowerCamelCase : int = is_training _lowerCamelCase : Union[str, Any] = use_labels _lowerCamelCase : Dict = hidden_size _lowerCamelCase : Tuple = num_hidden_layers _lowerCamelCase : int = num_attention_heads _lowerCamelCase : Tuple = intermediate_size _lowerCamelCase : List[str] = hidden_act _lowerCamelCase : Dict = hidden_dropout_prob _lowerCamelCase : int = attention_probs_dropout_prob _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : List[str] = scope _lowerCamelCase : Optional[int] = frequency_stride _lowerCamelCase : List[Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase : Union[str, Any] = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase : Any = frequency_out_dimension * time_out_dimension _lowerCamelCase : List[Any] = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : int = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase : str = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Optional[int] = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : List[Any] = ASTModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : List[Any] = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : int = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Optional[Any] = config_and_inputs _lowerCamelCase : int = {'''input_values''': input_values} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) snake_case__ : Tuple = ( {"audio-classification": ASTForAudioClassification, "feature-extraction": ASTModel} if is_torch_available() else {} ) snake_case__ : Any = False snake_case__ : List[Any] = False snake_case__ : Optional[Any] = False snake_case__ : Optional[Any] = False def SCREAMING_SNAKE_CASE ( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Optional[Any] ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[int] = ASTModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" pass def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Dict = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : str = model_class(__lowerCAmelCase ) _lowerCamelCase : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : Any = [*signature.parameters.keys()] _lowerCamelCase : str = ['''input_values'''] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Union[str, Any] = ASTModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase : str = torchaudio.load(A_ ) return audio, sampling_rate @require_torch @require_torchaudio class __snake_case ( unittest.TestCase): @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : int = self.default_feature_extractor _lowerCamelCase : Union[str, Any] = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(__lowerCAmelCase ) _lowerCamelCase : List[Any] = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase : List[Any] = prepare_audio() _lowerCamelCase : Dict = audio.squeeze().numpy() _lowerCamelCase : Tuple = feature_extractor(__lowerCAmelCase , sampling_rate=__lowerCAmelCase , return_tensors='''pt''' ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): _lowerCamelCase : Tuple = model(**__lowerCAmelCase ) # verify the logits _lowerCamelCase : Tuple = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) _lowerCamelCase : Optional[int] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4 ) )
83
0
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _lowerCAmelCase = 1_6 _lowerCAmelCase = 3_2 def lowerCamelCase__ ( _lowerCamelCase , _lowerCamelCase = 16 , _lowerCamelCase = "bert-base-cased" ): '''simple docstring''' _lowerCAmelCase : Optional[int] = AutoTokenizer.from_pretrained(A_ ) _lowerCAmelCase : Tuple = load_dataset('glue' , 'mrpc' ) def tokenize_function(_lowerCamelCase ): # max_length=None => use the model max length (it's actually the default) _lowerCAmelCase : str = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=A_ , max_length=A_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset _lowerCAmelCase : Any = datasets.map( A_ , batched=A_ , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=A_ ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library _lowerCAmelCase : List[Any] = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_lowerCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(A_ , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(A_ , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. _lowerCAmelCase : Optional[Any] = DataLoader( tokenized_datasets['train'] , shuffle=A_ , collate_fn=A_ , batch_size=A_ ) _lowerCAmelCase : int = DataLoader( tokenized_datasets['validation'] , shuffle=A_ , collate_fn=A_ , batch_size=A_ ) return train_dataloader, eval_dataloader def lowerCamelCase__ ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Union[str, Any] = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs _lowerCAmelCase : List[Any] = config['''lr'''] _lowerCAmelCase : Union[str, Any] = int(config['num_epochs'] ) _lowerCAmelCase : Optional[int] = int(config['seed'] ) _lowerCAmelCase : Tuple = int(config['batch_size'] ) _lowerCAmelCase : List[Any] = args.model_name_or_path set_seed(A_ ) _lowerCAmelCase : List[str] = get_dataloaders(A_ , A_ , A_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) _lowerCAmelCase : Union[str, Any] = AutoModelForSequenceClassification.from_pretrained(A_ , return_dict=A_ ) # Instantiate optimizer _lowerCAmelCase : Optional[Any] = ( AdamW if accelerator.state.deepspeed_plugin is None or '''optimizer''' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) _lowerCAmelCase : str = optimizer_cls(params=model.parameters() , lr=A_ ) if accelerator.state.deepspeed_plugin is not None: _lowerCAmelCase : Any = accelerator.state.deepspeed_plugin.deepspeed_config[ '''gradient_accumulation_steps''' ] else: _lowerCAmelCase : Union[str, Any] = 1 _lowerCAmelCase : Tuple = (len(A_ ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): _lowerCAmelCase : Union[str, Any] = get_linear_schedule_with_warmup( optimizer=A_ , num_warmup_steps=0 , num_training_steps=A_ , ) else: _lowerCAmelCase : Tuple = DummyScheduler(A_ , total_num_steps=A_ , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. _lowerCAmelCase : Any = accelerator.prepare( A_ , A_ , A_ , A_ , A_ ) # We need to keep track of how many total steps we have iterated over _lowerCAmelCase : Union[str, Any] = 0 # We also need to keep track of the stating epoch so files are named properly _lowerCAmelCase : Tuple = 0 # Now we train the model _lowerCAmelCase : int = evaluate.load('glue' , 'mrpc' ) _lowerCAmelCase : Union[str, Any] = 0 _lowerCAmelCase : List[str] = {} for epoch in range(A_ , A_ ): model.train() for step, batch in enumerate(A_ ): _lowerCAmelCase : Tuple = model(**A_ ) _lowerCAmelCase : List[str] = outputs.loss _lowerCAmelCase : Any = loss / gradient_accumulation_steps accelerator.backward(A_ ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 model.eval() _lowerCAmelCase : Any = 0 for step, batch in enumerate(A_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): _lowerCAmelCase : str = model(**A_ ) _lowerCAmelCase : Tuple = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times _lowerCAmelCase : Tuple = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(A_ ) - 1: _lowerCAmelCase : List[Any] = predictions[: len(eval_dataloader.dataset ) - samples_seen] _lowerCAmelCase : Optional[Any] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=A_ , references=A_ , ) _lowerCAmelCase : Optional[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , A_ ) _lowerCAmelCase : Optional[Any] = eval_metric['''accuracy'''] if best_performance < eval_metric["accuracy"]: _lowerCAmelCase : Optional[int] = eval_metric['''accuracy'''] if args.performance_lower_bound is not None: assert ( args.performance_lower_bound <= best_performance ), f"""Best performance metric {best_performance} is lower than the lower bound {args.performance_lower_bound}""" accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , 'all_results.json' ) , 'w' ) as f: json.dump(A_ , A_ ) def lowerCamelCase__ ( ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=A_ , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=A_ , ) parser.add_argument( '--output_dir' , type=A_ , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--performance_lower_bound' , type=A_ , default=A_ , help='Optional lower bound for the performance metric. If set, the training will throw error when the performance metric drops below this value.' , ) parser.add_argument( '--num_epochs' , type=A_ , default=3 , help='Number of train epochs.' , ) _lowerCAmelCase : Tuple = parser.parse_args() _lowerCAmelCase : str = {'''lr''': 2e-5, '''num_epochs''': args.num_epochs, '''seed''': 42, '''batch_size''': 16} training_function(A_ , A_ ) if __name__ == "__main__": main()
259
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def snake_case_ ( A_ : Dict, A_ : bool = True, A_ : float = math.inf, A_ : float = -math.inf, A_ : float = math.inf, A_ : float = -math.inf, A_ : bool = False, A_ : float = 1_00, A_ : float = 0.01, A_ : float = 1, ): '''simple docstring''' _lowerCamelCase : Optional[int] = False _lowerCamelCase : str = search_prob _lowerCamelCase : str = start_temperate _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : int = 0 _lowerCamelCase : Any = None while not search_end: _lowerCamelCase : Dict = current_state.score() if best_state is None or current_score > best_state.score(): _lowerCamelCase : Tuple = current_state scores.append(A_ ) iterations += 1 _lowerCamelCase : List[Any] = None _lowerCamelCase : Optional[int] = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to _lowerCamelCase : List[Any] = random.randint(0, len(A_ ) - 1 ) # picking a random neighbor _lowerCamelCase : Dict = neighbors.pop(A_ ) _lowerCamelCase : Union[str, Any] = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: _lowerCamelCase : str = change * -1 # in case we are finding minimum if change > 0: # improves the solution _lowerCamelCase : Optional[Any] = picked_neighbor else: _lowerCamelCase : Optional[int] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability _lowerCamelCase : Union[str, Any] = picked_neighbor _lowerCamelCase : List[str] = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor _lowerCamelCase : Tuple = True else: _lowerCamelCase : Optional[Any] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(A_ ), A_ ) plt.xlabel('''Iterations''' ) plt.ylabel('''Function values''' ) plt.show() return best_state if __name__ == "__main__": def snake_case_ ( A_ : int, A_ : Tuple ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def snake_case_ ( A_ : Optional[int], A_ : List[Any] ): '''simple docstring''' return (3 * x**2) - (6 * y) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" ) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" )
83
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( VersatileDiffusionDualGuidedPipeline, VersatileDiffusionImageVariationPipeline, VersatileDiffusionPipeline, VersatileDiffusionTextToImagePipeline, ) else: from .modeling_text_unet import UNetFlatConditionModel from .pipeline_versatile_diffusion import VersatileDiffusionPipeline from .pipeline_versatile_diffusion_dual_guided import VersatileDiffusionDualGuidedPipeline from .pipeline_versatile_diffusion_image_variation import VersatileDiffusionImageVariationPipeline from .pipeline_versatile_diffusion_text_to_image import VersatileDiffusionTextToImagePipeline
439
"""simple docstring""" from collections import namedtuple lowerCAmelCase__ = namedtuple('''from_to''', '''from_ to''') lowerCAmelCase__ = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.0_0_1, 1000), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.0_0_4_5_4, 2_6_4.1_7_2), '''cubicyard''': from_to(0.7_6_4_5_5, 1.3_0_7_9_5), '''cubicfoot''': from_to(0.0_2_8, 3_5.3_1_4_7), '''cup''': from_to(0.0_0_0_2_3_6_5_8_8, 4_2_2_6.7_5), } def snake_case_ ( A_ : float, A_ : str, A_ : str ): '''simple docstring''' if from_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + ''', '''.join(A_ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + ''', '''.join(A_ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
83
0
from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def lowercase_( ): '''simple docstring''' lowerCamelCase : Optional[int] = 9, 14 # noqa: F841 lowerCamelCase : Dict = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] lowerCamelCase : Union[str, Any] = defaultdict(A_ ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) lowerCamelCase : List[str] = mst(A_ ) lowerCamelCase : Union[str, Any] = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: lowerCamelCase : str = tuple(answer[:2] ) lowerCamelCase : Union[str, Any] = tuple(edge[::-1] ) assert edge in result or reverse in result
340
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case ( _lowercase): def __init__( self : Dict , __lowerCAmelCase : Any , __lowerCAmelCase : int=1_3 , __lowerCAmelCase : Optional[int]=7 , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : str=True , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : int=True , __lowerCAmelCase : List[str]=9_9 , __lowerCAmelCase : List[Any]=3_2 , __lowerCAmelCase : Union[str, Any]=5 , __lowerCAmelCase : List[str]=4 , __lowerCAmelCase : Optional[int]=3_7 , __lowerCAmelCase : List[Any]="gelu" , __lowerCAmelCase : Tuple=0.1 , __lowerCAmelCase : List[Any]=0.1 , __lowerCAmelCase : int=5_1_2 , __lowerCAmelCase : Tuple=1_6 , __lowerCAmelCase : Tuple=2 , __lowerCAmelCase : Tuple=0.02 , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : Any="None" , __lowerCAmelCase : str=3 , __lowerCAmelCase : Dict=4 , __lowerCAmelCase : Optional[Any]=None , ): """simple docstring""" _lowerCamelCase : Dict = parent _lowerCamelCase : Union[str, Any] = batch_size _lowerCamelCase : Optional[Any] = seq_length _lowerCamelCase : Optional[Any] = is_training _lowerCamelCase : Dict = use_input_mask _lowerCamelCase : Tuple = use_token_type_ids _lowerCamelCase : Optional[Any] = use_labels _lowerCamelCase : List[str] = vocab_size _lowerCamelCase : Any = hidden_size _lowerCamelCase : int = num_hidden_layers _lowerCamelCase : Optional[Any] = num_attention_heads _lowerCamelCase : int = intermediate_size _lowerCamelCase : Optional[int] = hidden_act _lowerCamelCase : int = hidden_dropout_prob _lowerCamelCase : Dict = attention_probs_dropout_prob _lowerCamelCase : List[Any] = max_position_embeddings _lowerCamelCase : str = type_vocab_size _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : List[Any] = initializer_range _lowerCamelCase : Optional[int] = num_labels _lowerCamelCase : Any = num_choices _lowerCamelCase : int = relative_attention _lowerCamelCase : Union[str, Any] = position_biased_input _lowerCamelCase : str = pos_att_type _lowerCamelCase : Tuple = scope def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : List[Any] = None if self.use_input_mask: _lowerCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _lowerCamelCase : Any = None if self.use_token_type_ids: _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : Any = None _lowerCamelCase : int = None _lowerCamelCase : Union[str, Any] = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase : List[str] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : str ): """simple docstring""" self.parent.assertListEqual(list(result.loss.size() ) , [] ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : str = model(__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : List[Any] = model(__lowerCAmelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __lowerCAmelCase : Any , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaForMaskedLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[int] = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = self.num_labels _lowerCamelCase : Dict = DebertaVaForSequenceClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : List[str] , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.num_labels _lowerCamelCase : Tuple = DebertaVaForTokenClassification(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Any = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaForQuestionAnswering(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Tuple = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , start_positions=__lowerCAmelCase , end_positions=__lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : Optional[int] = DebertaVaForMultipleChoice(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Union[str, Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : List[Any] = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Any = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCamelCase : Optional[int] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) snake_case__ : int = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : List[str] = True snake_case__ : List[Any] = False snake_case__ : int = False snake_case__ : Optional[Any] = False snake_case__ : str = False def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class __snake_case ( unittest.TestCase): @unittest.skip(reason='''Model not available yet''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" pass @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained('''microsoft/deberta-v2-xlarge''' ) _lowerCamelCase : List[str] = torch.tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) _lowerCamelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _lowerCamelCase : Tuple = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase )[0] # compare the actual values for a slice. _lowerCamelCase : Union[str, Any] = torch.tensor( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowerCAmelCase , atol=1E-4 ) , f'''{output[:, 1:4, 1:4]}''' )
83
0
from .constants import ( MODEL_NAME, OPTIMIZER_NAME, RNG_STATE_NAME, SAFE_WEIGHTS_INDEX_NAME, SAFE_WEIGHTS_NAME, SCALER_NAME, SCHEDULER_NAME, TORCH_LAUNCH_PARAMS, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ) from .dataclasses import ( BnbQuantizationConfig, ComputeEnvironment, CustomDtype, DeepSpeedPlugin, DistributedDataParallelKwargs, DistributedType, DynamoBackend, FPaRecipeKwargs, FullyShardedDataParallelPlugin, GradientAccumulationPlugin, GradScalerKwargs, InitProcessGroupKwargs, KwargsHandler, LoggerType, MegatronLMPlugin, PrecisionType, ProjectConfiguration, RNGType, SageMakerDistributedType, TensorInformation, TorchDynamoPlugin, ) from .environment import get_int_from_env, parse_choice_from_env, parse_flag_from_env from .imports import ( get_ccl_version, is_abit_bnb_available, is_abit_bnb_available, is_aim_available, is_bfaa_available, is_bnb_available, is_botoa_available, is_ccl_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_fpa_available, is_ipex_available, is_megatron_lm_available, is_mlflow_available, is_mps_available, is_npu_available, is_rich_available, is_safetensors_available, is_sagemaker_available, is_tensorboard_available, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) from .modeling import ( check_device_map, check_tied_parameters_in_config, check_tied_parameters_on_same_device, compute_module_sizes, convert_file_size_to_int, dtype_byte_size, find_tied_parameters, get_balanced_memory, get_max_layer_size, get_max_memory, get_mixed_precision_context_manager, id_tensor_storage, infer_auto_device_map, load_checkpoint_in_model, load_offloaded_weights, load_state_dict, named_module_tensors, retie_parameters, set_module_tensor_to_device, shard_checkpoint, ) from .offload import ( OffloadedWeightsLoader, PrefixedDataset, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, save_offload_index, ) from .operations import ( broadcast, broadcast_object_list, concatenate, convert_outputs_to_fpaa, convert_to_fpaa, find_batch_size, find_device, gather, gather_object, get_data_structure, honor_type, initialize_tensors, is_namedtuple, is_tensor_information, is_torch_tensor, listify, pad_across_processes, recursively_apply, reduce, send_to_device, slice_tensors, ) from .versions import compare_versions, is_torch_version if is_deepspeed_available(): from .deepspeed import ( DeepSpeedEngineWrapper, DeepSpeedOptimizerWrapper, DeepSpeedSchedulerWrapper, DummyOptim, DummyScheduler, HfDeepSpeedConfig, ) from .bnb import has_abit_bnb_layers, load_and_quantize_model from .fsdp_utils import load_fsdp_model, load_fsdp_optimizer, save_fsdp_model, save_fsdp_optimizer from .launch import ( PrepareForLaunch, _filter_args, prepare_deepspeed_cmd_env, prepare_multi_gpu_env, prepare_sagemager_args_inputs, prepare_simple_launcher_cmd_env, prepare_tpu, ) from .megatron_lm import ( AbstractTrainStep, BertTrainStep, GPTTrainStep, MegatronEngine, MegatronLMDummyDataLoader, MegatronLMDummyScheduler, MegatronLMOptimizerWrapper, MegatronLMSchedulerWrapper, TaTrainStep, avg_losses_across_data_parallel_group, gather_across_data_parallel_groups, ) from .megatron_lm import initialize as megatron_lm_initialize from .megatron_lm import prepare_data_loader as megatron_lm_prepare_data_loader from .megatron_lm import prepare_model as megatron_lm_prepare_model from .megatron_lm import prepare_optimizer as megatron_lm_prepare_optimizer from .megatron_lm import prepare_scheduler as megatron_lm_prepare_scheduler from .memory import find_executable_batch_size, release_memory from .other import ( extract_model_from_parallel, get_pretty_name, is_port_in_use, merge_dicts, patch_environment, save, wait_for_everyone, write_basic_config, ) from .random import set_seed, synchronize_rng_state, synchronize_rng_states from .torch_xla import install_xla from .tqdm import tqdm from .transformer_engine import convert_model, has_transformer_engine_layers
9
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCAmelCase__ = '''hf-internal-testing/tiny-random-bert''' lowerCAmelCase__ = os.path.join(TRANSFORMERS_CACHE, '''models--hf-internal-testing--tiny-random-bert''') lowerCAmelCase__ = '''9b8c223d42b2188cb49d29af482996f9d0f3e5a6''' class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = cached_file(__lowerCAmelCase , __lowerCAmelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(__lowerCAmelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) ) ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : Optional[int] = f.read() self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) self.assertTrue(os.path.isfile(__lowerCAmelCase ) ) # File is cached at the same place the second time. _lowerCamelCase : Tuple = cached_file(__lowerCAmelCase , __lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) # Using a specific revision to test the full commit hash. _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''9b8c223''' ) self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): _lowerCamelCase : Optional[int] = cached_file('''tiny-random-bert''' , __lowerCAmelCase ) with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): _lowerCamelCase : str = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''aaaa''' ) with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : int = cached_file(__lowerCAmelCase , '''conf''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , '''conf''' ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : List[Any] = f.read() self.assertTrue(os.path.isfile(os.path.join(__lowerCAmelCase , '''.no_exist''' , __lowerCAmelCase , '''conf''' ) ) ) _lowerCamelCase : str = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Optional[int] = cached_file(__lowerCAmelCase , '''conf''' , local_files_only=__lowerCAmelCase , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Any = mock.Mock() _lowerCamelCase : Optional[Any] = 5_0_0 _lowerCamelCase : Dict = {} _lowerCamelCase : List[Any] = HTTPError _lowerCamelCase : int = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=__lowerCAmelCase ) as mock_head: _lowerCamelCase : Union[str, Any] = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_connection_errors=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , __lowerCAmelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase , revision='''ahaha''' ) _lowerCamelCase : Dict = get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. _lowerCamelCase : Dict = json.loads(open(__lowerCAmelCase , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_6_8 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase : Any = Path(__lowerCAmelCase ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(__lowerCAmelCase , '''a.txt''' ) , str(__lowerCAmelCase ) ) self.assertIsNone(get_file_from_repo(__lowerCAmelCase , '''b.txt''' ) )
83
0
import warnings from ...utils import logging from .image_processing_chinese_clip import ChineseCLIPImageProcessor lowerCamelCase__ : Tuple = logging.get_logger(__name__) class __magic_name__ (_lowercase ): '''simple docstring''' def __init__( self:Dict , *_a:List[Any] , **_a:Union[str, Any] ): warnings.warn( '''The class ChineseCLIPFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use ChineseCLIPImageProcessor instead.''' , __lowerCAmelCase , ) super().__init__(*__lowerCAmelCase , **__lowerCAmelCase )
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class __snake_case ( _lowercase): snake_case__ : List[str] = "cvt" def __init__( self : Any , __lowerCAmelCase : Union[str, Any]=3 , __lowerCAmelCase : List[str]=[7, 3, 3] , __lowerCAmelCase : int=[4, 2, 2] , __lowerCAmelCase : int=[2, 1, 1] , __lowerCAmelCase : str=[6_4, 1_9_2, 3_8_4] , __lowerCAmelCase : Dict=[1, 3, 6] , __lowerCAmelCase : Optional[Any]=[1, 2, 1_0] , __lowerCAmelCase : Dict=[4.0, 4.0, 4.0] , __lowerCAmelCase : Dict=[0.0, 0.0, 0.0] , __lowerCAmelCase : Union[str, Any]=[0.0, 0.0, 0.0] , __lowerCAmelCase : int=[0.0, 0.0, 0.1] , __lowerCAmelCase : Union[str, Any]=[True, True, True] , __lowerCAmelCase : str=[False, False, True] , __lowerCAmelCase : List[str]=["dw_bn", "dw_bn", "dw_bn"] , __lowerCAmelCase : List[Any]=[3, 3, 3] , __lowerCAmelCase : Dict=[1, 1, 1] , __lowerCAmelCase : str=[2, 2, 2] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : str=0.02 , __lowerCAmelCase : Any=1E-12 , **__lowerCAmelCase : int , ): """simple docstring""" super().__init__(**__lowerCAmelCase ) _lowerCamelCase : Optional[int] = num_channels _lowerCamelCase : int = patch_sizes _lowerCamelCase : Optional[Any] = patch_stride _lowerCamelCase : str = patch_padding _lowerCamelCase : Any = embed_dim _lowerCamelCase : Optional[Any] = num_heads _lowerCamelCase : Dict = depth _lowerCamelCase : Optional[int] = mlp_ratio _lowerCamelCase : Any = attention_drop_rate _lowerCamelCase : Any = drop_rate _lowerCamelCase : Dict = drop_path_rate _lowerCamelCase : Optional[int] = qkv_bias _lowerCamelCase : int = cls_token _lowerCamelCase : int = qkv_projection_method _lowerCamelCase : Optional[Any] = kernel_qkv _lowerCamelCase : List[str] = padding_kv _lowerCamelCase : Tuple = stride_kv _lowerCamelCase : Union[str, Any] = padding_q _lowerCamelCase : Optional[Any] = stride_q _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Optional[int] = layer_norm_eps
83
0
from __future__ import annotations from fractions import Fraction def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Union[str, Any]: return ( num != den and num % 10 == den // 10 and (num // 10) / (den % 10) == num / den ) def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE ) -> Optional[int]: SCREAMING_SNAKE_CASE_ : Optional[int] = [] SCREAMING_SNAKE_CASE_ : int = 11 SCREAMING_SNAKE_CASE_ : Any = int('1' + '0' * digit_len ) for num in range(A_ , A_ ): while den <= 99: if (num != den) and (num % 10 == den // 10) and (den % 10 != 0): if is_digit_cancelling(A_ , A_ ): solutions.append(f'{num}/{den}' ) den += 1 num += 1 SCREAMING_SNAKE_CASE_ : Dict = 10 return solutions def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE = 2 ) -> Union[str, Any]: SCREAMING_SNAKE_CASE_ : List[str] = 1.0 for fraction in fraction_list(A_ ): SCREAMING_SNAKE_CASE_ : List[Any] = Fraction(A_ ) result *= frac.denominator / frac.numerator return int(A_ ) if __name__ == "__main__": print(solution())
345
"""simple docstring""" # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class __snake_case ( _lowercase): snake_case__ : torch.FloatTensor snake_case__ : torch.FloatTensor class __snake_case ( _lowercase , _lowercase): snake_case__ : int = 1 @register_to_config def __init__( self : str , __lowerCAmelCase : int = 2_0_0_0 , __lowerCAmelCase : float = 0.15 , __lowerCAmelCase : float = 0.01 , __lowerCAmelCase : float = 13_48.0 , __lowerCAmelCase : float = 1E-5 , __lowerCAmelCase : int = 1 , ): """simple docstring""" _lowerCamelCase : Optional[int] = sigma_max # setable values _lowerCamelCase : Dict = None self.set_sigmas(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[int] = None ): """simple docstring""" return sample def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : Union[str, torch.device] = None ): """simple docstring""" _lowerCamelCase : Tuple = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCamelCase : Optional[int] = torch.linspace(1 , __lowerCAmelCase , __lowerCAmelCase , device=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None ): """simple docstring""" _lowerCamelCase : List[str] = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCamelCase : int = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCamelCase : Any = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCamelCase : Optional[int] = torch.exp(torch.linspace(math.log(__lowerCAmelCase ) , math.log(__lowerCAmelCase ) , __lowerCAmelCase ) ) _lowerCamelCase : Tuple = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def SCREAMING_SNAKE_CASE ( self : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[Any] ): """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : int , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) _lowerCamelCase : Tuple = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCamelCase : Dict = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCamelCase : Optional[int] = timesteps.to(self.discrete_sigmas.device ) _lowerCamelCase : Any = self.discrete_sigmas[timesteps].to(sample.device ) _lowerCamelCase : int = self.get_adjacent_sigma(__lowerCAmelCase , __lowerCAmelCase ).to(sample.device ) _lowerCamelCase : Any = torch.zeros_like(__lowerCAmelCase ) _lowerCamelCase : Any = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCamelCase : Union[str, Any] = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _lowerCamelCase : List[Any] = diffusion.unsqueeze(-1 ) _lowerCamelCase : int = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCamelCase : List[str] = randn_tensor( sample.shape , layout=sample.layout , generator=__lowerCAmelCase , device=sample.device , dtype=sample.dtype ) _lowerCamelCase : List[Any] = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCamelCase : int = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=__lowerCAmelCase , prev_sample_mean=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCamelCase : Union[str, Any] = randn_tensor(sample.shape , layout=sample.layout , generator=__lowerCAmelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _lowerCamelCase : Union[str, Any] = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : Tuple = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : str = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCamelCase : Tuple = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCamelCase : Union[str, Any] = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _lowerCamelCase : str = step_size.unsqueeze(-1 ) _lowerCamelCase : Any = sample + step_size * model_output _lowerCamelCase : int = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , ): """simple docstring""" _lowerCamelCase : Dict = timesteps.to(original_samples.device ) _lowerCamelCase : Union[str, Any] = self.discrete_sigmas.to(original_samples.device )[timesteps] _lowerCamelCase : Union[str, Any] = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(__lowerCAmelCase ) * sigmas[:, None, None, None] ) _lowerCamelCase : int = noise + original_samples return noisy_samples def __len__( self : Optional[int] ): """simple docstring""" return self.config.num_train_timesteps
83
0
def _UpperCAmelCase ( A ): '''simple docstring''' return 10 - x * x def _UpperCAmelCase ( A , A ): '''simple docstring''' if equation(A_ ) * equation(A_ ) >= 0: raise ValueError("Wrong space!" ) UpperCAmelCase__ =a while (b - a) >= 0.01: # Find middle point UpperCAmelCase__ =(a + b) / 2 # Check if middle point is root if equation(A_ ) == 0.0: break # Decide the side to repeat the steps if equation(A_ ) * equation(A_ ) < 0: UpperCAmelCase__ =c else: UpperCAmelCase__ =c return c if __name__ == "__main__": import doctest doctest.testmod() print(bisection(-2, 5)) print(bisection(0, 6))
625
"""simple docstring""" from torch import nn def snake_case_ ( A_ : int ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'''Unsupported activation function: {act_fn}''' )
83
0
import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class A ( _lowercase , unittest.TestCase ): UpperCamelCase__ : List[str] =VideoToVideoSDPipeline UpperCamelCase__ : Any =TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({'video'} ) - {"image", "width", "height"} UpperCamelCase__ : List[Any] =TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'video'} ) - {"image"} UpperCamelCase__ : Tuple =PipelineTesterMixin.required_optional_params - {"latents"} UpperCamelCase__ : str =False # No `output_type`. UpperCamelCase__ : int =frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def lowerCamelCase ( self : List[Any] ) -> Any: """simple docstring""" torch.manual_seed(0 ) _lowerCamelCase : Any =UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'CrossAttnDownBlock3D', 'DownBlock3D') , up_block_types=('UpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D', 'CrossAttnUpBlock3D') , cross_attention_dim=32 , attention_head_dim=4 , ) _lowerCamelCase : str =DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=__lowerCAmelCase , set_alpha_to_one=__lowerCAmelCase , ) torch.manual_seed(0 ) _lowerCamelCase : Optional[int] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) _lowerCamelCase : List[Any] =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act='gelu' , projection_dim=512 , ) _lowerCamelCase : List[str] =CLIPTextModel(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) _lowerCamelCase : int ={ '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def lowerCamelCase ( self : int , lowercase_ : Dict , lowercase_ : Any=0 ) -> Dict: """simple docstring""" _lowerCamelCase : str =floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(__lowerCAmelCase ) ).to(__lowerCAmelCase ) if str(__lowerCAmelCase ).startswith('mps' ): _lowerCamelCase : List[Any] =torch.manual_seed(__lowerCAmelCase ) else: _lowerCamelCase : Tuple =torch.Generator(device=__lowerCAmelCase ).manual_seed(__lowerCAmelCase ) _lowerCamelCase : Optional[int] ={ '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def lowerCamelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" _lowerCamelCase : int ='''cpu''' # ensure determinism for the device-dependent torch.Generator _lowerCamelCase : Dict =self.get_dummy_components() _lowerCamelCase : int =VideoToVideoSDPipeline(**__lowerCAmelCase ) _lowerCamelCase : str =sd_pipe.to(__lowerCAmelCase ) sd_pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : Dict =self.get_dummy_inputs(__lowerCAmelCase ) _lowerCamelCase : Optional[Any] ='''np''' _lowerCamelCase : Tuple =sd_pipe(**__lowerCAmelCase ).frames _lowerCamelCase : Dict =frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _lowerCamelCase : Dict =np.array([106, 117, 113, 174, 137, 112, 148, 151, 131] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf( torch_device != 'cuda' or not is_xformers_available() , reason='XFormers attention is only available with CUDA and `xformers` installed' , ) def lowerCamelCase ( self : List[str] ) -> int: """simple docstring""" self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=__lowerCAmelCase , expected_max_diff=5E-3 ) @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def lowerCamelCase ( self : List[Any] ) -> Optional[Any]: """simple docstring""" pass @unittest.skip(reason='Batching needs to be properly figured out first for this pipeline.' ) def lowerCamelCase ( self : List[str] ) -> Union[str, Any]: """simple docstring""" pass @unittest.skip(reason='`num_images_per_prompt` argument is not supported for this pipeline.' ) def lowerCamelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" pass def lowerCamelCase ( self : Tuple ) -> List[str]: """simple docstring""" return super().test_progress_bar() @slow @skip_mps class A ( unittest.TestCase ): def lowerCamelCase ( self : Optional[Any] ) -> Dict: """simple docstring""" _lowerCamelCase : Any =VideoToVideoSDPipeline.from_pretrained('cerspense/zeroscope_v2_XL' , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames _lowerCamelCase : Union[str, Any] =torch.Generator(device='cpu' ).manual_seed(0 ) _lowerCamelCase : Optional[int] =torch.randn((1, 10, 3, 1024, 576) , generator=__lowerCAmelCase ) _lowerCamelCase : List[Any] =video.to('cuda' ) _lowerCamelCase : Tuple ='''Spiderman is surfing''' _lowerCamelCase : Any =pipe(__lowerCAmelCase , video=__lowerCAmelCase , generator=__lowerCAmelCase , num_inference_steps=3 , output_type='pt' ).frames _lowerCamelCase : Dict =np.array([-1.0458984, -1.1279297, -0.9663086, -0.91503906, -0.75097656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1E-2
464
"""simple docstring""" def snake_case_ ( A_ : int, A_ : int ): '''simple docstring''' return int(input_a == input_a == 0 ) def snake_case_ ( ): '''simple docstring''' print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(F'''| 0 | 0 | {nor_gate(0, 0 )} |''' ) print(F'''| 0 | 1 | {nor_gate(0, 1 )} |''' ) print(F'''| 1 | 0 | {nor_gate(1, 0 )} |''' ) print(F'''| 1 | 1 | {nor_gate(1, 1 )} |''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
83
0
'''simple docstring''' from __future__ import annotations from cmath import sqrt def _UpperCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): if a == 0: raise ValueError("""Coefficient \'a\' must not be zero.""" ) UpperCAmelCase__ : List[str] = b * b - 4 * a * c UpperCAmelCase__ : List[Any] = (-b + sqrt(A_ )) / (2 * a) UpperCAmelCase__ : Optional[Any] = (-b - sqrt(A_ )) / (2 * a) return ( root_a.real if not root_a.imag else root_a, root_a.real if not root_a.imag else root_a, ) def _UpperCamelCase ( ): UpperCAmelCase__ : Optional[Any] = quadratic_roots(a=5 , b=6 , c=1 ) print(f'''The solutions are: {solutiona} and {solutiona}''' ) if __name__ == "__main__": main()
407
"""simple docstring""" from __future__ import annotations def snake_case_ ( A_ : list[list[int]] ): '''simple docstring''' for i in range(1, len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1, len(A_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1, len(A_ ) ): for j in range(1, len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j], matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
83
0
'''simple docstring''' import os import tempfile import unittest import numpy as np from diffusers.utils import is_flax_available from diffusers.utils.testing_utils import require_flax, slow if is_flax_available(): import jax import jax.numpy as jnp from flax.jax_utils import replicate from flax.training.common_utils import shard from diffusers import FlaxDDIMScheduler, FlaxDiffusionPipeline, FlaxStableDiffusionPipeline @require_flax class UpperCAmelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase__ ( self : Dict ) -> str: with tempfile.TemporaryDirectory() as tmpdirname: # pipeline has Flax weights _UpperCamelCase =FlaxDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__lowerCAmelCase , cache_dir=__lowerCAmelCase ) _UpperCamelCase =[t[-1] for t in os.walk(os.path.join(__lowerCAmelCase , os.listdir(__lowerCAmelCase )[0] , '''snapshots''' ) )] _UpperCamelCase =[item for sublist in all_root_files for item in sublist] # None of the downloaded files should be a PyTorch file even if we have some here: # https://huggingface.co/hf-internal-testing/tiny-stable-diffusion-pipe/blob/main/unet/diffusion_pytorch_model.bin assert not any(f.endswith('''.bin''' ) for f in files ) @slow @require_flax class UpperCAmelCase ( unittest.TestCase): """simple docstring""" def UpperCamelCase__ ( self : Any ) -> int: _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''hf-internal-testing/tiny-stable-diffusion-pipe''' , safety_checker=__lowerCAmelCase ) _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.random.PRNGKey(0 ) _UpperCamelCase =4 _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) # shard inputs and rng _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =jax.random.split(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 64, 64, 3) if jax.device_count() == 8: assert np.abs(np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 4.1514745 ) < 1E-3 assert np.abs(np.abs(__lowerCAmelCase , dtype=np.floataa ).sum() - 4_9947.875 ) < 5E-1 _UpperCamelCase =pipeline.numpy_to_pil(np.asarray(images.reshape((num_samples,) + images.shape[-3:] ) ) ) assert len(__lowerCAmelCase ) == num_samples def UpperCamelCase__ ( self : Optional[Any] ) -> List[str]: _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''flax''' , safety_checker=__lowerCAmelCase ) _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.random.PRNGKey(0 ) _UpperCamelCase =50 _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) # shard inputs and rng _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =jax.random.split(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.05652401) ) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa ).sum() - 238_3808.2) ) < 5E-1 def UpperCamelCase__ ( self : List[str] ) -> str: _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase ) _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.random.PRNGKey(0 ) _UpperCamelCase =50 _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) # shard inputs and rng _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =jax.random.split(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04003906) ) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def UpperCamelCase__ ( self : Any ) -> int: _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa ) _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.random.PRNGKey(0 ) _UpperCamelCase =50 _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) # shard inputs and rng _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =jax.random.split(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.04003906) ) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa ).sum() - 237_3516.75) ) < 5E-1 def UpperCamelCase__ ( self : List[Any] ) -> Dict: _UpperCamelCase =FlaxDDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule='''scaled_linear''' , set_alpha_to_one=__lowerCAmelCase , steps_offset=1 , ) _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , scheduler=__lowerCAmelCase , safety_checker=__lowerCAmelCase , ) _UpperCamelCase =scheduler.create_state() _UpperCamelCase =scheduler_state _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.random.PRNGKey(0 ) _UpperCamelCase =50 _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) # shard inputs and rng _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =jax.random.split(__lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) if jax.device_count() == 8: assert np.abs((np.abs(images[0, 0, :2, :2, -2:] , dtype=np.floataa ).sum() - 0.045043945) ) < 1E-3 assert np.abs((np.abs(__lowerCAmelCase , dtype=np.floataa ).sum() - 234_7693.5) ) < 5E-1 def UpperCamelCase__ ( self : Tuple ) -> int: _UpperCamelCase =( '''A cinematic film still of Morgan Freeman starring as Jimi Hendrix, portrait, 40mm lens, shallow depth of''' ''' field, close up, split lighting, cinematic''' ) _UpperCamelCase =jax.device_count() _UpperCamelCase =num_samples * [prompt] _UpperCamelCase =jax.random.split(jax.random.PRNGKey(0 ) , __lowerCAmelCase ) _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase , ) _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images.shape == (num_samples, 1, 512, 512, 3) _UpperCamelCase =images[2, 0, 256, 10:17, 1] # With memory efficient attention _UpperCamelCase =FlaxStableDiffusionPipeline.from_pretrained( '''CompVis/stable-diffusion-v1-4''' , revision='''bf16''' , dtype=jnp.bfloataa , safety_checker=__lowerCAmelCase , use_memory_efficient_attention=__lowerCAmelCase , ) _UpperCamelCase =replicate(__lowerCAmelCase ) _UpperCamelCase =pipeline.prepare_inputs(__lowerCAmelCase ) _UpperCamelCase =shard(__lowerCAmelCase ) _UpperCamelCase =pipeline(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , jit=__lowerCAmelCase ).images assert images_eff.shape == (num_samples, 1, 512, 512, 3) _UpperCamelCase =images[2, 0, 256, 10:17, 1] # I checked the results visually and they are very similar. However, I saw that the max diff is `1` and the `sum` # over the 8 images is exactly `256`, which is very suspicious. Testing a random slice for now. assert abs(slice_eff - slice ).max() < 1E-2
404
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''T''') class __snake_case ( Generic[T]): def __init__( self : int , __lowerCAmelCase : T ): """simple docstring""" _lowerCamelCase : Optional[int] = data _lowerCamelCase : Node[T] | None = None def __str__( self : Optional[Any] ): """simple docstring""" return f'''{self.data}''' class __snake_case ( Generic[T]): def __init__( self : int ): """simple docstring""" _lowerCamelCase : Node[T] | None = None def __iter__( self : str ): """simple docstring""" _lowerCamelCase : List[str] = self.top while node: yield node.data _lowerCamelCase : Any = node.next def __str__( self : int ): """simple docstring""" return "->".join([str(__lowerCAmelCase ) for item in self] ) def __len__( self : int ): """simple docstring""" return len(tuple(iter(self ) ) ) def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" return self.top is None def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : T ): """simple docstring""" _lowerCamelCase : Tuple = Node(__lowerCAmelCase ) if not self.is_empty(): _lowerCamelCase : Optional[int] = self.top _lowerCamelCase : List[str] = node def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" if self.is_empty(): raise IndexError('''pop from empty stack''' ) assert isinstance(self.top , __lowerCAmelCase ) _lowerCamelCase : Any = self.top _lowerCamelCase : Any = self.top.next return pop_node.data def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" if self.is_empty(): raise IndexError('''peek from empty stack''' ) assert self.top is not None return self.top.data def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : List[str] = None if __name__ == "__main__": from doctest import testmod testmod()
83
0
import unittest from transformers import AlbertTokenizer, AlbertTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin lowerCamelCase__ = get_tests_dir('''fixtures/spiece.model''') @require_sentencepiece @require_tokenizers class __magic_name__ (_lowercase , unittest.TestCase ): lowerCamelCase__ = AlbertTokenizer lowerCamelCase__ = AlbertTokenizerFast lowerCamelCase__ = True lowerCamelCase__ = True lowerCamelCase__ = True def __a ( self ) -> Optional[int]: super().setUp() # We have a SentencePiece fixture for testing lowerCAmelCase_ = AlbertTokenizer(__lowerCAmelCase ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self , _a ) -> Dict: lowerCAmelCase_ = '''this is a test''' lowerCAmelCase_ = '''this is a test''' return input_text, output_text def __a ( self ) -> Dict: lowerCAmelCase_ = '''<pad>''' lowerCAmelCase_ = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__lowerCAmelCase ) , __lowerCAmelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__lowerCAmelCase ) , __lowerCAmelCase ) def __a ( self ) -> int: lowerCAmelCase_ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<pad>" ) self.assertEqual(vocab_keys[1] , "<unk>" ) self.assertEqual(vocab_keys[-1] , "▁eloquent" ) self.assertEqual(len(__lowerCAmelCase ) , 30000 ) def __a ( self ) -> Optional[Any]: self.assertEqual(self.get_tokenizer().vocab_size , 30000 ) def __a ( self ) -> Optional[int]: if not self.test_rust_tokenizer: return lowerCAmelCase_ = self.get_tokenizer() lowerCAmelCase_ = self.get_rust_tokenizer() lowerCAmelCase_ = '''I was born in 92000, and this is falsé.''' lowerCAmelCase_ = tokenizer.tokenize(__lowerCAmelCase ) lowerCAmelCase_ = rust_tokenizer.tokenize(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) lowerCAmelCase_ = tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ) lowerCAmelCase_ = rust_tokenizer.encode(__lowerCAmelCase , add_special_tokens=__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) lowerCAmelCase_ = self.get_rust_tokenizer() lowerCAmelCase_ = tokenizer.encode(__lowerCAmelCase ) lowerCAmelCase_ = rust_tokenizer.encode(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase ) def __a ( self ) -> Any: lowerCAmelCase_ = AlbertTokenizer(__lowerCAmelCase , keep_accents=__lowerCAmelCase ) lowerCAmelCase_ = tokenizer.tokenize("This is a test" ) self.assertListEqual(__lowerCAmelCase , ["▁this", "▁is", "▁a", "▁test"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) , [48, 25, 21, 1289] ) lowerCAmelCase_ = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( __lowerCAmelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "é", "."] ) lowerCAmelCase_ = tokenizer.convert_tokens_to_ids(__lowerCAmelCase ) self.assertListEqual(__lowerCAmelCase , [31, 23, 386, 19, 561, 3050, 15, 17, 48, 25, 8256, 18, 1, 9] ) lowerCAmelCase_ = tokenizer.convert_ids_to_tokens(__lowerCAmelCase ) self.assertListEqual( __lowerCAmelCase , ["▁i", "▁was", "▁born", "▁in", "▁9", "2000", ",", "▁and", "▁this", "▁is", "▁fal", "s", "<unk>", "."] , ) def __a ( self ) -> Dict: lowerCAmelCase_ = AlbertTokenizer(__lowerCAmelCase ) lowerCAmelCase_ = tokenizer.encode("sequence builders" ) lowerCAmelCase_ = tokenizer.encode("multi-sequence build" ) lowerCAmelCase_ = tokenizer.build_inputs_with_special_tokens(__lowerCAmelCase ) lowerCAmelCase_ = tokenizer.build_inputs_with_special_tokens(__lowerCAmelCase , __lowerCAmelCase ) assert encoded_sentence == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] assert encoded_pair == [tokenizer.cls_token_id] + text + [tokenizer.sep_token_id] + text_a + [ tokenizer.sep_token_id ] @slow def __a ( self ) -> List[str]: lowerCAmelCase_ = {'''attention_mask''': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''input_ids''': [[2, 21970, 13, 5, 6092, 167, 28, 7103, 2153, 673, 8, 7028, 12051, 18, 17, 7103, 2153, 673, 8, 3515, 18684, 8, 4461, 6, 1927, 297, 8, 12060, 2607, 18, 13, 5, 4461, 15, 10538, 38, 8, 135, 15, 822, 58, 15, 993, 10363, 15, 1460, 8005, 4461, 15, 993, 255, 2328, 9, 9, 9, 6, 26, 1112, 816, 3260, 13, 5, 103, 2377, 6, 17, 1112, 816, 2782, 13, 5, 103, 10641, 6, 29, 84, 2512, 2430, 782, 18684, 2761, 19, 808, 2430, 2556, 17, 855, 1480, 9477, 4091, 128, 11712, 15, 7103, 2153, 673, 17, 24883, 9990, 9, 3], [2, 11502, 25, 1006, 20, 782, 8, 11809, 855, 1732, 19393, 18667, 37, 367, 21018, 69, 1854, 34, 11860, 19124, 27, 156, 225, 17, 193, 4141, 19, 65, 9124, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [2, 14, 2231, 886, 2385, 17659, 84, 14, 16792, 1952, 9, 3, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], '''token_type_ids''': [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__lowerCAmelCase , model_name="albert-base-v2" , revision="6b6560eaf5ff2e250b00c50f380c5389a9c2d82e" , )
122
"""simple docstring""" import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCAmelCase__ = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. lowerCAmelCase__ = importlib.util.spec_from_file_location( '''transformers''', os.path.join(PATH_TO_TRANSFORMERS, '''__init__.py'''), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) lowerCAmelCase__ = spec.loader.load_module() lowerCAmelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCAmelCase__ = re.compile('''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') lowerCAmelCase__ = { '''CLIPConfigMixin''', '''DecisionTransformerConfigMixin''', '''EncoderDecoderConfigMixin''', '''RagConfigMixin''', '''SpeechEncoderDecoderConfigMixin''', '''VisionEncoderDecoderConfigMixin''', '''VisionTextDualEncoderConfigMixin''', } def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Any = [] for config_class in list(CONFIG_MAPPING.values() ): _lowerCamelCase : Tuple = False # source code of `config_class` _lowerCamelCase : int = inspect.getsource(A_ ) _lowerCamelCase : str = _re_checkpoint.findall(A_ ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` _lowerCamelCase , _lowerCamelCase : Tuple = checkpoint # verify the checkpoint name corresponds to the checkpoint link _lowerCamelCase : Tuple = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: _lowerCamelCase : Union[str, Any] = True break _lowerCamelCase : Tuple = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(A_ ) if len(A_ ) > 0: _lowerCamelCase : Union[str, Any] = '''\n'''.join(sorted(A_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
83
0
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def lowerCamelCase__ ( _lowerCamelCase ): '''simple docstring''' random.seed(A_ ) np.random.seed(A_ ) torch.manual_seed(A_ ) torch.cuda.manual_seed_all(A_ ) # ^^ safe to call this function even if cuda is not available class __UpperCamelCase : def __init__( self ,_A ,_A = 0.9_9_9_9 ,_A = 0.0 ,_A = 0 ,_A = False ,_A = 1.0 ,_A = 2 / 3 ,_A = None ,_A = None ,**_A ,): '''simple docstring''' if isinstance(__lowerCAmelCase ,torch.nn.Module ): _lowerCAmelCase : Dict = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage`' ,'1.0.0' ,__lowerCAmelCase ,standard_warn=__lowerCAmelCase ,) _lowerCAmelCase : int = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility _lowerCAmelCase : Optional[int] = True if kwargs.get('max_value' ,__lowerCAmelCase ) is not None: _lowerCAmelCase : str = '''The `max_value` argument is deprecated. Please use `decay` instead.''' deprecate('max_value' ,'1.0.0' ,__lowerCAmelCase ,standard_warn=__lowerCAmelCase ) _lowerCAmelCase : Any = kwargs['''max_value'''] if kwargs.get('min_value' ,__lowerCAmelCase ) is not None: _lowerCAmelCase : Optional[int] = '''The `min_value` argument is deprecated. Please use `min_decay` instead.''' deprecate('min_value' ,'1.0.0' ,__lowerCAmelCase ,standard_warn=__lowerCAmelCase ) _lowerCAmelCase : Optional[int] = kwargs['''min_value'''] _lowerCAmelCase : int = list(__lowerCAmelCase ) _lowerCAmelCase : int = [p.clone().detach() for p in parameters] if kwargs.get('device' ,__lowerCAmelCase ) is not None: _lowerCAmelCase : Tuple = '''The `device` argument is deprecated. Please use `to` instead.''' deprecate('device' ,'1.0.0' ,__lowerCAmelCase ,standard_warn=__lowerCAmelCase ) self.to(device=kwargs['device'] ) _lowerCAmelCase : Union[str, Any] = None _lowerCAmelCase : Tuple = decay _lowerCAmelCase : Any = min_decay _lowerCAmelCase : str = update_after_step _lowerCAmelCase : Any = use_ema_warmup _lowerCAmelCase : Optional[Any] = inv_gamma _lowerCAmelCase : Tuple = power _lowerCAmelCase : Any = 0 _lowerCAmelCase : Any = None # set in `step()` _lowerCAmelCase : int = model_cls _lowerCAmelCase : Dict = model_config @classmethod def __lowerCamelCase ( cls ,_A ,_A ): '''simple docstring''' _lowerCAmelCase : Optional[int] = model_cls.load_config(__lowerCAmelCase ,return_unused_kwargs=__lowerCAmelCase ) _lowerCAmelCase : int = model_cls.from_pretrained(__lowerCAmelCase ) _lowerCAmelCase : Union[str, Any] = cls(model.parameters() ,model_cls=__lowerCAmelCase ,model_config=model.config ) ema_model.load_state_dict(__lowerCAmelCase ) return ema_model def __lowerCamelCase ( self ,_A ): '''simple docstring''' if self.model_cls is None: raise ValueError('`save_pretrained` can only be used if `model_cls` was defined at __init__.' ) if self.model_config is None: raise ValueError('`save_pretrained` can only be used if `model_config` was defined at __init__.' ) _lowerCAmelCase : Dict = self.model_cls.from_config(self.model_config ) _lowerCAmelCase : int = self.state_dict() state_dict.pop('shadow_params' ,__lowerCAmelCase ) model.register_to_config(**__lowerCAmelCase ) self.copy_to(model.parameters() ) model.save_pretrained(__lowerCAmelCase ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : List[str] = max(0 ,optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: _lowerCAmelCase : List[Any] = 1 - (1 + step / self.inv_gamma) ** -self.power else: _lowerCAmelCase : str = (1 + step) / (10 + step) _lowerCAmelCase : List[Any] = min(__lowerCAmelCase ,self.decay ) # make sure decay is not smaller than min_decay _lowerCAmelCase : str = max(__lowerCAmelCase ,self.min_decay ) return cur_decay_value @torch.no_grad() def __lowerCamelCase ( self ,_A ): '''simple docstring''' if isinstance(__lowerCAmelCase ,torch.nn.Module ): _lowerCAmelCase : int = ( '''Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. ''' '''Please pass the parameters of the module instead.''' ) deprecate( 'passing a `torch.nn.Module` to `ExponentialMovingAverage.step`' ,'1.0.0' ,__lowerCAmelCase ,standard_warn=__lowerCAmelCase ,) _lowerCAmelCase : Dict = parameters.parameters() _lowerCAmelCase : List[Any] = list(__lowerCAmelCase ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. _lowerCAmelCase : Optional[int] = self.get_decay(self.optimization_step ) _lowerCAmelCase : Optional[Any] = decay _lowerCAmelCase : Union[str, Any] = 1 - decay _lowerCAmelCase : Dict = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params ,__lowerCAmelCase ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): _lowerCAmelCase : Optional[int] = deepspeed.zero.GatheredParameters(__lowerCAmelCase ,modifier_rank=__lowerCAmelCase ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(__lowerCAmelCase ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : List[str] = list(__lowerCAmelCase ) for s_param, param in zip(self.shadow_params ,__lowerCAmelCase ): param.data.copy_(s_param.to(param.device ).data ) def __lowerCamelCase ( self ,_A=None ,_A=None ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = [ p.to(device=__lowerCAmelCase ,dtype=__lowerCAmelCase ) if p.is_floating_point() else p.to(device=__lowerCAmelCase ) for p in self.shadow_params ] def __lowerCamelCase ( self ): '''simple docstring''' return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Optional[Any] = [param.detach().cpu().clone() for param in parameters] def __lowerCamelCase ( self ,_A ): '''simple docstring''' if self.temp_stored_params is None: raise RuntimeError('This ExponentialMovingAverage has no `store()`ed weights ' 'to `restore()`' ) for c_param, param in zip(self.temp_stored_params ,__lowerCAmelCase ): param.data.copy_(c_param.data ) # Better memory-wise. _lowerCAmelCase : str = None def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Any = copy.deepcopy(__lowerCAmelCase ) _lowerCAmelCase : str = state_dict.get('decay' ,self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError('Decay must be between 0 and 1' ) _lowerCAmelCase : Any = state_dict.get('min_decay' ,self.min_decay ) if not isinstance(self.min_decay ,__lowerCAmelCase ): raise ValueError('Invalid min_decay' ) _lowerCAmelCase : Union[str, Any] = state_dict.get('optimization_step' ,self.optimization_step ) if not isinstance(self.optimization_step ,__lowerCAmelCase ): raise ValueError('Invalid optimization_step' ) _lowerCAmelCase : Optional[Any] = state_dict.get('update_after_step' ,self.update_after_step ) if not isinstance(self.update_after_step ,__lowerCAmelCase ): raise ValueError('Invalid update_after_step' ) _lowerCAmelCase : Tuple = state_dict.get('use_ema_warmup' ,self.use_ema_warmup ) if not isinstance(self.use_ema_warmup ,__lowerCAmelCase ): raise ValueError('Invalid use_ema_warmup' ) _lowerCAmelCase : int = state_dict.get('inv_gamma' ,self.inv_gamma ) if not isinstance(self.inv_gamma ,(float, int) ): raise ValueError('Invalid inv_gamma' ) _lowerCAmelCase : Tuple = state_dict.get('power' ,self.power ) if not isinstance(self.power ,(float, int) ): raise ValueError('Invalid power' ) _lowerCAmelCase : List[Any] = state_dict.get('shadow_params' ,__lowerCAmelCase ) if shadow_params is not None: _lowerCAmelCase : Optional[Any] = shadow_params if not isinstance(self.shadow_params ,__lowerCAmelCase ): raise ValueError('shadow_params must be a list' ) if not all(isinstance(__lowerCAmelCase ,torch.Tensor ) for p in self.shadow_params ): raise ValueError('shadow_params must all be Tensors' )
259
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCAmelCase__ = False class __snake_case ( unittest.TestCase): pass @nightly @require_torch_gpu class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : int = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : str = '''A painting of a squirrel eating a burger ''' _lowerCamelCase : Dict = torch.manual_seed(0 ) _lowerCamelCase : List[Any] = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__lowerCAmelCase ) _lowerCamelCase : Tuple = VersatileDiffusionTextToImagePipeline.from_pretrained(__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : int = generator.manual_seed(0 ) _lowerCamelCase : List[str] = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Optional[int] = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = '''A painting of a squirrel eating a burger ''' _lowerCamelCase : Optional[int] = torch.manual_seed(0 ) _lowerCamelCase : int = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='''numpy''' ).images _lowerCamelCase : List[str] = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase : Dict = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
83
0
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() a_ : Optional[int] = logging.get_logger(__name__) def __lowerCAmelCase ( _UpperCamelCase : Any ) -> List[str]: '''simple docstring''' SCREAMING_SNAKE_CASE = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: SCREAMING_SNAKE_CASE = [1_44, 1_92, 2_40] SCREAMING_SNAKE_CASE = [16, 32, 64, 96, 1_28, 1_60, 6_40] elif "mobilevit_xs" in mobilevit_name: SCREAMING_SNAKE_CASE = [96, 1_20, 1_44] SCREAMING_SNAKE_CASE = [16, 32, 48, 64, 80, 96, 3_84] elif "mobilevit_xxs" in mobilevit_name: SCREAMING_SNAKE_CASE = [64, 80, 96] SCREAMING_SNAKE_CASE = [16, 16, 24, 48, 64, 80, 3_20] SCREAMING_SNAKE_CASE = 0.05 SCREAMING_SNAKE_CASE = 2.0 if mobilevit_name.startswith('deeplabv3_' ): SCREAMING_SNAKE_CASE = 5_12 SCREAMING_SNAKE_CASE = 16 SCREAMING_SNAKE_CASE = 21 SCREAMING_SNAKE_CASE = '''pascal-voc-id2label.json''' else: SCREAMING_SNAKE_CASE = 10_00 SCREAMING_SNAKE_CASE = '''imagenet-1k-id2label.json''' SCREAMING_SNAKE_CASE = '''huggingface/label-files''' SCREAMING_SNAKE_CASE = json.load(open(hf_hub_download(A_ , A_ , repo_type='dataset' ) , 'r' ) ) SCREAMING_SNAKE_CASE = {int(A_ ): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE = idalabel SCREAMING_SNAKE_CASE = {v: k for k, v in idalabel.items()} return config def __lowerCAmelCase ( _UpperCamelCase : Tuple , _UpperCamelCase : Union[str, Any]=False ) -> Any: '''simple docstring''' for i in range(1 , 6 ): if f"""layer_{i}.""" in name: SCREAMING_SNAKE_CASE = name.replace(f"""layer_{i}.""" , f"""encoder.layer.{i - 1}.""" ) if "conv_1." in name: SCREAMING_SNAKE_CASE = name.replace('conv_1.' , 'conv_stem.' ) if ".block." in name: SCREAMING_SNAKE_CASE = name.replace('.block.' , '.' ) if "exp_1x1" in name: SCREAMING_SNAKE_CASE = name.replace('exp_1x1' , 'expand_1x1' ) if "red_1x1" in name: SCREAMING_SNAKE_CASE = name.replace('red_1x1' , 'reduce_1x1' ) if ".local_rep.conv_3x3." in name: SCREAMING_SNAKE_CASE = name.replace('.local_rep.conv_3x3.' , '.conv_kxk.' ) if ".local_rep.conv_1x1." in name: SCREAMING_SNAKE_CASE = name.replace('.local_rep.conv_1x1.' , '.conv_1x1.' ) if ".norm." in name: SCREAMING_SNAKE_CASE = name.replace('.norm.' , '.normalization.' ) if ".conv." in name: SCREAMING_SNAKE_CASE = name.replace('.conv.' , '.convolution.' ) if ".conv_proj." in name: SCREAMING_SNAKE_CASE = name.replace('.conv_proj.' , '.conv_projection.' ) for i in range(0 , 2 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: SCREAMING_SNAKE_CASE = name.replace(f""".{i}.{j}.""" , f""".{i}.layer.{j}.""" ) for i in range(2 , 6 ): for j in range(0 , 4 ): if f""".{i}.{j}.""" in name: SCREAMING_SNAKE_CASE = name.replace(f""".{i}.{j}.""" , f""".{i}.""" ) if "expand_1x1" in name: SCREAMING_SNAKE_CASE = name.replace('expand_1x1' , 'downsampling_layer.expand_1x1' ) if "conv_3x3" in name: SCREAMING_SNAKE_CASE = name.replace('conv_3x3' , 'downsampling_layer.conv_3x3' ) if "reduce_1x1" in name: SCREAMING_SNAKE_CASE = name.replace('reduce_1x1' , 'downsampling_layer.reduce_1x1' ) for i in range(2 , 5 ): if f""".global_rep.{i}.weight""" in name: SCREAMING_SNAKE_CASE = name.replace(f""".global_rep.{i}.weight""" , '.layernorm.weight' ) if f""".global_rep.{i}.bias""" in name: SCREAMING_SNAKE_CASE = name.replace(f""".global_rep.{i}.bias""" , '.layernorm.bias' ) if ".global_rep." in name: SCREAMING_SNAKE_CASE = name.replace('.global_rep.' , '.transformer.' ) if ".pre_norm_mha.0." in name: SCREAMING_SNAKE_CASE = name.replace('.pre_norm_mha.0.' , '.layernorm_before.' ) if ".pre_norm_mha.1.out_proj." in name: SCREAMING_SNAKE_CASE = name.replace('.pre_norm_mha.1.out_proj.' , '.attention.output.dense.' ) if ".pre_norm_ffn.0." in name: SCREAMING_SNAKE_CASE = name.replace('.pre_norm_ffn.0.' , '.layernorm_after.' ) if ".pre_norm_ffn.1." in name: SCREAMING_SNAKE_CASE = name.replace('.pre_norm_ffn.1.' , '.intermediate.dense.' ) if ".pre_norm_ffn.4." in name: SCREAMING_SNAKE_CASE = name.replace('.pre_norm_ffn.4.' , '.output.dense.' ) if ".transformer." in name: SCREAMING_SNAKE_CASE = name.replace('.transformer.' , '.transformer.layer.' ) if ".aspp_layer." in name: SCREAMING_SNAKE_CASE = name.replace('.aspp_layer.' , '.' ) if ".aspp_pool." in name: SCREAMING_SNAKE_CASE = name.replace('.aspp_pool.' , '.' ) if "seg_head." in name: SCREAMING_SNAKE_CASE = name.replace('seg_head.' , 'segmentation_head.' ) if "segmentation_head.classifier.classifier." in name: SCREAMING_SNAKE_CASE = name.replace('segmentation_head.classifier.classifier.' , 'segmentation_head.classifier.' ) if "classifier.fc." in name: SCREAMING_SNAKE_CASE = name.replace('classifier.fc.' , 'classifier.' ) elif (not base_model) and ("segmentation_head." not in name): SCREAMING_SNAKE_CASE = '''mobilevit.''' + name return name def __lowerCAmelCase ( _UpperCamelCase : Union[str, Any] , _UpperCamelCase : str , _UpperCamelCase : Union[str, Any]=False ) -> List[Any]: '''simple docstring''' if base_model: SCREAMING_SNAKE_CASE = '''''' else: SCREAMING_SNAKE_CASE = '''mobilevit.''' for key in orig_state_dict.copy().keys(): SCREAMING_SNAKE_CASE = orig_state_dict.pop(A_ ) if key[:8] == "encoder.": SCREAMING_SNAKE_CASE = key[8:] if "qkv" in key: SCREAMING_SNAKE_CASE = key.split('.' ) SCREAMING_SNAKE_CASE = int(key_split[0][6:] ) - 1 SCREAMING_SNAKE_CASE = int(key_split[3] ) SCREAMING_SNAKE_CASE = model.get_submodule(f"""{model_prefix}encoder.layer.{layer_num}""" ) SCREAMING_SNAKE_CASE = layer.transformer.layer[transformer_num].attention.attention.all_head_size SCREAMING_SNAKE_CASE = ( f"""{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.""" ) if "weight" in key: SCREAMING_SNAKE_CASE = val[:dim, :] SCREAMING_SNAKE_CASE = val[dim : dim * 2, :] SCREAMING_SNAKE_CASE = val[-dim:, :] else: SCREAMING_SNAKE_CASE = val[:dim] SCREAMING_SNAKE_CASE = val[dim : dim * 2] SCREAMING_SNAKE_CASE = val[-dim:] else: SCREAMING_SNAKE_CASE = val return orig_state_dict def __lowerCAmelCase ( ) -> Optional[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = '''http://images.cocodataset.org/val2017/000000039769.jpg''' SCREAMING_SNAKE_CASE = Image.open(requests.get(A_ , stream=A_ ).raw ) return im @torch.no_grad() def __lowerCAmelCase ( _UpperCamelCase : Any , _UpperCamelCase : Tuple , _UpperCamelCase : Tuple , _UpperCamelCase : Any=False ) -> Optional[int]: '''simple docstring''' SCREAMING_SNAKE_CASE = get_mobilevit_config(A_ ) # load original state_dict SCREAMING_SNAKE_CASE = torch.load(A_ , map_location='cpu' ) # load 🤗 model if mobilevit_name.startswith('deeplabv3_' ): SCREAMING_SNAKE_CASE = MobileViTForSemanticSegmentation(A_ ).eval() else: SCREAMING_SNAKE_CASE = MobileViTForImageClassification(A_ ).eval() SCREAMING_SNAKE_CASE = convert_state_dict(A_ , A_ ) model.load_state_dict(A_ ) # Check outputs on an image, prepared by MobileViTImageProcessor SCREAMING_SNAKE_CASE = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) SCREAMING_SNAKE_CASE = image_processor(images=prepare_img() , return_tensors='pt' ) SCREAMING_SNAKE_CASE = model(**A_ ) SCREAMING_SNAKE_CASE = outputs.logits if mobilevit_name.startswith('deeplabv3_' ): assert logits.shape == (1, 21, 32, 32) if mobilevit_name == "deeplabv3_mobilevit_s": SCREAMING_SNAKE_CASE = torch.tensor( [ [[6.20_65, 6.12_92, 6.20_70], [6.10_79, 6.12_54, 6.17_47], [6.00_42, 6.10_71, 6.10_34]], [[-6.92_53, -6.86_53, -7.03_98], [-7.32_18, -7.39_83, -7.36_70], [-7.19_61, -7.24_82, -7.15_69]], [[-4.47_23, -4.43_48, -4.37_69], [-5.36_29, -5.46_32, -5.45_98], [-5.15_87, -5.34_02, -5.50_59]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": SCREAMING_SNAKE_CASE = torch.tensor( [ [[5.44_49, 5.57_33, 5.63_14], [5.18_15, 5.39_30, 5.59_63], [5.16_56, 5.43_33, 5.48_53]], [[-9.44_23, -9.77_66, -9.67_14], [-9.15_81, -9.57_20, -9.55_19], [-9.10_06, -9.64_58, -9.57_03]], [[-7.77_21, -7.37_16, -7.15_83], [-8.45_99, -8.06_24, -7.79_44], [-8.41_72, -7.83_66, -7.50_25]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": SCREAMING_SNAKE_CASE = torch.tensor( [ [[6.98_11, 6.97_43, 7.31_23], [7.17_77, 7.19_31, 7.39_38], [7.56_33, 7.80_50, 7.89_01]], [[-10.55_36, -10.23_32, -10.29_24], [-10.23_36, -9.86_24, -9.59_64], [-10.88_40, -10.81_58, -10.66_59]], [[-3.49_38, -3.06_31, -2.86_20], [-3.42_05, -2.81_35, -2.68_75], [-3.41_79, -2.79_45, -2.87_50]], ] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3, :3, :3] , A_ , atol=1e-4 ) else: assert logits.shape == (1, 10_00) if mobilevit_name == "mobilevit_s": SCREAMING_SNAKE_CASE = torch.tensor([-0.98_66, 0.23_92, -1.12_41] ) elif mobilevit_name == "mobilevit_xs": SCREAMING_SNAKE_CASE = torch.tensor([-2.47_61, -0.93_99, -1.95_87] ) elif mobilevit_name == "mobilevit_xxs": SCREAMING_SNAKE_CASE = torch.tensor([-1.93_64, -1.23_27, -0.46_53] ) else: raise ValueError(f"""Unknown mobilevit_name: {mobilevit_name}""" ) assert torch.allclose(logits[0, :3] , A_ , atol=1e-4 ) Path(A_ ).mkdir(exist_ok=A_ ) print(f"""Saving model {mobilevit_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(A_ ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(A_ ) if push_to_hub: SCREAMING_SNAKE_CASE = { '''mobilevit_s''': '''mobilevit-small''', '''mobilevit_xs''': '''mobilevit-x-small''', '''mobilevit_xxs''': '''mobilevit-xx-small''', '''deeplabv3_mobilevit_s''': '''deeplabv3-mobilevit-small''', '''deeplabv3_mobilevit_xs''': '''deeplabv3-mobilevit-x-small''', '''deeplabv3_mobilevit_xxs''': '''deeplabv3-mobilevit-xx-small''', } print('Pushing to the hub...' ) SCREAMING_SNAKE_CASE = model_mapping[mobilevit_name] image_processor.push_to_hub(A_ , organization='apple' ) model.push_to_hub(A_ , organization='apple' ) if __name__ == "__main__": a_ : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--mobilevit_name", default="mobilevit_s", type=str, help=( "Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\'," " \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'." ), ) parser.add_argument( "--checkpoint_path", required=True, type=str, help="Path to the original state dict (.pt file)." ) parser.add_argument( "--pytorch_dump_folder_path", required=True, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) a_ : List[str] = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
439
"""simple docstring""" import os import time import numpy as np import onnxruntime as ort lowerCAmelCase__ = '''1''' lowerCAmelCase__ = '''0''' lowerCAmelCase__ = '''1''' lowerCAmelCase__ = ort.SessionOptions() lowerCAmelCase__ = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print('''Create inference session...''') lowerCAmelCase__ = ['''TensorrtExecutionProvider''', '''CUDAExecutionProvider'''] lowerCAmelCase__ = ort.InferenceSession('''model.onnx''', sess_options=sess_opt, providers=execution_provider) lowerCAmelCase__ = ort.RunOptions() lowerCAmelCase__ = 128 lowerCAmelCase__ = 1 lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) print('''Warm up phase...''') sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Start inference...''') lowerCAmelCase__ = time.time() lowerCAmelCase__ = 2000 lowerCAmelCase__ = {} for iter in range(max_iters): lowerCAmelCase__ = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Average Inference Time = {:.3f} ms'''.format((time.time() - start_time) * 1000 / max_iters))
83
0
import tempfile import unittest from make_student import create_student_by_copying_alternating_layers from transformers import AutoConfig from transformers.file_utils import cached_property from transformers.testing_utils import require_torch _snake_case = '''sshleifer/bart-tiny-random''' _snake_case = '''patrickvonplaten/t5-tiny-random''' @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' @cached_property def _snake_case ( self ): """simple docstring""" return AutoConfig.from_pretrained(__lowerCAmelCase ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[Any] = create_student_by_copying_alternating_layers(__lowerCAmelCase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.num_hidden_layers , 1 ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Union[str, Any] = create_student_by_copying_alternating_layers(__lowerCAmelCase , tempfile.mkdtemp() , e=1 , d=__lowerCAmelCase ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = create_student_by_copying_alternating_layers(__lowerCAmelCase , tempfile.mkdtemp() , e=1 , d=__lowerCAmelCase ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , self.teacher_config.encoder_layers ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : str = create_student_by_copying_alternating_layers(__lowerCAmelCase , tempfile.mkdtemp() , e=1 , d=1 ) self.assertEqual(student.config.encoder_layers , 1 ) self.assertEqual(student.config.decoder_layers , 1 ) def _snake_case ( self ): """simple docstring""" with self.assertRaises(__lowerCAmelCase ): create_student_by_copying_alternating_layers(__lowerCAmelCase , tempfile.mkdtemp() , e=__lowerCAmelCase , d=__lowerCAmelCase )
340
"""simple docstring""" import colorsys from PIL import Image # type: ignore def snake_case_ ( A_ : float, A_ : float, A_ : int ): '''simple docstring''' _lowerCamelCase : List[Any] = x _lowerCamelCase : List[Any] = y for step in range(A_ ): # noqa: B007 _lowerCamelCase : Dict = a * a - b * b + x _lowerCamelCase : List[str] = 2 * a * b + y _lowerCamelCase : Any = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case_ ( A_ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def snake_case_ ( A_ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(A_, 1, 1 ) ) def snake_case_ ( A_ : int = 8_00, A_ : int = 6_00, A_ : float = -0.6, A_ : float = 0, A_ : float = 3.2, A_ : int = 50, A_ : bool = True, ): '''simple docstring''' _lowerCamelCase : Tuple = Image.new('''RGB''', (image_width, image_height) ) _lowerCamelCase : int = img.load() # loop through the image-coordinates for image_x in range(A_ ): for image_y in range(A_ ): # determine the figure-coordinates based on the image-coordinates _lowerCamelCase : Optional[Any] = figure_width / image_width * image_height _lowerCamelCase : List[Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width _lowerCamelCase : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height _lowerCamelCase : str = get_distance(A_, A_, A_ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _lowerCamelCase : Dict = get_color_coded_rgb(A_ ) else: _lowerCamelCase : str = get_black_and_white_rgb(A_ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure lowerCAmelCase__ = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
0
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
9
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case_ ( A_ : Tuple, A_ : List[str], A_ : Optional[Any], A_ : Dict, A_ : Dict=True, A_ : int="pt" ): '''simple docstring''' _lowerCamelCase : str = {'''add_prefix_space''': True} if isinstance(A_, A_ ) and not line.startswith(''' ''' ) else {} _lowerCamelCase : Union[str, Any] = padding_side return tokenizer( [line], max_length=A_, padding='''max_length''' if pad_to_max_length else None, truncation=A_, return_tensors=A_, add_special_tokens=A_, **A_, ) def snake_case_ ( A_ : Any, A_ : Optional[int], A_ : List[Any]=None, ): '''simple docstring''' _lowerCamelCase : Optional[int] = input_ids.ne(A_ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __snake_case ( _lowercase): def __init__( self : Tuple , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Tuple="train" , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[int]=None , __lowerCAmelCase : Any=None , __lowerCAmelCase : Union[str, Any]="" , ): """simple docstring""" super().__init__() _lowerCamelCase : Optional[int] = Path(__lowerCAmelCase ).joinpath(type_path + '''.source''' ) _lowerCamelCase : List[str] = Path(__lowerCAmelCase ).joinpath(type_path + '''.target''' ) _lowerCamelCase : List[Any] = self.get_char_lens(self.src_file ) _lowerCamelCase : Optional[int] = max_source_length _lowerCamelCase : Optional[Any] = max_target_length assert min(self.src_lens ) > 0, f'''found empty line in {self.src_file}''' _lowerCamelCase : List[Any] = tokenizer _lowerCamelCase : List[Any] = prefix if n_obs is not None: _lowerCamelCase : List[str] = self.src_lens[:n_obs] _lowerCamelCase : int = src_lang _lowerCamelCase : Union[str, Any] = tgt_lang def __len__( self : int ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : Dict , __lowerCAmelCase : Optional[Any] ): """simple docstring""" _lowerCamelCase : str = index + 1 # linecache starts at 1 _lowerCamelCase : Union[str, Any] = self.prefix + linecache.getline(str(self.src_file ) , __lowerCAmelCase ).rstrip('''\n''' ) _lowerCamelCase : Optional[Any] = linecache.getline(str(self.tgt_file ) , __lowerCAmelCase ).rstrip('''\n''' ) assert source_line, f'''empty source line for index {index}''' assert tgt_line, f'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer , __lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right _lowerCamelCase : Optional[int] = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer ) _lowerCamelCase : Union[str, Any] = self.tokenizer.generator if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_source_length , '''right''' ) _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_target_length , '''right''' ) _lowerCamelCase : Optional[Any] = source_inputs['''input_ids'''].squeeze() _lowerCamelCase : Union[str, Any] = target_inputs['''input_ids'''].squeeze() _lowerCamelCase : Any = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : str ): """simple docstring""" return [len(__lowerCAmelCase ) for x in Path(__lowerCAmelCase ).open().readlines()] def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[Any] = torch.stack([x['''input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = torch.stack([x['''attention_mask'''] for x in batch] ) _lowerCamelCase : Union[str, Any] = torch.stack([x['''decoder_input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Tuple = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Union[str, Any] = trim_batch(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : List[str] = trim_batch(__lowerCAmelCase , __lowerCAmelCase , attention_mask=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch lowerCAmelCase__ = getLogger(__name__) def snake_case_ ( A_ : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(A_ ) ) def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : Dict = get_git_info() save_json(A_, os.path.join(A_, '''git_log.json''' ) ) def snake_case_ ( A_ : str, A_ : Union[str, Any], A_ : int=4, **A_ : Optional[int] ): '''simple docstring''' with open(A_, '''w''' ) as f: json.dump(A_, A_, indent=A_, **A_ ) def snake_case_ ( A_ : Any ): '''simple docstring''' with open(A_ ) as f: return json.load(A_ ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = git.Repo(search_parent_directories=A_ ) _lowerCamelCase : str = { '''repo_id''': str(A_ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case_ ( A_ : Callable, A_ : Iterable ): '''simple docstring''' return list(map(A_, A_ ) ) def snake_case_ ( A_ : str, A_ : Tuple ): '''simple docstring''' with open(A_, '''wb''' ) as f: return pickle.dump(A_, A_ ) def snake_case_ ( A_ : List[str] ): '''simple docstring''' def remove_articles(A_ : str ): return re.sub(R'''\b(a|an|the)\b''', ''' ''', A_ ) def white_space_fix(A_ : Any ): return " ".join(text.split() ) def remove_punc(A_ : List[Any] ): _lowerCamelCase : Any = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(A_ : Optional[int] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(A_ ) ) ) ) def snake_case_ ( A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : str = normalize_answer(A_ ).split() _lowerCamelCase : int = normalize_answer(A_ ).split() _lowerCamelCase : str = Counter(A_ ) & Counter(A_ ) _lowerCamelCase : Any = sum(common.values() ) if num_same == 0: return 0 _lowerCamelCase : int = 1.0 * num_same / len(A_ ) _lowerCamelCase : str = 1.0 * num_same / len(A_ ) _lowerCamelCase : List[Any] = (2 * precision * recall) / (precision + recall) return fa def snake_case_ ( A_ : Dict, A_ : str ): '''simple docstring''' return normalize_answer(A_ ) == normalize_answer(A_ ) def snake_case_ ( A_ : List[str], A_ : List[str] ): '''simple docstring''' assert len(A_ ) == len(A_ ) _lowerCamelCase : Optional[Any] = 0 for hypo, pred in zip(A_, A_ ): em += exact_match_score(A_, A_ ) if len(A_ ) > 0: em /= len(A_ ) return {"em": em} def snake_case_ ( A_ : Optional[int] ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def snake_case_ ( A_ : Dict, A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : Dict = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead _lowerCamelCase : Tuple = '''dropout_rate''' for p in extra_params: if getattr(A_, A_, A_ ): if not hasattr(A_, A_ ) and not hasattr(A_, equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(A_ ) ) delattr(A_, A_ ) continue _lowerCamelCase : Union[str, Any] = p if hasattr(A_, A_ ) else equivalent_param[p] setattr(A_, A_, getattr(A_, A_ ) ) delattr(A_, A_ ) return hparams, config
83
0
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase = 100_0000 ) -> List[str]: snake_case__ = set(range(3 , A_ , 2 ) ) primes.add(2 ) for p in range(3 , A_ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A_ , A_ ) ) ) snake_case__ = [float(A_ ) for n in range(limit + 1 )] for p in primes: for n in range(A_ , limit + 1 , A_ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F"""{solution() = }""")
33
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/config.json''', '''umberto-commoncrawl-cased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json''' ), '''umberto-wikipedia-uncased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json''' ), } class __snake_case ( _lowercase): snake_case__ : Optional[Any] = "camembert" def __init__( self : Optional[Any] , __lowerCAmelCase : Any=3_0_5_2_2 , __lowerCAmelCase : List[str]=7_6_8 , __lowerCAmelCase : List[str]=1_2 , __lowerCAmelCase : Optional[int]=1_2 , __lowerCAmelCase : List[Any]=3_0_7_2 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Union[str, Any]=0.1 , __lowerCAmelCase : Optional[int]=0.1 , __lowerCAmelCase : Optional[int]=5_1_2 , __lowerCAmelCase : str=2 , __lowerCAmelCase : int=0.02 , __lowerCAmelCase : List[Any]=1E-12 , __lowerCAmelCase : Union[str, Any]=1 , __lowerCAmelCase : Optional[Any]=0 , __lowerCAmelCase : List[Any]=2 , __lowerCAmelCase : str="absolute" , __lowerCAmelCase : Any=True , __lowerCAmelCase : Optional[int]=None , **__lowerCAmelCase : Optional[int] , ): """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) _lowerCamelCase : Tuple = vocab_size _lowerCamelCase : str = hidden_size _lowerCamelCase : Union[str, Any] = num_hidden_layers _lowerCamelCase : Any = num_attention_heads _lowerCamelCase : Optional[Any] = hidden_act _lowerCamelCase : List[str] = intermediate_size _lowerCamelCase : Optional[Any] = hidden_dropout_prob _lowerCamelCase : List[Any] = attention_probs_dropout_prob _lowerCamelCase : Optional[Any] = max_position_embeddings _lowerCamelCase : Tuple = type_vocab_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Dict = layer_norm_eps _lowerCamelCase : List[Any] = position_embedding_type _lowerCamelCase : int = use_cache _lowerCamelCase : List[str] = classifier_dropout class __snake_case ( _lowercase): @property def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" if self.task == "multiple-choice": _lowerCamelCase : Optional[Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase : Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
83
0
import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class snake_case_ ( nn.Module ): __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : float = 0.0 __lowerCamelCase : int = 1 __lowerCamelCase : int = 1 __lowerCamelCase : bool = True __lowerCamelCase : bool = False __lowerCamelCase : bool = False __lowerCamelCase : bool = False __lowerCamelCase : jnp.dtype = jnp.floataa def __A ( self ): SCREAMING_SNAKE_CASE_ : List[str] = [] SCREAMING_SNAKE_CASE_ : Optional[Any] = [] for i in range(self.num_layers ): SCREAMING_SNAKE_CASE_ : Optional[int] = self.in_channels if i == 0 else self.out_channels SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxResnetBlockaD( in_channels=__lowerCAmelCase , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : List[str] = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] = resnets SCREAMING_SNAKE_CASE_ : List[Any] = attentions if self.add_downsample: SCREAMING_SNAKE_CASE_ : str = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True ): SCREAMING_SNAKE_CASE_ : List[Any] = () for resnet, attn in zip(self.resnets , self.attentions ): SCREAMING_SNAKE_CASE_ : Optional[int] = resnet(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = attn(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) output_states += (hidden_states,) if self.add_downsample: SCREAMING_SNAKE_CASE_ : Optional[Any] = self.downsamplers_a(__lowerCAmelCase ) output_states += (hidden_states,) return hidden_states, output_states class snake_case_ ( nn.Module ): __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : float = 0.0 __lowerCamelCase : int = 1 __lowerCamelCase : bool = True __lowerCamelCase : jnp.dtype = jnp.floataa def __A ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] = [] for i in range(self.num_layers ): SCREAMING_SNAKE_CASE_ : Any = self.in_channels if i == 0 else self.out_channels SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxResnetBlockaD( in_channels=__lowerCAmelCase , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Dict = resnets if self.add_downsample: SCREAMING_SNAKE_CASE_ : Tuple = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True ): SCREAMING_SNAKE_CASE_ : List[str] = () for resnet in self.resnets: SCREAMING_SNAKE_CASE_ : int = resnet(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) output_states += (hidden_states,) if self.add_downsample: SCREAMING_SNAKE_CASE_ : Union[str, Any] = self.downsamplers_a(__lowerCAmelCase ) output_states += (hidden_states,) return hidden_states, output_states class snake_case_ ( nn.Module ): __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : float = 0.0 __lowerCamelCase : int = 1 __lowerCamelCase : int = 1 __lowerCamelCase : bool = True __lowerCamelCase : bool = False __lowerCamelCase : bool = False __lowerCamelCase : bool = False __lowerCamelCase : jnp.dtype = jnp.floataa def __A ( self ): SCREAMING_SNAKE_CASE_ : Optional[int] = [] SCREAMING_SNAKE_CASE_ : List[str] = [] for i in range(self.num_layers ): SCREAMING_SNAKE_CASE_ : List[str] = self.in_channels if (i == self.num_layers - 1) else self.out_channels SCREAMING_SNAKE_CASE_ : str = self.prev_output_channel if i == 0 else self.out_channels SCREAMING_SNAKE_CASE_ : Any = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Dict = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Optional[Any] = resnets SCREAMING_SNAKE_CASE_ : Optional[Any] = attentions if self.add_upsample: SCREAMING_SNAKE_CASE_ : str = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states SCREAMING_SNAKE_CASE_ : List[Any] = res_hidden_states_tuple[-1] SCREAMING_SNAKE_CASE_ : Any = res_hidden_states_tuple[:-1] SCREAMING_SNAKE_CASE_ : Union[str, Any] = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) SCREAMING_SNAKE_CASE_ : List[str] = resnet(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Tuple = attn(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) if self.add_upsample: SCREAMING_SNAKE_CASE_ : List[Any] = self.upsamplers_a(__lowerCAmelCase ) return hidden_states class snake_case_ ( nn.Module ): __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : int __lowerCamelCase : float = 0.0 __lowerCamelCase : int = 1 __lowerCamelCase : bool = True __lowerCamelCase : jnp.dtype = jnp.floataa def __A ( self ): SCREAMING_SNAKE_CASE_ : Dict = [] for i in range(self.num_layers ): SCREAMING_SNAKE_CASE_ : Optional[int] = self.in_channels if (i == self.num_layers - 1) else self.out_channels SCREAMING_SNAKE_CASE_ : Optional[int] = self.prev_output_channel if i == 0 else self.out_channels SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Tuple = resnets if self.add_upsample: SCREAMING_SNAKE_CASE_ : str = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True ): for resnet in self.resnets: # pop res hidden states SCREAMING_SNAKE_CASE_ : Tuple = res_hidden_states_tuple[-1] SCREAMING_SNAKE_CASE_ : Union[str, Any] = res_hidden_states_tuple[:-1] SCREAMING_SNAKE_CASE_ : List[Any] = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) SCREAMING_SNAKE_CASE_ : List[Any] = resnet(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) if self.add_upsample: SCREAMING_SNAKE_CASE_ : int = self.upsamplers_a(__lowerCAmelCase ) return hidden_states class snake_case_ ( nn.Module ): __lowerCamelCase : int __lowerCamelCase : float = 0.0 __lowerCamelCase : int = 1 __lowerCamelCase : int = 1 __lowerCamelCase : bool = False __lowerCamelCase : bool = False __lowerCamelCase : jnp.dtype = jnp.floataa def __A ( self ): SCREAMING_SNAKE_CASE_ : List[Any] = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] SCREAMING_SNAKE_CASE_ : Dict = [] for _ in range(self.num_layers ): SCREAMING_SNAKE_CASE_ : int = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Union[str, Any] = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Any = resnets SCREAMING_SNAKE_CASE_ : Tuple = attentions def __call__( self , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=True ): SCREAMING_SNAKE_CASE_ : int = self.resnets[0](__lowerCAmelCase , __lowerCAmelCase ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): SCREAMING_SNAKE_CASE_ : Any = attn(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : Tuple = resnet(__lowerCAmelCase , __lowerCAmelCase , deterministic=__lowerCAmelCase ) return hidden_states
345
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase__ = '''<<<<<<< This should probably be modified because it mentions: ''' lowerCAmelCase__ = '''======= >>>>>>> ''' lowerCAmelCase__ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] lowerCAmelCase__ = [ # (pattern, replacement) # Order is important here for some replacements (R'''tfds\.core''', R'''datasets'''), (R'''tf\.io\.gfile\.GFile''', R'''open'''), (R'''tf\.([\w\d]+)''', R'''datasets.Value(\'\1\')'''), (R'''tfds\.features\.Text\(\)''', R'''datasets.Value(\'string\')'''), (R'''tfds\.features\.Text\(''', R'''datasets.Value(\'string\'),'''), (R'''features\s*=\s*tfds.features.FeaturesDict\(''', R'''features=datasets.Features('''), (R'''tfds\.features\.FeaturesDict\(''', R'''dict('''), (R'''The TensorFlow Datasets Authors''', R'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (R'''tfds\.''', R'''datasets.'''), (R'''dl_manager\.manual_dir''', R'''self.config.data_dir'''), (R'''self\.builder_config''', R'''self.config'''), ] def snake_case_ ( A_ : Namespace ): '''simple docstring''' return ConvertCommand(args.tfds_path, args.datasets_directory ) class __snake_case ( _lowercase): @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : ArgumentParser ): """simple docstring""" _lowerCamelCase : List[str] = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=__lowerCAmelCase ) def __init__( self : str , __lowerCAmelCase : str , __lowerCAmelCase : str , *__lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[str] = get_logger('''datasets-cli/converting''' ) _lowerCamelCase : int = tfds_path _lowerCamelCase : Dict = datasets_directory def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" if os.path.isdir(self._tfds_path ): _lowerCamelCase : Union[str, Any] = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): _lowerCamelCase : Dict = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) _lowerCamelCase : int = os.path.abspath(self._datasets_directory ) self._logger.info(f'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) _lowerCamelCase : str = [] _lowerCamelCase : Union[str, Any] = [] _lowerCamelCase : Union[str, Any] = {} if os.path.isdir(self._tfds_path ): _lowerCamelCase : List[str] = os.listdir(__lowerCAmelCase ) else: _lowerCamelCase : Optional[Any] = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f'''Looking at file {f_name}''' ) _lowerCamelCase : Union[str, Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : Optional[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) if not os.path.isfile(__lowerCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(__lowerCAmelCase , encoding='''utf-8''' ) as f: _lowerCamelCase : Tuple = f.readlines() _lowerCamelCase : Optional[int] = [] _lowerCamelCase : Union[str, Any] = False _lowerCamelCase : int = False _lowerCamelCase : Tuple = [] for line in lines: _lowerCamelCase : Optional[int] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: _lowerCamelCase : Union[str, Any] = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here _lowerCamelCase : List[str] = '''''' continue elif "from absl import logging" in out_line: _lowerCamelCase : str = '''from datasets import logging\n''' elif "getLogger" in out_line: _lowerCamelCase : Union[str, Any] = out_line.replace('''getLogger''' , '''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): _lowerCamelCase : Dict = True _lowerCamelCase : Optional[int] = list(filter(lambda __lowerCAmelCase : e in out_line , __lowerCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowerCAmelCase ) + '''\n''' ) out_lines.append(__lowerCAmelCase ) out_lines.append(__lowerCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: _lowerCamelCase : str = re.sub(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: _lowerCamelCase : Dict = re.match(R'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , __lowerCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) _lowerCamelCase : Union[str, Any] = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: _lowerCamelCase : Any = True out_lines.append(__lowerCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset _lowerCamelCase : Union[str, Any] = f_name.replace('''.py''' , '''''' ) _lowerCamelCase : List[str] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) self._logger.info(f'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowerCAmelCase ) if needs_manual_update: with_manual_update.append(__lowerCAmelCase ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.writelines(__lowerCAmelCase ) self._logger.info(f'''Converted in {output_file}''' ) for utils_file in utils_files: try: _lowerCamelCase : Optional[int] = os.path.basename(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = imports_to_builder_map[f_name.replace('''.py''' , '''''' )] self._logger.info(f'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowerCAmelCase , __lowerCAmelCase ) except KeyError: self._logger.error(f'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
83
0
import argparse from pathlib import Path import requests import torch from PIL import Image from transformers import ( RobertaTokenizer, TrOCRConfig, TrOCRForCausalLM, TrOCRProcessor, VisionEncoderDecoderModel, ViTConfig, ViTImageProcessor, ViTModel, ) from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase_ = logging.get_logger(__name__) def _UpperCAmelCase ( A , A ): '''simple docstring''' UpperCAmelCase__ =[] for i in range(encoder_config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (F"""encoder.deit.blocks.{i}.norm1.weight""", F"""encoder.encoder.layer.{i}.layernorm_before.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.norm1.bias""", F"""encoder.encoder.layer.{i}.layernorm_before.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.attn.proj.weight""", F"""encoder.encoder.layer.{i}.attention.output.dense.weight""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.attn.proj.bias""", F"""encoder.encoder.layer.{i}.attention.output.dense.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.norm2.weight""", F"""encoder.encoder.layer.{i}.layernorm_after.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.norm2.bias""", F"""encoder.encoder.layer.{i}.layernorm_after.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc1.weight""", F"""encoder.encoder.layer.{i}.intermediate.dense.weight""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc1.bias""", F"""encoder.encoder.layer.{i}.intermediate.dense.bias""") ) rename_keys.append( (F"""encoder.deit.blocks.{i}.mlp.fc2.weight""", F"""encoder.encoder.layer.{i}.output.dense.weight""") ) rename_keys.append((F"""encoder.deit.blocks.{i}.mlp.fc2.bias""", F"""encoder.encoder.layer.{i}.output.dense.bias""") ) # cls token, position embeddings and patch embeddings of encoder rename_keys.extend( [ ("encoder.deit.cls_token", "encoder.embeddings.cls_token"), ("encoder.deit.pos_embed", "encoder.embeddings.position_embeddings"), ("encoder.deit.patch_embed.proj.weight", "encoder.embeddings.patch_embeddings.projection.weight"), ("encoder.deit.patch_embed.proj.bias", "encoder.embeddings.patch_embeddings.projection.bias"), ("encoder.deit.norm.weight", "encoder.layernorm.weight"), ("encoder.deit.norm.bias", "encoder.layernorm.bias"), ] ) return rename_keys def _UpperCAmelCase ( A , A ): '''simple docstring''' for i in range(encoder_config.num_hidden_layers ): # queries, keys and values (only weights, no biases) UpperCAmelCase__ =state_dict.pop(F"""encoder.deit.blocks.{i}.attn.qkv.weight""" ) UpperCAmelCase__ =in_proj_weight[ : encoder_config.hidden_size, : ] UpperCAmelCase__ =in_proj_weight[ encoder_config.hidden_size : encoder_config.hidden_size * 2, : ] UpperCAmelCase__ =in_proj_weight[ -encoder_config.hidden_size :, : ] def _UpperCAmelCase ( A , A , A ): '''simple docstring''' UpperCAmelCase__ =dct.pop(A_ ) UpperCAmelCase__ =val def _UpperCAmelCase ( A ): '''simple docstring''' if "handwritten" in checkpoint_url: UpperCAmelCase__ ='''https://fki.tic.heia-fr.ch/static/img/a01-122-02-00.jpg''' # industry # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-12.jpg" # have # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02-10.jpg" # let # url = "https://fki.tic.heia-fr.ch/static/img/a01-122-02.jpg" # # url = "https://fki.tic.heia-fr.ch/static/img/a01-122.jpg" elif "printed" in checkpoint_url or "stage1" in checkpoint_url: UpperCAmelCase__ ='''https://www.researchgate.net/profile/Dinh-Sang/publication/338099565/figure/fig8/AS:840413229350922@1577381536857/An-receipt-example-in-the-SROIE-2019-dataset_Q640.jpg''' UpperCAmelCase__ =Image.open(requests.get(A_ , stream=A_ ).raw ).convert("RGB" ) return im @torch.no_grad() def _UpperCAmelCase ( A , A ): '''simple docstring''' UpperCAmelCase__ =ViTConfig(image_size=384 , qkv_bias=A_ ) UpperCAmelCase__ =TrOCRConfig() # size of the architecture if "base" in checkpoint_url: UpperCAmelCase__ =768 elif "large" in checkpoint_url: # use ViT-large encoder UpperCAmelCase__ =1024 UpperCAmelCase__ =4096 UpperCAmelCase__ =24 UpperCAmelCase__ =16 UpperCAmelCase__ =1024 else: raise ValueError("Should either find \'base\' or \'large\' in checkpoint URL" ) # the large-printed + stage1 checkpoints uses sinusoidal position embeddings, no layernorm afterwards if "large-printed" in checkpoint_url or "stage1" in checkpoint_url: UpperCAmelCase__ =False UpperCAmelCase__ ='''relu''' UpperCAmelCase__ =1024 UpperCAmelCase__ =True UpperCAmelCase__ =False UpperCAmelCase__ =False # load HuggingFace model UpperCAmelCase__ =ViTModel(A_ , add_pooling_layer=A_ ) UpperCAmelCase__ =TrOCRForCausalLM(A_ ) UpperCAmelCase__ =VisionEncoderDecoderModel(encoder=A_ , decoder=A_ ) model.eval() # load state_dict of original model, rename some keys UpperCAmelCase__ =torch.hub.load_state_dict_from_url(A_ , map_location="cpu" , check_hash=A_ )['''model'''] UpperCAmelCase__ =create_rename_keys(A_ , A_ ) for src, dest in rename_keys: rename_key(A_ , A_ , A_ ) read_in_q_k_v(A_ , A_ ) # remove parameters we don't need del state_dict["encoder.deit.head.weight"] del state_dict["encoder.deit.head.bias"] del state_dict["decoder.version"] # add prefix to decoder keys for key, val in state_dict.copy().items(): UpperCAmelCase__ =state_dict.pop(A_ ) if key.startswith("decoder" ) and "output_projection" not in key: UpperCAmelCase__ =val else: UpperCAmelCase__ =val # load state dict model.load_state_dict(A_ ) # Check outputs on an image UpperCAmelCase__ =ViTImageProcessor(size=encoder_config.image_size ) UpperCAmelCase__ =RobertaTokenizer.from_pretrained("roberta-large" ) UpperCAmelCase__ =TrOCRProcessor(A_ , A_ ) UpperCAmelCase__ =processor(images=prepare_img(A_ ) , return_tensors="pt" ).pixel_values # verify logits UpperCAmelCase__ =torch.tensor([[model.config.decoder.decoder_start_token_id]] ) UpperCAmelCase__ =model(pixel_values=A_ , decoder_input_ids=A_ ) UpperCAmelCase__ =outputs.logits UpperCAmelCase__ =torch.Size([1, 1, 50265] ) if "trocr-base-handwritten" in checkpoint_url: UpperCAmelCase__ =torch.tensor( [-1.45_02, -4.66_83, -0.53_47, -2.92_91, 9.14_35, -3.05_71, 8.97_64, 1.75_60, 8.73_58, -1.53_11] ) elif "trocr-large-handwritten" in checkpoint_url: UpperCAmelCase__ =torch.tensor( [-2.64_37, -1.31_29, -2.25_96, -5.34_55, 6.35_39, 1.76_04, 5.49_91, 1.47_02, 5.61_13, 2.01_70] ) elif "trocr-base-printed" in checkpoint_url: UpperCAmelCase__ =torch.tensor( [-5.68_16, -5.83_88, 1.13_98, -6.90_34, 6.85_05, -2.43_93, 1.22_84, -1.02_32, -1.96_61, -3.92_10] ) elif "trocr-large-printed" in checkpoint_url: UpperCAmelCase__ =torch.tensor( [-6.01_62, -7.09_59, 4.41_55, -5.10_63, 7.04_68, -3.16_31, 2.64_66, -0.30_81, -0.81_06, -1.75_35] ) if "stage1" not in checkpoint_url: assert logits.shape == expected_shape, "Shape of logits not as expected" assert torch.allclose(logits[0, 0, :10] , A_ , atol=1e-3 ), "First elements of logits not as expected" Path(A_ ).mkdir(exist_ok=A_ ) print(F"""Saving model to {pytorch_dump_folder_path}""" ) model.save_pretrained(A_ ) print(F"""Saving processor to {pytorch_dump_folder_path}""" ) processor.save_pretrained(A_ ) if __name__ == "__main__": UpperCamelCase_ = argparse.ArgumentParser() parser.add_argument( '--checkpoint_url', default='https://layoutlm.blob.core.windows.net/trocr/model_zoo/fairseq/trocr-base-handwritten.pt', type=str, help='URL to the original PyTorch checkpoint (.pth file).', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) UpperCamelCase_ = parser.parse_args() convert_tr_ocr_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path)
625
"""simple docstring""" def snake_case_ ( A_ : list, A_ : list, A_ : int, A_ : int, A_ : int ): '''simple docstring''' if index == number_of_items: return 0 _lowerCamelCase : int = 0 _lowerCamelCase : str = 0 _lowerCamelCase : Dict = knapsack(A_, A_, A_, A_, index + 1 ) if weights[index] <= max_weight: _lowerCamelCase : Tuple = values[index] + knapsack( A_, A_, A_, max_weight - weights[index], index + 1 ) return max(A_, A_ ) if __name__ == "__main__": import doctest doctest.testmod()
83
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase = logging.get_logger(__name__) lowerCamelCase = { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/config.json' ), # See all Speech2Text models at https://huggingface.co/models?filter=speech2text2 } class A ( _lowercase ): UpperCamelCase__ : List[str] ="speech_to_text_2" UpperCamelCase__ : Dict =["past_key_values"] UpperCamelCase__ : Optional[Any] ={"num_attention_heads": "decoder_attention_heads", "hidden_size": "d_model"} def __init__( self : List[str] , lowercase_ : Optional[Any]=1_0000 , lowercase_ : str=6 , lowercase_ : Union[str, Any]=2048 , lowercase_ : List[str]=4 , lowercase_ : List[str]=0.0 , lowercase_ : Any=True , lowercase_ : str="relu" , lowercase_ : Optional[Any]=256 , lowercase_ : Tuple=0.1 , lowercase_ : List[Any]=0.0 , lowercase_ : Any=0.0 , lowercase_ : int=0.02 , lowercase_ : int=2 , lowercase_ : Union[str, Any]=True , lowercase_ : int=1 , lowercase_ : List[str]=0 , lowercase_ : int=2 , lowercase_ : int=1024 , **lowercase_ : List[Any] , ) -> Optional[Any]: """simple docstring""" _lowerCamelCase : int =vocab_size _lowerCamelCase : Dict =d_model _lowerCamelCase : Optional[Any] =decoder_ffn_dim _lowerCamelCase : Any =decoder_layers _lowerCamelCase : int =decoder_attention_heads _lowerCamelCase : Union[str, Any] =dropout _lowerCamelCase : int =attention_dropout _lowerCamelCase : Any =activation_dropout _lowerCamelCase : List[Any] =activation_function _lowerCamelCase : Optional[Any] =init_std _lowerCamelCase : List[Any] =decoder_layerdrop _lowerCamelCase : Any =use_cache _lowerCamelCase : Tuple =decoder_layers _lowerCamelCase : int =scale_embedding # scale factor will be sqrt(d_model) if True _lowerCamelCase : Dict =max_target_positions super().__init__( pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , decoder_start_token_id=__lowerCAmelCase , **__lowerCAmelCase , )
464
"""simple docstring""" from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=_lowercase): snake_case__ : Optional[Any] = ["transformers", "torch", "note_seq"] def __init__( self : Union[str, Any] , *__lowerCAmelCase : Optional[Any] , **__lowerCAmelCase : List[Any] ): """simple docstring""" requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : List[Any] , *__lowerCAmelCase : str , **__lowerCAmelCase : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : int , *__lowerCAmelCase : List[str] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
83
0
'''simple docstring''' __A ={ 'A': '.-', 'B': '-...', 'C': '-.-.', 'D': '-..', 'E': '.', 'F': '..-.', 'G': '--.', 'H': '....', 'I': '..', 'J': '.---', 'K': '-.-', 'L': '.-..', 'M': '--', 'N': '-.', 'O': '---', 'P': '.--.', 'Q': '--.-', 'R': '.-.', 'S': '...', 'T': '-', 'U': '..-', 'V': '...-', 'W': '.--', 'X': '-..-', 'Y': '-.--', 'Z': '--..', '1': '.----', '2': '..---', '3': '...--', '4': '....-', '5': '.....', '6': '-....', '7': '--...', '8': '---..', '9': '----.', '0': '-----', '&': '.-...', '@': '.--.-.', ':': '---...', ',': '--..--', '.': '.-.-.-', '\'': '.----.', '"': '.-..-.', '?': '..--..', '/': '-..-.', '=': '-...-', '+': '.-.-.', '-': '-....-', '(': '-.--.', ')': '-.--.-', '!': '-.-.--', ' ': '/' } # Exclamation mark is not in ITU-R recommendation # fmt: on __A ={value: key for key, value in MORSE_CODE_DICT.items()} def _UpperCamelCase ( UpperCamelCase__ ): return " ".join(MORSE_CODE_DICT[char] for char in message.upper() ) def _UpperCamelCase ( UpperCamelCase__ ): return "".join(REVERSE_DICT[char] for char in message.split() ) def _UpperCamelCase ( ): UpperCAmelCase__ : List[Any] = '''Morse code here!''' print(A_ ) UpperCAmelCase__ : Optional[Any] = encrypt(A_ ) print(A_ ) UpperCAmelCase__ : Optional[Any] = decrypt(A_ ) print(A_ ) if __name__ == "__main__": main()
407
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __snake_case ( unittest.TestCase): @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : Any = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Any = AutoTokenizer.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Union[str, Any] = tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids _lowerCamelCase : Optional[int] = tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids _lowerCamelCase : List[Any] = shift_tokens_right(__lowerCAmelCase , model.config.pad_token_id , model.config.decoder_start_token_id ) _lowerCamelCase : int = model(__lowerCAmelCase , decoder_input_ids=__lowerCAmelCase ).logits _lowerCamelCase : Optional[Any] = optax.softmax_cross_entropy(__lowerCAmelCase , onehot(__lowerCAmelCase , logits.shape[-1] ) ).mean() _lowerCamelCase : Dict = -(labels.shape[-1] * loss.item()) _lowerCamelCase : Dict = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
83
0
'''simple docstring''' def _a (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCamelCase =set() # edges = list of graph's edges _UpperCamelCase =get_edges(A_ ) # While there are still elements in edges list, take an arbitrary edge # (from_node, to_node) and add his extremity to chosen_vertices and then # remove all arcs adjacent to the from_node and to_node while edges: _UpperCamelCase =edges.pop() chosen_vertices.add(A_ ) chosen_vertices.add(A_ ) for edge in edges.copy(): if from_node in edge or to_node in edge: edges.discard(A_ ) return chosen_vertices def _a (__SCREAMING_SNAKE_CASE ): """simple docstring""" _UpperCamelCase =set() for from_node, to_nodes in graph.items(): for to_node in to_nodes: edges.add((from_node, to_node) ) return edges if __name__ == "__main__": import doctest doctest.testmod() # graph = {0: [1, 3], 1: [0, 3], 2: [0, 3, 4], 3: [0, 1, 2], 4: [2, 3]} # print(f"Matching vertex cover:\n{matching_min_vertex_cover(graph)}")
404
"""simple docstring""" import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( '''files''', [ ['''full:README.md''', '''dataset_infos.json'''], ['''empty:README.md''', '''dataset_infos.json'''], ['''dataset_infos.json'''], ['''full:README.md'''], ], ) def snake_case_ ( A_ : Dict, A_ : List[str] ): '''simple docstring''' _lowerCamelCase : int = tmp_path_factory.mktemp('''dset_infos_dir''' ) if "full:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''---\ndataset_info:\n dataset_size: 42\n---''' ) if "empty:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''''' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / '''dataset_infos.json''', '''w''' ) as f: f.write('''{"default": {"dataset_size": 42}}''' ) _lowerCamelCase : str = DatasetInfosDict.from_directory(A_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( '''dataset_info''', [ DatasetInfo(), DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ), ], ) def snake_case_ ( A_ : str, A_ : DatasetInfo ): '''simple docstring''' _lowerCamelCase : Optional[Any] = str(A_ ) dataset_info.write_to_directory(A_ ) _lowerCamelCase : str = DatasetInfo.from_directory(A_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(A_, '''dataset_info.json''' ) ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = DatasetInfo( description='''foo''', citation='''bar''', homepage='''https://foo.bar''', license='''CC0''', features=Features({'''a''': Value('''int32''' )} ), post_processed={}, supervised_keys=(), task_templates=[], builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train''', '''num_examples''': 42}], download_checksums={}, download_size=13_37, post_processing_size=4_42, dataset_size=12_34, size_in_bytes=13_37 + 4_42 + 12_34, ) _lowerCamelCase : Optional[Any] = dataset_info._to_yaml_dict() assert sorted(A_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key], (list, dict, int, str) ) _lowerCamelCase : str = yaml.safe_dump(A_ ) _lowerCamelCase : Tuple = yaml.safe_load(A_ ) assert dataset_info_yaml_dict == reloaded def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : int = DatasetInfo() _lowerCamelCase : Dict = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( '''dataset_infos_dict''', [ DatasetInfosDict(), DatasetInfosDict({'''default''': DatasetInfo()} ), DatasetInfosDict({'''my_config_name''': DatasetInfo()} ), DatasetInfosDict( { '''default''': DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ) } ), DatasetInfosDict( { '''v1''': DatasetInfo(dataset_size=42 ), '''v2''': DatasetInfo(dataset_size=13_37 ), } ), ], ) def snake_case_ ( A_ : Optional[Any], A_ : DatasetInfosDict ): '''simple docstring''' _lowerCamelCase : List[str] = str(A_ ) dataset_infos_dict.write_to_directory(A_ ) _lowerCamelCase : List[Any] = DatasetInfosDict.from_directory(A_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): _lowerCamelCase : str = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml _lowerCamelCase : Any = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(A_, '''README.md''' ) )
83
0
from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCamelCase__ = logging.get_logger(__name__) lowerCamelCase__ = { '''microsoft/biogpt''': '''https://huggingface.co/microsoft/biogpt/resolve/main/config.json''', # See all BioGPT models at https://huggingface.co/models?filter=biogpt } class __magic_name__ (_lowercase ): lowerCamelCase__ = "biogpt" def __init__( self , _a=42384 , _a=1024 , _a=24 , _a=16 , _a=4096 , _a="gelu" , _a=0.1 , _a=0.1 , _a=1024 , _a=0.0_2 , _a=1E-12 , _a=True , _a=True , _a=0.0 , _a=0.0 , _a=1 , _a=0 , _a=2 , **_a , ) -> Dict: lowerCAmelCase_ = vocab_size lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_act lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = scale_embedding lowerCAmelCase_ = use_cache lowerCAmelCase_ = layerdrop lowerCAmelCase_ = activation_dropout super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase )
122
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __snake_case : def __init__( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple=1_3 , __lowerCAmelCase : Any=2 , __lowerCAmelCase : List[str]=2_4 , __lowerCAmelCase : str=1_6 , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : Union[str, Any]=True , __lowerCAmelCase : Optional[Any]=3_2 , __lowerCAmelCase : List[Any]=5 , __lowerCAmelCase : int=4 , __lowerCAmelCase : int=3_7 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Dict=0.1 , __lowerCAmelCase : str=0.1 , __lowerCAmelCase : int=1_0 , __lowerCAmelCase : List[Any]=0.02 , __lowerCAmelCase : str=None , __lowerCAmelCase : List[str]=2 , __lowerCAmelCase : Union[str, Any]=2 , ): """simple docstring""" _lowerCamelCase : List[str] = parent _lowerCamelCase : str = batch_size _lowerCamelCase : Tuple = patch_size _lowerCamelCase : Optional[int] = max_length _lowerCamelCase : List[Any] = num_mel_bins _lowerCamelCase : int = is_training _lowerCamelCase : Union[str, Any] = use_labels _lowerCamelCase : Dict = hidden_size _lowerCamelCase : Tuple = num_hidden_layers _lowerCamelCase : int = num_attention_heads _lowerCamelCase : Tuple = intermediate_size _lowerCamelCase : List[str] = hidden_act _lowerCamelCase : Dict = hidden_dropout_prob _lowerCamelCase : int = attention_probs_dropout_prob _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : List[str] = scope _lowerCamelCase : Optional[int] = frequency_stride _lowerCamelCase : List[Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase : Union[str, Any] = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase : Any = frequency_out_dimension * time_out_dimension _lowerCamelCase : List[Any] = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : int = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase : str = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Optional[int] = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : List[Any] = ASTModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : List[Any] = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : int = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Optional[Any] = config_and_inputs _lowerCamelCase : int = {'''input_values''': input_values} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) snake_case__ : Tuple = ( {"audio-classification": ASTForAudioClassification, "feature-extraction": ASTModel} if is_torch_available() else {} ) snake_case__ : Any = False snake_case__ : List[Any] = False snake_case__ : Optional[Any] = False snake_case__ : Optional[Any] = False def SCREAMING_SNAKE_CASE ( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Optional[Any] ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[int] = ASTModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" pass def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Dict = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : str = model_class(__lowerCAmelCase ) _lowerCamelCase : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : Any = [*signature.parameters.keys()] _lowerCamelCase : str = ['''input_values'''] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Union[str, Any] = ASTModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase : str = torchaudio.load(A_ ) return audio, sampling_rate @require_torch @require_torchaudio class __snake_case ( unittest.TestCase): @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : int = self.default_feature_extractor _lowerCamelCase : Union[str, Any] = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(__lowerCAmelCase ) _lowerCamelCase : List[Any] = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase : List[Any] = prepare_audio() _lowerCamelCase : Dict = audio.squeeze().numpy() _lowerCamelCase : Tuple = feature_extractor(__lowerCAmelCase , sampling_rate=__lowerCAmelCase , return_tensors='''pt''' ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): _lowerCamelCase : Tuple = model(**__lowerCAmelCase ) # verify the logits _lowerCamelCase : Tuple = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) _lowerCamelCase : Optional[int] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4 ) )
83
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging _lowerCAmelCase = logging.get_logger(__name__) _lowerCAmelCase = """▁""" _lowerCAmelCase = {"""vocab_file""": """sentencepiece.bpe.model"""} _lowerCAmelCase = { """vocab_file""": { """facebook/mbart-large-50-one-to-many-mmt""": ( """https://huggingface.co/facebook/mbart-large-50-one-to-many-mmt/resolve/main/sentencepiece.bpe.model""" ), } } _lowerCAmelCase = { """facebook/mbart-large-50-one-to-many-mmt""": 1_0_2_4, } # fmt: off _lowerCAmelCase = ["""ar_AR""", """cs_CZ""", """de_DE""", """en_XX""", """es_XX""", """et_EE""", """fi_FI""", """fr_XX""", """gu_IN""", """hi_IN""", """it_IT""", """ja_XX""", """kk_KZ""", """ko_KR""", """lt_LT""", """lv_LV""", """my_MM""", """ne_NP""", """nl_XX""", """ro_RO""", """ru_RU""", """si_LK""", """tr_TR""", """vi_VN""", """zh_CN""", """af_ZA""", """az_AZ""", """bn_IN""", """fa_IR""", """he_IL""", """hr_HR""", """id_ID""", """ka_GE""", """km_KH""", """mk_MK""", """ml_IN""", """mn_MN""", """mr_IN""", """pl_PL""", """ps_AF""", """pt_XX""", """sv_SE""", """sw_KE""", """ta_IN""", """te_IN""", """th_TH""", """tl_XX""", """uk_UA""", """ur_PK""", """xh_ZA""", """gl_ES""", """sl_SI"""] class __UpperCamelCase ( _lowercase ): _UpperCAmelCase = VOCAB_FILES_NAMES _UpperCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCAmelCase = PRETRAINED_VOCAB_FILES_MAP _UpperCAmelCase = ["input_ids", "attention_mask"] _UpperCAmelCase = [] _UpperCAmelCase = [] def __init__( self ,_A ,_A=None ,_A=None ,_A="</s>" ,_A="</s>" ,_A="<s>" ,_A="<unk>" ,_A="<pad>" ,_A="<mask>" ,_A = None ,**_A ,): '''simple docstring''' _lowerCAmelCase : Optional[int] = AddedToken(__lowerCAmelCase ,lstrip=__lowerCAmelCase ,rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase ,__lowerCAmelCase ) else mask_token _lowerCAmelCase : Dict = {} if sp_model_kwargs is None else sp_model_kwargs _lowerCAmelCase : List[Any] = kwargs.get('additional_special_tokens' ,[] ) kwargs["additional_special_tokens"] += [ code for code in FAIRSEQ_LANGUAGE_CODES if code not in kwargs["additional_special_tokens"] ] super().__init__( src_lang=__lowerCAmelCase ,tgt_lang=__lowerCAmelCase ,eos_token=__lowerCAmelCase ,unk_token=__lowerCAmelCase ,sep_token=__lowerCAmelCase ,cls_token=__lowerCAmelCase ,pad_token=__lowerCAmelCase ,mask_token=__lowerCAmelCase ,sp_model_kwargs=self.sp_model_kwargs ,**__lowerCAmelCase ,) _lowerCAmelCase : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__lowerCAmelCase ) ) _lowerCAmelCase : List[str] = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token _lowerCAmelCase : Optional[int] = {'''<s>''': 0, '''<pad>''': 1, '''</s>''': 2, '''<unk>''': 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab _lowerCAmelCase : Tuple = 1 _lowerCAmelCase : List[Any] = len(self.sp_model ) _lowerCAmelCase : Any = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(__lowerCAmelCase ) } _lowerCAmelCase : List[str] = {v: k for k, v in self.lang_code_to_id.items()} _lowerCAmelCase : Tuple = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) _lowerCAmelCase : Any = {v: k for k, v in self.fairseq_tokens_to_ids.items()} _lowerCAmelCase : Tuple = src_lang if src_lang is not None else '''en_XX''' _lowerCAmelCase : Optional[int] = self.lang_code_to_id[self._src_lang] _lowerCAmelCase : List[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) @property def __lowerCamelCase ( self ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def __lowerCamelCase ( self ): '''simple docstring''' return self._src_lang @src_lang.setter def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : List[Any] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self ): '''simple docstring''' _lowerCAmelCase : List[str] = self.__dict__.copy() _lowerCAmelCase : List[Any] = None return state def __setstate__( self ,_A ): '''simple docstring''' _lowerCAmelCase : int = d # for backward compatibility if not hasattr(self ,'sp_model_kwargs' ): _lowerCAmelCase : List[Any] = {} _lowerCAmelCase : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def __lowerCamelCase ( self ): '''simple docstring''' _lowerCAmelCase : str = {self.convert_ids_to_tokens(__lowerCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __lowerCamelCase ( self ,_A ): '''simple docstring''' return self.sp_model.encode(__lowerCAmelCase ,out_type=__lowerCAmelCase ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _lowerCAmelCase : Dict = self.sp_model.PieceToId(__lowerCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def __lowerCamelCase ( self ,_A ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Dict = [] _lowerCAmelCase : Any = '''''' _lowerCAmelCase : str = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__lowerCAmelCase ) + token _lowerCAmelCase : Dict = True _lowerCAmelCase : str = [] else: current_sub_tokens.append(__lowerCAmelCase ) _lowerCAmelCase : Union[str, Any] = False out_string += self.sp_model.decode(__lowerCAmelCase ) return out_string.strip() def __lowerCamelCase ( self ,_A ,_A = None ): '''simple docstring''' if not os.path.isdir(__lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return _lowerCAmelCase : str = os.path.join( __lowerCAmelCase ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__lowerCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,__lowerCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__lowerCAmelCase ,'wb' ) as fi: _lowerCAmelCase : Optional[int] = self.sp_model.serialized_model_proto() fi.write(__lowerCAmelCase ) return (out_vocab_file,) def __lowerCamelCase ( self ,_A ,_A = None ,_A = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__lowerCAmelCase ,token_ids_a=__lowerCAmelCase ,already_has_special_tokens=__lowerCAmelCase ) _lowerCAmelCase : int = [1] * len(self.prefix_tokens ) _lowerCAmelCase : Optional[int] = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__lowerCAmelCase )) + suffix_ones return prefix_ones + ([0] * len(__lowerCAmelCase )) + ([0] * len(__lowerCAmelCase )) + suffix_ones def __lowerCamelCase ( self ,_A ,_A = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def __lowerCamelCase ( self ,_A ,_A ,_A ,_A ,**_A ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError('Translation requires a `src_lang` and a `tgt_lang` for this model' ) _lowerCAmelCase : int = src_lang _lowerCAmelCase : Tuple = self(__lowerCAmelCase ,add_special_tokens=__lowerCAmelCase ,return_tensors=__lowerCAmelCase ,**__lowerCAmelCase ) _lowerCAmelCase : int = self.convert_tokens_to_ids(__lowerCAmelCase ) _lowerCAmelCase : Union[str, Any] = tgt_lang_id return inputs def __lowerCamelCase ( self ,_A ,_A = "en_XX" ,_A = None ,_A = "ro_RO" ,**_A ,): '''simple docstring''' _lowerCAmelCase : Optional[int] = src_lang _lowerCAmelCase : Union[str, Any] = tgt_lang return super().prepare_seqaseq_batch(__lowerCAmelCase ,__lowerCAmelCase ,**__lowerCAmelCase ) def __lowerCamelCase ( self ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def __lowerCamelCase ( self ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : str = self.lang_code_to_id[src_lang] _lowerCAmelCase : Tuple = [self.cur_lang_code_id] _lowerCAmelCase : str = [self.eos_token_id] def __lowerCamelCase ( self ,_A ): '''simple docstring''' _lowerCAmelCase : Optional[int] = self.lang_code_to_id[tgt_lang] _lowerCAmelCase : List[Any] = [self.cur_lang_code_id] _lowerCAmelCase : Any = [self.eos_token_id]
259
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def snake_case_ ( A_ : Dict, A_ : bool = True, A_ : float = math.inf, A_ : float = -math.inf, A_ : float = math.inf, A_ : float = -math.inf, A_ : bool = False, A_ : float = 1_00, A_ : float = 0.01, A_ : float = 1, ): '''simple docstring''' _lowerCamelCase : Optional[int] = False _lowerCamelCase : str = search_prob _lowerCamelCase : str = start_temperate _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : int = 0 _lowerCamelCase : Any = None while not search_end: _lowerCamelCase : Dict = current_state.score() if best_state is None or current_score > best_state.score(): _lowerCamelCase : Tuple = current_state scores.append(A_ ) iterations += 1 _lowerCamelCase : List[Any] = None _lowerCamelCase : Optional[int] = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to _lowerCamelCase : List[Any] = random.randint(0, len(A_ ) - 1 ) # picking a random neighbor _lowerCamelCase : Dict = neighbors.pop(A_ ) _lowerCamelCase : Union[str, Any] = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: _lowerCamelCase : str = change * -1 # in case we are finding minimum if change > 0: # improves the solution _lowerCamelCase : Optional[Any] = picked_neighbor else: _lowerCamelCase : Optional[int] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability _lowerCamelCase : Union[str, Any] = picked_neighbor _lowerCamelCase : List[str] = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor _lowerCamelCase : Tuple = True else: _lowerCamelCase : Optional[Any] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(A_ ), A_ ) plt.xlabel('''Iterations''' ) plt.ylabel('''Function values''' ) plt.show() return best_state if __name__ == "__main__": def snake_case_ ( A_ : int, A_ : Tuple ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def snake_case_ ( A_ : Optional[int], A_ : List[Any] ): '''simple docstring''' return (3 * x**2) - (6 * y) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" ) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" )
83
0
import unittest from transformers import GPTNeoXJapaneseConfig, is_torch_available from transformers.models.gpt_neox_japanese.tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseModel class UpperCamelCase : def __init__( self : Dict , snake_case__ : str , snake_case__ : Tuple=1_3 , snake_case__ : Optional[int]=7 , snake_case__ : Union[str, Any]=True , snake_case__ : List[str]=True , snake_case__ : Tuple=True , snake_case__ : Tuple=True , snake_case__ : Optional[Any]=9_9 , snake_case__ : str=3_2 , snake_case__ : Dict=5 , snake_case__ : int=4 , snake_case__ : Tuple=4 , snake_case__ : List[str]="gelu" , snake_case__ : Dict=0.0 , snake_case__ : Union[str, Any]=0.1 , snake_case__ : str=True , snake_case__ : str=5_1_2 , snake_case__ : str=1_6 , snake_case__ : List[str]=2 , snake_case__ : Optional[Any]=0.02 , snake_case__ : int=3 , snake_case__ : int=4 , snake_case__ : Union[str, Any]=None , ): """simple docstring""" SCREAMING_SNAKE_CASE = parent SCREAMING_SNAKE_CASE = batch_size SCREAMING_SNAKE_CASE = seq_length SCREAMING_SNAKE_CASE = is_training SCREAMING_SNAKE_CASE = use_input_mask SCREAMING_SNAKE_CASE = use_token_type_ids SCREAMING_SNAKE_CASE = use_labels SCREAMING_SNAKE_CASE = vocab_size SCREAMING_SNAKE_CASE = hidden_size SCREAMING_SNAKE_CASE = num_hidden_layers SCREAMING_SNAKE_CASE = num_attention_heads SCREAMING_SNAKE_CASE = intermediate_multiple_size SCREAMING_SNAKE_CASE = hidden_act SCREAMING_SNAKE_CASE = hidden_dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = weight_tying SCREAMING_SNAKE_CASE = max_position_embeddings SCREAMING_SNAKE_CASE = type_vocab_size SCREAMING_SNAKE_CASE = type_sequence_label_size SCREAMING_SNAKE_CASE = initializer_range SCREAMING_SNAKE_CASE = num_labels SCREAMING_SNAKE_CASE = num_choices SCREAMING_SNAKE_CASE = scope def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE = None if self.use_input_mask: SCREAMING_SNAKE_CASE = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE = None if self.use_labels: SCREAMING_SNAKE_CASE = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE = self.get_config() return config, input_ids, input_mask, token_labels def UpperCamelCase ( self : List[str] ): """simple docstring""" return GPTNeoXJapaneseConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_multiple_size=self.intermediate_multiple_size , hidden_act=self.hidden_act , hidden_dropout=self.hidden_dropout , attention_dropout=self.attention_dropout , weight_tying=self.weight_tying , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , ) def UpperCamelCase ( self : Optional[int] ): """simple docstring""" SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE = True return config, input_ids, input_mask, token_labels def UpperCamelCase ( self : List[str] , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : Any ): """simple docstring""" SCREAMING_SNAKE_CASE = GPTNeoXJapaneseModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase ) SCREAMING_SNAKE_CASE = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self : List[str] , snake_case__ : str , snake_case__ : List[Any] , snake_case__ : Any ): """simple docstring""" SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = GPTNeoXJapaneseModel(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCamelCase ( self : Tuple , snake_case__ : str , snake_case__ : str , snake_case__ : List[Any] , snake_case__ : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE = GPTNeoXJapaneseForCausalLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() SCREAMING_SNAKE_CASE = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def UpperCamelCase ( self : Tuple , snake_case__ : Union[str, Any] , snake_case__ : Optional[Any] , snake_case__ : Any ): """simple docstring""" SCREAMING_SNAKE_CASE = True SCREAMING_SNAKE_CASE = GPTNeoXJapaneseForCausalLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() # first forward pass SCREAMING_SNAKE_CASE = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , use_cache=__lowerCAmelCase ) SCREAMING_SNAKE_CASE = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and SCREAMING_SNAKE_CASE = torch.cat([input_ids, next_tokens] , dim=-1 ) SCREAMING_SNAKE_CASE = torch.cat([input_mask, next_mask] , dim=-1 ) SCREAMING_SNAKE_CASE = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , output_hidden_states=__lowerCAmelCase ) SCREAMING_SNAKE_CASE = output_from_no_past['''hidden_states'''][0] SCREAMING_SNAKE_CASE = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , past_key_values=__lowerCAmelCase , output_hidden_states=__lowerCAmelCase , )['''hidden_states'''][0] # select random slice SCREAMING_SNAKE_CASE = ids_tensor((1,) , output_from_past.shape[-1] ).item() SCREAMING_SNAKE_CASE = output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) def UpperCamelCase ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE = self.prepare_config_and_inputs() SCREAMING_SNAKE_CASE = config_and_inputs SCREAMING_SNAKE_CASE = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class UpperCamelCase ( _lowercase , _lowercase , unittest.TestCase ): __UpperCamelCase =(GPTNeoXJapaneseModel, GPTNeoXJapaneseForCausalLM) if is_torch_available() else () __UpperCamelCase =(GPTNeoXJapaneseForCausalLM,) if is_torch_available() else () __UpperCamelCase =( {"feature-extraction": GPTNeoXJapaneseModel, "text-generation": GPTNeoXJapaneseForCausalLM} if is_torch_available() else {} ) __UpperCamelCase =False __UpperCamelCase =False __UpperCamelCase =False __UpperCamelCase =False def UpperCamelCase ( self : Dict ): """simple docstring""" SCREAMING_SNAKE_CASE = GPTNeoXJapaneseModelTester(self ) SCREAMING_SNAKE_CASE = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=3_7 ) def UpperCamelCase ( self : str ): """simple docstring""" self.config_tester.run_common_tests() def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase ( self : Tuple ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs_for_decoder() SCREAMING_SNAKE_CASE = None self.model_tester.create_and_check_model_as_decoder(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def UpperCamelCase ( self : Optional[Any] ): """simple docstring""" SCREAMING_SNAKE_CASE = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*__lowerCAmelCase ) @slow def UpperCamelCase ( self : str ): """simple docstring""" SCREAMING_SNAKE_CASE = '''abeja/gpt-neox-japanese-2.7b''' SCREAMING_SNAKE_CASE = ['''データサイエンティストとは、''', '''100年後に必要とされる会社は、''', '''フルリモートの環境で働くために必要なことは、''', '''国境の長いトンネルを抜けると''', '''美味しい日本食といえば、'''] SCREAMING_SNAKE_CASE = [ '''データサイエンティストとは、データを分析し、ビジネスに役立つ知見を導き出す専門家のことです。''', '''100年後に必要とされる会社は、「人」が中心の会社です。''', '''フルリモートの環境で働くために必要なことは、「自分の時間をコントロールする」ことです。''', '''国境の長いトンネルを抜けると、そこは雪国だった。''', '''美味しい日本食といえば、やっぱりお寿司ですよね。''', ] SCREAMING_SNAKE_CASE = GPTNeoXJapaneseTokenizer.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE = GPTNeoXJapaneseForCausalLM.from_pretrained(__lowerCAmelCase ) SCREAMING_SNAKE_CASE = [] for prompt in prompts: SCREAMING_SNAKE_CASE = tokenizer(__lowerCAmelCase , return_tensors='pt' ).input_ids SCREAMING_SNAKE_CASE = model.generate(__lowerCAmelCase , max_length=5_0 ) SCREAMING_SNAKE_CASE = tokenizer.batch_decode(__lowerCAmelCase , skip_special_tokens=__lowerCAmelCase ) predicted_outputs += generated_string self.assertListEqual(__lowerCAmelCase , __lowerCAmelCase )
439
"""simple docstring""" from collections import namedtuple lowerCAmelCase__ = namedtuple('''from_to''', '''from_ to''') lowerCAmelCase__ = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.0_0_1, 1000), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.0_0_4_5_4, 2_6_4.1_7_2), '''cubicyard''': from_to(0.7_6_4_5_5, 1.3_0_7_9_5), '''cubicfoot''': from_to(0.0_2_8, 3_5.3_1_4_7), '''cup''': from_to(0.0_0_0_2_3_6_5_8_8, 4_2_2_6.7_5), } def snake_case_ ( A_ : float, A_ : str, A_ : str ): '''simple docstring''' if from_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + ''', '''.join(A_ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + ''', '''.join(A_ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
83
0
def lowercase_( ): '''simple docstring''' lowerCamelCase : Optional[int] = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] lowerCamelCase : Tuple = 6 lowerCamelCase : int = 1 lowerCamelCase : Any = 1901 lowerCamelCase : int = 0 while year < 2001: day += 7 if (year % 4 == 0 and year % 100 != 0) or (year % 400 == 0): if day > days_per_month[month - 1] and month != 2: month += 1 lowerCamelCase : str = day - days_per_month[month - 2] elif day > 29 and month == 2: month += 1 lowerCamelCase : Dict = day - 29 else: if day > days_per_month[month - 1]: month += 1 lowerCamelCase : Union[str, Any] = day - days_per_month[month - 2] if month > 12: year += 1 lowerCamelCase : Optional[int] = 1 if year < 2001 and day == 1: sundays += 1 return sundays if __name__ == "__main__": print(solution())
340
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case ( _lowercase): def __init__( self : Dict , __lowerCAmelCase : Any , __lowerCAmelCase : int=1_3 , __lowerCAmelCase : Optional[int]=7 , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : str=True , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : int=True , __lowerCAmelCase : List[str]=9_9 , __lowerCAmelCase : List[Any]=3_2 , __lowerCAmelCase : Union[str, Any]=5 , __lowerCAmelCase : List[str]=4 , __lowerCAmelCase : Optional[int]=3_7 , __lowerCAmelCase : List[Any]="gelu" , __lowerCAmelCase : Tuple=0.1 , __lowerCAmelCase : List[Any]=0.1 , __lowerCAmelCase : int=5_1_2 , __lowerCAmelCase : Tuple=1_6 , __lowerCAmelCase : Tuple=2 , __lowerCAmelCase : Tuple=0.02 , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : Any="None" , __lowerCAmelCase : str=3 , __lowerCAmelCase : Dict=4 , __lowerCAmelCase : Optional[Any]=None , ): """simple docstring""" _lowerCamelCase : Dict = parent _lowerCamelCase : Union[str, Any] = batch_size _lowerCamelCase : Optional[Any] = seq_length _lowerCamelCase : Optional[Any] = is_training _lowerCamelCase : Dict = use_input_mask _lowerCamelCase : Tuple = use_token_type_ids _lowerCamelCase : Optional[Any] = use_labels _lowerCamelCase : List[str] = vocab_size _lowerCamelCase : Any = hidden_size _lowerCamelCase : int = num_hidden_layers _lowerCamelCase : Optional[Any] = num_attention_heads _lowerCamelCase : int = intermediate_size _lowerCamelCase : Optional[int] = hidden_act _lowerCamelCase : int = hidden_dropout_prob _lowerCamelCase : Dict = attention_probs_dropout_prob _lowerCamelCase : List[Any] = max_position_embeddings _lowerCamelCase : str = type_vocab_size _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : List[Any] = initializer_range _lowerCamelCase : Optional[int] = num_labels _lowerCamelCase : Any = num_choices _lowerCamelCase : int = relative_attention _lowerCamelCase : Union[str, Any] = position_biased_input _lowerCamelCase : str = pos_att_type _lowerCamelCase : Tuple = scope def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : List[Any] = None if self.use_input_mask: _lowerCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _lowerCamelCase : Any = None if self.use_token_type_ids: _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : Any = None _lowerCamelCase : int = None _lowerCamelCase : Union[str, Any] = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase : List[str] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : str ): """simple docstring""" self.parent.assertListEqual(list(result.loss.size() ) , [] ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : str = model(__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : List[Any] = model(__lowerCAmelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __lowerCAmelCase : Any , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaForMaskedLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[int] = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = self.num_labels _lowerCamelCase : Dict = DebertaVaForSequenceClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : List[str] , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.num_labels _lowerCamelCase : Tuple = DebertaVaForTokenClassification(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Any = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaForQuestionAnswering(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Tuple = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , start_positions=__lowerCAmelCase , end_positions=__lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : Optional[int] = DebertaVaForMultipleChoice(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Union[str, Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : List[Any] = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Any = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCamelCase : Optional[int] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) snake_case__ : int = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : List[str] = True snake_case__ : List[Any] = False snake_case__ : int = False snake_case__ : Optional[Any] = False snake_case__ : str = False def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class __snake_case ( unittest.TestCase): @unittest.skip(reason='''Model not available yet''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" pass @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained('''microsoft/deberta-v2-xlarge''' ) _lowerCamelCase : List[str] = torch.tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) _lowerCamelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _lowerCamelCase : Tuple = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase )[0] # compare the actual values for a slice. _lowerCamelCase : Union[str, Any] = torch.tensor( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowerCAmelCase , atol=1E-4 ) , f'''{output[:, 1:4, 1:4]}''' )
83
0
import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger SCREAMING_SNAKE_CASE__ = '''<<<<<<< This should probably be modified because it mentions: ''' SCREAMING_SNAKE_CASE__ = '''======= >>>>>>> ''' SCREAMING_SNAKE_CASE__ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] SCREAMING_SNAKE_CASE__ = [ # (pattern, replacement) # Order is important here for some replacements (r'''tfds\.core''', r'''datasets'''), (r'''tf\.io\.gfile\.GFile''', r'''open'''), (r'''tf\.([\w\d]+)''', r'''datasets.Value(\'\1\')'''), (r'''tfds\.features\.Text\(\)''', r'''datasets.Value(\'string\')'''), (r'''tfds\.features\.Text\(''', r'''datasets.Value(\'string\'),'''), (r'''features\s*=\s*tfds.features.FeaturesDict\(''', r'''features=datasets.Features('''), (r'''tfds\.features\.FeaturesDict\(''', r'''dict('''), (r'''The TensorFlow Datasets Authors''', r'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (r'''tfds\.''', r'''datasets.'''), (r'''dl_manager\.manual_dir''', r'''self.config.data_dir'''), (r'''self\.builder_config''', r'''self.config'''), ] def A ( __UpperCamelCase ) -> Optional[Any]: return ConvertCommand(args.tfds_path , args.datasets_directory ) class __lowerCAmelCase ( _lowercase ): """simple docstring""" @staticmethod def _a ( _snake_case : ArgumentParser ): """simple docstring""" A__ = parser.add_parser( 'convert' , help='Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.' , ) train_parser.add_argument( '--tfds_path' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.' , ) train_parser.add_argument( '--datasets_directory' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='Path to the HuggingFace Datasets folder.' ) train_parser.set_defaults(func=__lowerCAmelCase ) def __init__( self : str , _snake_case : str , _snake_case : str , *_snake_case : int ): """simple docstring""" A__ = get_logger('datasets-cli/converting' ) A__ = tfds_path A__ = datasets_directory def _a ( self : Dict ): """simple docstring""" if os.path.isdir(self._tfds_path ): A__ = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): A__ = os.path.dirname(self._tfds_path ) else: raise ValueError('--tfds_path is neither a directory nor a file. Please check path.' ) A__ = os.path.abspath(self._datasets_directory ) self._logger.info(F'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) A__ = [] A__ = [] A__ = {} if os.path.isdir(self._tfds_path ): A__ = os.listdir(__lowerCAmelCase ) else: A__ = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(F'''Looking at file {f_name}''' ) A__ = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) A__ = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) if not os.path.isfile(__lowerCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('Skipping file' ) continue with open(__lowerCAmelCase , encoding='utf-8' ) as f: A__ = f.readlines() A__ = [] A__ = False A__ = False A__ = [] for line in lines: A__ = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: A__ = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here A__ = '''''' continue elif "from absl import logging" in out_line: A__ = '''from datasets import logging\n''' elif "getLogger" in out_line: A__ = out_line.replace('getLogger' , 'get_logger' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): A__ = True A__ = list(filter(lambda _snake_case : e in out_line , __lowerCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowerCAmelCase ) + '\n' ) out_lines.append(__lowerCAmelCase ) out_lines.append(__lowerCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: A__ = re.sub(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: A__ = re.match(R'from\stensorflow_datasets.*import\s([^\.\r\n]+)' , __lowerCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(',' ) ) A__ = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(F'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: A__ = True out_lines.append(__lowerCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset A__ = f_name.replace('.py' , '' ) A__ = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) A__ = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) self._logger.info(F'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowerCAmelCase ) if needs_manual_update: with_manual_update.append(__lowerCAmelCase ) with open(__lowerCAmelCase , 'w' , encoding='utf-8' ) as f: f.writelines(__lowerCAmelCase ) self._logger.info(F'''Converted in {output_file}''' ) for utils_file in utils_files: try: A__ = os.path.basename(__lowerCAmelCase ) A__ = imports_to_builder_map[f_name.replace('.py' , '' )] self._logger.info(F'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowerCAmelCase , __lowerCAmelCase ) except KeyError: self._logger.error(F'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( F'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
9
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCAmelCase__ = '''hf-internal-testing/tiny-random-bert''' lowerCAmelCase__ = os.path.join(TRANSFORMERS_CACHE, '''models--hf-internal-testing--tiny-random-bert''') lowerCAmelCase__ = '''9b8c223d42b2188cb49d29af482996f9d0f3e5a6''' class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = cached_file(__lowerCAmelCase , __lowerCAmelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(__lowerCAmelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) ) ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : Optional[int] = f.read() self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) self.assertTrue(os.path.isfile(__lowerCAmelCase ) ) # File is cached at the same place the second time. _lowerCamelCase : Tuple = cached_file(__lowerCAmelCase , __lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) # Using a specific revision to test the full commit hash. _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''9b8c223''' ) self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): _lowerCamelCase : Optional[int] = cached_file('''tiny-random-bert''' , __lowerCAmelCase ) with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): _lowerCamelCase : str = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''aaaa''' ) with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : int = cached_file(__lowerCAmelCase , '''conf''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , '''conf''' ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : List[Any] = f.read() self.assertTrue(os.path.isfile(os.path.join(__lowerCAmelCase , '''.no_exist''' , __lowerCAmelCase , '''conf''' ) ) ) _lowerCamelCase : str = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Optional[int] = cached_file(__lowerCAmelCase , '''conf''' , local_files_only=__lowerCAmelCase , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Any = mock.Mock() _lowerCamelCase : Optional[Any] = 5_0_0 _lowerCamelCase : Dict = {} _lowerCamelCase : List[Any] = HTTPError _lowerCamelCase : int = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=__lowerCAmelCase ) as mock_head: _lowerCamelCase : Union[str, Any] = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_connection_errors=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , __lowerCAmelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase , revision='''ahaha''' ) _lowerCamelCase : Dict = get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. _lowerCamelCase : Dict = json.loads(open(__lowerCAmelCase , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_6_8 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase : Any = Path(__lowerCAmelCase ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(__lowerCAmelCase , '''a.txt''' ) , str(__lowerCAmelCase ) ) self.assertIsNone(get_file_from_repo(__lowerCAmelCase , '''b.txt''' ) )
83
0
import json import os from typing import Dict, List, Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCamelCase__ : Optional[int] = logging.get_logger(__name__) lowerCamelCase__ : str = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", """tokenizer_config_file""": """tokenizer_config.json""", } lowerCamelCase__ : str = { """vocab_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json""" }, """merges_file""": { """facebook/blenderbot_small-90M""": """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt""" }, """tokenizer_config_file""": { """facebook/blenderbot_small-90M""": ( """https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json""" ) }, } lowerCamelCase__ : Union[str, Any] = {"""facebook/blenderbot_small-90M""": 5_1_2} def SCREAMING_SNAKE_CASE ( __lowerCAmelCase ) -> Tuple: snake_case__ = set() snake_case__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) snake_case__ = char snake_case__ = set(A_ ) return pairs class __magic_name__ (_lowercase ): '''simple docstring''' __lowercase : Optional[Any] = VOCAB_FILES_NAMES __lowercase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __lowercase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self:int , _a:List[str] , _a:List[str] , _a:Any="__start__" , _a:Any="__end__" , _a:Any="__unk__" , _a:Tuple="__null__" , **_a:int , ): super().__init__(unk_token=__lowerCAmelCase , bos_token=__lowerCAmelCase , eos_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , **__lowerCAmelCase ) with open(__lowerCAmelCase , encoding='''utf-8''' ) as vocab_handle: snake_case__ = json.load(__lowerCAmelCase ) snake_case__ = {v: k for k, v in self.encoder.items()} with open(__lowerCAmelCase , encoding='''utf-8''' ) as merges_handle: snake_case__ = merges_handle.read().split('''\n''' )[1:-1] snake_case__ = [tuple(merge.split() ) for merge in merges] snake_case__ = dict(zip(__lowerCAmelCase , range(len(__lowerCAmelCase ) ) ) ) snake_case__ = {} @property def SCREAMING_SNAKE_CASE__ ( self:str ): return len(self.encoder ) def SCREAMING_SNAKE_CASE__ ( self:Tuple ): return dict(self.encoder , **self.added_tokens_encoder ) def SCREAMING_SNAKE_CASE__ ( self:Tuple , _a:str ): if token in self.cache: return self.cache[token] snake_case__ = re.sub('''([.,!?()])''' , r''' \1''' , __lowerCAmelCase ) snake_case__ = re.sub('''(\')''' , r''' \1 ''' , __lowerCAmelCase ) snake_case__ = re.sub(r'''\s{2,}''' , ''' ''' , __lowerCAmelCase ) if "\n" in token: snake_case__ = token.replace('''\n''' , ''' __newln__''' ) snake_case__ = token.split(''' ''' ) snake_case__ = [] for token in tokens: if not len(__lowerCAmelCase ): continue snake_case__ = token.lower() snake_case__ = tuple(__lowerCAmelCase ) snake_case__ = tuple(list(word[:-1] ) + [word[-1] + '''</w>'''] ) snake_case__ = get_pairs(__lowerCAmelCase ) if not pairs: words.append(__lowerCAmelCase ) continue while True: snake_case__ = min(__lowerCAmelCase , key=lambda _a : self.bpe_ranks.get(__lowerCAmelCase , float('''inf''' ) ) ) if bigram not in self.bpe_ranks: break snake_case__ = bigram snake_case__ = [] snake_case__ = 0 while i < len(__lowerCAmelCase ): try: snake_case__ = word.index(__lowerCAmelCase , __lowerCAmelCase ) new_word.extend(word[i:j] ) snake_case__ = j except ValueError: new_word.extend(word[i:] ) break if word[i] == first and i < len(__lowerCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 snake_case__ = tuple(__lowerCAmelCase ) snake_case__ = new_word if len(__lowerCAmelCase ) == 1: break else: snake_case__ = get_pairs(__lowerCAmelCase ) snake_case__ = '''@@ '''.join(__lowerCAmelCase ) snake_case__ = word[:-4] snake_case__ = word words.append(__lowerCAmelCase ) return " ".join(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE__ ( self:int , _a:str ): snake_case__ = [] snake_case__ = re.findall(r'''\S+\n?''' , __lowerCAmelCase ) for token in words: split_tokens.extend(list(self.bpe(__lowerCAmelCase ).split(''' ''' ) ) ) return split_tokens def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:str ): snake_case__ = token.lower() return self.encoder.get(__lowerCAmelCase , self.encoder.get(self.unk_token ) ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:int ): return self.decoder.get(__lowerCAmelCase , self.unk_token ) def SCREAMING_SNAKE_CASE__ ( self:Optional[Any] , _a:List[str] ): snake_case__ = ''' '''.join(__lowerCAmelCase ).replace('''@@ ''' , '''''' ).strip() return out_string def SCREAMING_SNAKE_CASE__ ( self:List[str] , _a:str , _a:Optional[str] = None ): if not os.path.isdir(__lowerCAmelCase ): logger.error(F"""Vocabulary path ({save_directory}) should be a directory""" ) return snake_case__ = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) snake_case__ = os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''merges_file'''] ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__lowerCAmelCase , ensure_ascii=__lowerCAmelCase ) + '''\n''' ) snake_case__ = 0 with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as writer: writer.write('''#version: 0.2\n''' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda _a : kv[1] ): if index != token_index: logger.warning( F"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ''' Please check that the tokenizer is not corrupted!''' ) snake_case__ = token_index writer.write(''' '''.join(__lowerCAmelCase ) + '''\n''' ) index += 1 return vocab_file, merge_file
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class __snake_case ( _lowercase): snake_case__ : List[str] = "cvt" def __init__( self : Any , __lowerCAmelCase : Union[str, Any]=3 , __lowerCAmelCase : List[str]=[7, 3, 3] , __lowerCAmelCase : int=[4, 2, 2] , __lowerCAmelCase : int=[2, 1, 1] , __lowerCAmelCase : str=[6_4, 1_9_2, 3_8_4] , __lowerCAmelCase : Dict=[1, 3, 6] , __lowerCAmelCase : Optional[Any]=[1, 2, 1_0] , __lowerCAmelCase : Dict=[4.0, 4.0, 4.0] , __lowerCAmelCase : Dict=[0.0, 0.0, 0.0] , __lowerCAmelCase : Union[str, Any]=[0.0, 0.0, 0.0] , __lowerCAmelCase : int=[0.0, 0.0, 0.1] , __lowerCAmelCase : Union[str, Any]=[True, True, True] , __lowerCAmelCase : str=[False, False, True] , __lowerCAmelCase : List[str]=["dw_bn", "dw_bn", "dw_bn"] , __lowerCAmelCase : List[Any]=[3, 3, 3] , __lowerCAmelCase : Dict=[1, 1, 1] , __lowerCAmelCase : str=[2, 2, 2] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : str=0.02 , __lowerCAmelCase : Any=1E-12 , **__lowerCAmelCase : int , ): """simple docstring""" super().__init__(**__lowerCAmelCase ) _lowerCamelCase : Optional[int] = num_channels _lowerCamelCase : int = patch_sizes _lowerCamelCase : Optional[Any] = patch_stride _lowerCamelCase : str = patch_padding _lowerCamelCase : Any = embed_dim _lowerCamelCase : Optional[Any] = num_heads _lowerCamelCase : Dict = depth _lowerCamelCase : Optional[int] = mlp_ratio _lowerCamelCase : Any = attention_drop_rate _lowerCamelCase : Any = drop_rate _lowerCamelCase : Dict = drop_path_rate _lowerCamelCase : Optional[int] = qkv_bias _lowerCamelCase : int = cls_token _lowerCamelCase : int = qkv_projection_method _lowerCamelCase : Optional[Any] = kernel_qkv _lowerCamelCase : List[str] = padding_kv _lowerCamelCase : Tuple = stride_kv _lowerCamelCase : Union[str, Any] = padding_q _lowerCamelCase : Optional[Any] = stride_q _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Optional[int] = layer_norm_eps
83
0
from typing import List, Optional, TypeVar from .arrow_dataset import Dataset, _concatenate_map_style_datasets, _interleave_map_style_datasets from .dataset_dict import DatasetDict, IterableDatasetDict from .info import DatasetInfo from .iterable_dataset import IterableDataset, _concatenate_iterable_datasets, _interleave_iterable_datasets from .splits import NamedSplit from .utils import logging from .utils.py_utils import Literal lowerCAmelCase__: str = logging.get_logger(__name__) lowerCAmelCase__: List[Any] = TypeVar("DatasetType", Dataset, IterableDataset) def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = "first_exhausted" , ) -> Dict: from .arrow_dataset import Dataset from .iterable_dataset import IterableDataset if not datasets: raise ValueError('Unable to interleave an empty list of datasets.' ) for i, dataset in enumerate(A_ ): if not isinstance(A_ , (Dataset, IterableDataset) ): if isinstance(A_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ' 'is an empty dataset dictionary.' ) raise ValueError( f'Dataset at position {i} has at least one split: {list(A_ )}\n' f'Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(A_ ) )}\']' ) raise ValueError( f'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(A_ ).__name__}.' ) if i == 0: SCREAMING_SNAKE_CASE_ : Any = ( (Dataset, IterableDataset) if isinstance(A_ , A_ ) else (IterableDataset, Dataset) ) elif not isinstance(A_ , A_ ): raise ValueError( f'Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.' ) if stopping_strategy not in ["first_exhausted", "all_exhausted"]: raise ValueError(f'{stopping_strategy} is not supported. Please enter a valid stopping_strategy.' ) if dataset_type is Dataset: return _interleave_map_style_datasets( A_ , A_ , A_ , info=A_ , split=A_ , stopping_strategy=A_ ) else: return _interleave_iterable_datasets( A_ , A_ , A_ , info=A_ , split=A_ , stopping_strategy=A_ ) def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = 0 , ) -> List[str]: if not dsets: raise ValueError('Unable to concatenate an empty list of datasets.' ) for i, dataset in enumerate(A_ ): if not isinstance(A_ , (Dataset, IterableDataset) ): if isinstance(A_ , (DatasetDict, IterableDatasetDict) ): if not dataset: raise ValueError( f'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} ' 'is an empty dataset dictionary.' ) raise ValueError( f'Dataset at position {i} has at least one split: {list(A_ )}\n' f'Please pick one to interleave with the other datasets, for example: dataset[\'{next(iter(A_ ) )}\']' ) raise ValueError( f'Expected a list of Dataset objects or a list of IterableDataset objects, but element at position {i} is a {type(A_ ).__name__}.' ) if i == 0: SCREAMING_SNAKE_CASE_ : str = ( (Dataset, IterableDataset) if isinstance(A_ , A_ ) else (IterableDataset, Dataset) ) elif not isinstance(A_ , A_ ): raise ValueError( f'Unable to interleave a {dataset_type.__name__} (at position 0) with a {other_type.__name__} (at position {i}). Expected a list of Dataset objects or a list of IterableDataset objects.' ) if dataset_type is Dataset: return _concatenate_map_style_datasets(A_ , info=A_ , split=A_ , axis=A_ ) else: return _concatenate_iterable_datasets(A_ , info=A_ , split=A_ , axis=A_ )
345
"""simple docstring""" # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class __snake_case ( _lowercase): snake_case__ : torch.FloatTensor snake_case__ : torch.FloatTensor class __snake_case ( _lowercase , _lowercase): snake_case__ : int = 1 @register_to_config def __init__( self : str , __lowerCAmelCase : int = 2_0_0_0 , __lowerCAmelCase : float = 0.15 , __lowerCAmelCase : float = 0.01 , __lowerCAmelCase : float = 13_48.0 , __lowerCAmelCase : float = 1E-5 , __lowerCAmelCase : int = 1 , ): """simple docstring""" _lowerCamelCase : Optional[int] = sigma_max # setable values _lowerCamelCase : Dict = None self.set_sigmas(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[int] = None ): """simple docstring""" return sample def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : Union[str, torch.device] = None ): """simple docstring""" _lowerCamelCase : Tuple = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCamelCase : Optional[int] = torch.linspace(1 , __lowerCAmelCase , __lowerCAmelCase , device=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None ): """simple docstring""" _lowerCamelCase : List[str] = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCamelCase : int = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCamelCase : Any = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCamelCase : Optional[int] = torch.exp(torch.linspace(math.log(__lowerCAmelCase ) , math.log(__lowerCAmelCase ) , __lowerCAmelCase ) ) _lowerCamelCase : Tuple = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def SCREAMING_SNAKE_CASE ( self : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[Any] ): """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : int , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) _lowerCamelCase : Tuple = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCamelCase : Dict = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCamelCase : Optional[int] = timesteps.to(self.discrete_sigmas.device ) _lowerCamelCase : Any = self.discrete_sigmas[timesteps].to(sample.device ) _lowerCamelCase : int = self.get_adjacent_sigma(__lowerCAmelCase , __lowerCAmelCase ).to(sample.device ) _lowerCamelCase : Any = torch.zeros_like(__lowerCAmelCase ) _lowerCamelCase : Any = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCamelCase : Union[str, Any] = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _lowerCamelCase : List[Any] = diffusion.unsqueeze(-1 ) _lowerCamelCase : int = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCamelCase : List[str] = randn_tensor( sample.shape , layout=sample.layout , generator=__lowerCAmelCase , device=sample.device , dtype=sample.dtype ) _lowerCamelCase : List[Any] = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCamelCase : int = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=__lowerCAmelCase , prev_sample_mean=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCamelCase : Union[str, Any] = randn_tensor(sample.shape , layout=sample.layout , generator=__lowerCAmelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _lowerCamelCase : Union[str, Any] = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : Tuple = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : str = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCamelCase : Tuple = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCamelCase : Union[str, Any] = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _lowerCamelCase : str = step_size.unsqueeze(-1 ) _lowerCamelCase : Any = sample + step_size * model_output _lowerCamelCase : int = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , ): """simple docstring""" _lowerCamelCase : Dict = timesteps.to(original_samples.device ) _lowerCamelCase : Union[str, Any] = self.discrete_sigmas.to(original_samples.device )[timesteps] _lowerCamelCase : Union[str, Any] = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(__lowerCAmelCase ) * sigmas[:, None, None, None] ) _lowerCamelCase : int = noise + original_samples return noisy_samples def __len__( self : Optional[int] ): """simple docstring""" return self.config.num_train_timesteps
83
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) UpperCamelCase_ = { 'configuration_mobilevit': ['MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileViTConfig', 'MobileViTOnnxConfig'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = ['MobileViTFeatureExtractor'] UpperCamelCase_ = ['MobileViTImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ 'MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileViTForImageClassification', 'MobileViTForSemanticSegmentation', 'MobileViTModel', 'MobileViTPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase_ = [ 'TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFMobileViTForImageClassification', 'TFMobileViTForSemanticSegmentation', 'TFMobileViTModel', 'TFMobileViTPreTrainedModel', ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys UpperCamelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
625
"""simple docstring""" from torch import nn def snake_case_ ( A_ : int ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'''Unsupported activation function: {act_fn}''' )
83
0
lowerCamelCase = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def a_ ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' _lowerCamelCase : List[str] =[False] * len(A_ ) _lowerCamelCase : Tuple =[s] _lowerCamelCase : Any =True while queue: _lowerCamelCase : List[Any] =queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(A_ ) _lowerCamelCase : Optional[int] =True _lowerCamelCase : Optional[int] =u return visited[t] def a_ ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Any ): '''simple docstring''' _lowerCamelCase : Any =[-1] * (len(A_ )) _lowerCamelCase : Any =0 _lowerCamelCase : Optional[int] =[] _lowerCamelCase : Tuple =[i[:] for i in graph] # Record original cut, copy. while bfs(A_ , A_ , A_ , A_ ): _lowerCamelCase : int =float('Inf' ) _lowerCamelCase : List[str] =sink while s != source: # Find the minimum value in select path _lowerCamelCase : List[Any] =min(A_ , graph[parent[s]][s] ) _lowerCamelCase : List[str] =parent[s] max_flow += path_flow _lowerCamelCase : Optional[Any] =sink while v != source: _lowerCamelCase : Tuple =parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow _lowerCamelCase : Optional[int] =parent[v] for i in range(len(A_ ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
464
"""simple docstring""" def snake_case_ ( A_ : int, A_ : int ): '''simple docstring''' return int(input_a == input_a == 0 ) def snake_case_ ( ): '''simple docstring''' print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(F'''| 0 | 0 | {nor_gate(0, 0 )} |''' ) print(F'''| 0 | 1 | {nor_gate(0, 1 )} |''' ) print(F'''| 1 | 0 | {nor_gate(1, 0 )} |''' ) print(F'''| 1 | 1 | {nor_gate(1, 1 )} |''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
83
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available __A ={ 'configuration_squeezebert': [ 'SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'SqueezeBertConfig', 'SqueezeBertOnnxConfig', ], 'tokenization_squeezebert': ['SqueezeBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =['SqueezeBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ 'SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'SqueezeBertForMaskedLM', 'SqueezeBertForMultipleChoice', 'SqueezeBertForQuestionAnswering', 'SqueezeBertForSequenceClassification', 'SqueezeBertForTokenClassification', 'SqueezeBertModel', 'SqueezeBertModule', 'SqueezeBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys __A =_LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
407
"""simple docstring""" from __future__ import annotations def snake_case_ ( A_ : list[list[int]] ): '''simple docstring''' for i in range(1, len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1, len(A_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1, len(A_ ) ): for j in range(1, len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j], matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
83
0
'''simple docstring''' import json import os import re import unicodedata from json.encoder import INFINITY from typing import Any, Dict, List, Optional, Tuple, Union import numpy as np import regex from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_flax_available, is_tf_available, is_torch_available, logging from ...utils.generic import _is_jax, _is_numpy __lowerCamelCase : Optional[int] = logging.get_logger(__name__) __lowerCamelCase : List[str] = { 'artists_file': 'artists.json', 'lyrics_file': 'lyrics.json', 'genres_file': 'genres.json', } __lowerCamelCase : Union[str, Any] = { 'artists_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/artists.json', }, 'genres_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/genres.json', }, 'lyrics_file': { 'jukebox': 'https://huggingface.co/ArthurZ/jukebox/blob/main/lyrics.json', }, } __lowerCamelCase : Dict = { 'jukebox': 512, } class UpperCAmelCase ( _lowercase): """simple docstring""" lowerCAmelCase_ = VOCAB_FILES_NAMES lowerCAmelCase_ = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase_ = PRETRAINED_LYRIC_TOKENS_SIZES lowerCAmelCase_ = ["input_ids", "attention_mask"] def __init__( self : str , UpperCamelCase__ : int , UpperCamelCase__ : Tuple , UpperCamelCase__ : str , UpperCamelCase__ : Optional[int]=["v3", "v2", "v2"] , UpperCamelCase__ : Any=512 , UpperCamelCase__ : Tuple=5 , UpperCamelCase__ : List[str]="<|endoftext|>" , **UpperCamelCase__ : Any , ) -> Union[str, Any]: _UpperCamelCase =AddedToken(__lowerCAmelCase , lstrip=__lowerCAmelCase , rstrip=__lowerCAmelCase ) if isinstance(__lowerCAmelCase , __lowerCAmelCase ) else unk_token super().__init__( unk_token=__lowerCAmelCase , n_genres=__lowerCAmelCase , version=__lowerCAmelCase , max_n_lyric_tokens=__lowerCAmelCase , **__lowerCAmelCase , ) _UpperCamelCase =version _UpperCamelCase =max_n_lyric_tokens _UpperCamelCase =n_genres with open(__lowerCAmelCase , encoding='''utf-8''' ) as vocab_handle: _UpperCamelCase =json.load(__lowerCAmelCase ) with open(__lowerCAmelCase , encoding='''utf-8''' ) as vocab_handle: _UpperCamelCase =json.load(__lowerCAmelCase ) with open(__lowerCAmelCase , encoding='''utf-8''' ) as vocab_handle: _UpperCamelCase =json.load(__lowerCAmelCase ) _UpperCamelCase =R'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''' # In v2, we had a n_vocab=80 and in v3 we missed + and so n_vocab=79 of characters. if len(self.lyrics_encoder ) == 79: _UpperCamelCase =oov.replace(R'''\-\'''' , R'''\-+\'''' ) _UpperCamelCase =regex.compile(__lowerCAmelCase ) _UpperCamelCase ={v: k for k, v in self.artists_encoder.items()} _UpperCamelCase ={v: k for k, v in self.genres_encoder.items()} _UpperCamelCase ={v: k for k, v in self.lyrics_encoder.items()} @property def UpperCamelCase__ ( self : List[Any] ) -> Union[str, Any]: return len(self.artists_encoder ) + len(self.genres_encoder ) + len(self.lyrics_encoder ) def UpperCamelCase__ ( self : Tuple ) -> Tuple: return dict(self.artists_encoder , self.genres_encoder , self.lyrics_encoder ) def UpperCamelCase__ ( self : Tuple , UpperCamelCase__ : str , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Union[str, Any] ) -> Optional[Any]: _UpperCamelCase =[self.artists_encoder.get(__lowerCAmelCase , 0 ) for artist in list_artists] for genres in range(len(__lowerCAmelCase ) ): _UpperCamelCase =[self.genres_encoder.get(__lowerCAmelCase , 0 ) for genre in list_genres[genres]] _UpperCamelCase =list_genres[genres] + [-1] * (self.n_genres - len(list_genres[genres] )) _UpperCamelCase =[[self.lyrics_encoder.get(__lowerCAmelCase , 0 ) for character in list_lyrics[0]], [], []] return artists_id, list_genres, lyric_ids def UpperCamelCase__ ( self : str , UpperCamelCase__ : Union[str, Any] ) -> int: return list(__lowerCAmelCase ) def UpperCamelCase__ ( self : Tuple , UpperCamelCase__ : List[str] , UpperCamelCase__ : Any , UpperCamelCase__ : Tuple , **UpperCamelCase__ : Union[str, Any] ) -> Optional[int]: _UpperCamelCase =self.prepare_for_tokenization(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =self._tokenize(__lowerCAmelCase ) return artist, genre, lyrics def UpperCamelCase__ ( self : Optional[int] , UpperCamelCase__ : str , UpperCamelCase__ : str , UpperCamelCase__ : str , UpperCamelCase__ : bool = False ) -> str: for idx in range(len(self.version ) ): if self.version[idx] == "v3": _UpperCamelCase =artists[idx].lower() _UpperCamelCase =[genres[idx].lower()] else: _UpperCamelCase =self._normalize(artists[idx] ) + '''.v2''' _UpperCamelCase =[ self._normalize(__lowerCAmelCase ) + '''.v2''' for genre in genres[idx].split('''_''' ) ] # split is for the full dictionary with combined genres if self.version[0] == "v2": _UpperCamelCase =regex.compile(R'''[^A-Za-z0-9.,:;!?\-\'\"()\[\] \t\n]+''' ) _UpperCamelCase ='''ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789.,:;!?-+\'\"()[] \t\n''' _UpperCamelCase ={vocab[index]: index + 1 for index in range(len(__lowerCAmelCase ) )} _UpperCamelCase =0 _UpperCamelCase =len(__lowerCAmelCase ) + 1 _UpperCamelCase =self.vocab _UpperCamelCase ={v: k for k, v in self.vocab.items()} _UpperCamelCase ='''''' else: _UpperCamelCase =regex.compile(R'''[^A-Za-z0-9.,:;!?\-+\'\"()\[\] \t\n]+''' ) _UpperCamelCase =self._run_strip_accents(__lowerCAmelCase ) _UpperCamelCase =lyrics.replace('''\\''' , '''\n''' ) _UpperCamelCase =self.out_of_vocab.sub('''''' , __lowerCAmelCase ), [], [] return artists, genres, lyrics def UpperCamelCase__ ( self : int , UpperCamelCase__ : Dict ) -> int: _UpperCamelCase =unicodedata.normalize('''NFD''' , __lowerCAmelCase ) _UpperCamelCase =[] for char in text: _UpperCamelCase =unicodedata.category(__lowerCAmelCase ) if cat == "Mn": continue output.append(__lowerCAmelCase ) return "".join(__lowerCAmelCase ) def UpperCamelCase__ ( self : Tuple , UpperCamelCase__ : str ) -> Optional[int]: _UpperCamelCase =( [chr(__lowerCAmelCase ) for i in range(ord('''a''' ) , ord('''z''' ) + 1 )] + [chr(__lowerCAmelCase ) for i in range(ord('''A''' ) , ord('''Z''' ) + 1 )] + [chr(__lowerCAmelCase ) for i in range(ord('''0''' ) , ord('''9''' ) + 1 )] + ['''.'''] ) _UpperCamelCase =frozenset(__lowerCAmelCase ) _UpperCamelCase =re.compile(R'''_+''' ) _UpperCamelCase =''''''.join([c if c in accepted else '''_''' for c in text.lower()] ) _UpperCamelCase =pattern.sub('''_''' , __lowerCAmelCase ).strip('''_''' ) return text def UpperCamelCase__ ( self : Dict , UpperCamelCase__ : List[str] ) -> Tuple: return " ".join(__lowerCAmelCase ) def UpperCamelCase__ ( self : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : Optional[Union[str, TensorType]] = None , UpperCamelCase__ : bool = False ) -> Optional[Any]: if not isinstance(__lowerCAmelCase , __lowerCAmelCase ): _UpperCamelCase =TensorType(__lowerCAmelCase ) # Get a function reference for the correct framework if tensor_type == TensorType.TENSORFLOW: if not is_tf_available(): raise ImportError( '''Unable to convert output to TensorFlow tensors format, TensorFlow is not installed.''' ) import tensorflow as tf _UpperCamelCase =tf.constant _UpperCamelCase =tf.is_tensor elif tensor_type == TensorType.PYTORCH: if not is_torch_available(): raise ImportError('''Unable to convert output to PyTorch tensors format, PyTorch is not installed.''' ) import torch _UpperCamelCase =torch.tensor _UpperCamelCase =torch.is_tensor elif tensor_type == TensorType.JAX: if not is_flax_available(): raise ImportError('''Unable to convert output to JAX tensors format, JAX is not installed.''' ) import jax.numpy as jnp # noqa: F811 _UpperCamelCase =jnp.array _UpperCamelCase =_is_jax else: _UpperCamelCase =np.asarray _UpperCamelCase =_is_numpy # Do the tensor conversion in batch try: if prepend_batch_axis: _UpperCamelCase =[inputs] if not is_tensor(__lowerCAmelCase ): _UpperCamelCase =as_tensor(__lowerCAmelCase ) except: # noqa E722 raise ValueError( '''Unable to create tensor, you should probably activate truncation and/or padding ''' '''with \'padding=True\' \'truncation=True\' to have batched tensors with the same length.''' ) return inputs def __call__( self : Tuple , UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : str , UpperCamelCase__ : int="" , UpperCamelCase__ : Union[str, Any]="pt" ) -> Any: _UpperCamelCase =[0, 0, 0] _UpperCamelCase =[artist] * len(self.version ) _UpperCamelCase =[genres] * len(self.version ) _UpperCamelCase =self.tokenize(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =self._convert_token_to_id(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) _UpperCamelCase =[-INFINITY] * len(full_tokens[-1] ) _UpperCamelCase =[ self.convert_to_tensors( [input_ids + [artists_id[i]] + genres_ids[i] + full_tokens[i]] , tensor_type=__lowerCAmelCase ) for i in range(len(self.version ) ) ] return BatchEncoding({'''input_ids''': input_ids, '''attention_masks''': attention_masks} ) def UpperCamelCase__ ( self : str , UpperCamelCase__ : str , UpperCamelCase__ : Optional[str] = None ) -> Optional[Any]: if not os.path.isdir(__lowerCAmelCase ): logger.error(F'''Vocabulary path ({save_directory}) should be a directory''' ) return _UpperCamelCase =os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''artists_file'''] ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.artists_encoder , ensure_ascii=__lowerCAmelCase ) ) _UpperCamelCase =os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''genres_file'''] ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.genres_encoder , ensure_ascii=__lowerCAmelCase ) ) _UpperCamelCase =os.path.join( __lowerCAmelCase , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''lyrics_file'''] ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(self.lyrics_encoder , ensure_ascii=__lowerCAmelCase ) ) return (artists_file, genres_file, lyrics_file) def UpperCamelCase__ ( self : Optional[Any] , UpperCamelCase__ : int , UpperCamelCase__ : Dict , UpperCamelCase__ : Dict ) -> Dict: _UpperCamelCase =self.artists_decoder.get(__lowerCAmelCase ) _UpperCamelCase =[self.genres_decoder.get(__lowerCAmelCase ) for genre in genres_index] _UpperCamelCase =[self.lyrics_decoder.get(__lowerCAmelCase ) for character in lyric_index] return artist, genres, lyrics
404
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar lowerCAmelCase__ = TypeVar('''T''') class __snake_case ( Generic[T]): def __init__( self : int , __lowerCAmelCase : T ): """simple docstring""" _lowerCamelCase : Optional[int] = data _lowerCamelCase : Node[T] | None = None def __str__( self : Optional[Any] ): """simple docstring""" return f'''{self.data}''' class __snake_case ( Generic[T]): def __init__( self : int ): """simple docstring""" _lowerCamelCase : Node[T] | None = None def __iter__( self : str ): """simple docstring""" _lowerCamelCase : List[str] = self.top while node: yield node.data _lowerCamelCase : Any = node.next def __str__( self : int ): """simple docstring""" return "->".join([str(__lowerCAmelCase ) for item in self] ) def __len__( self : int ): """simple docstring""" return len(tuple(iter(self ) ) ) def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" return self.top is None def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : T ): """simple docstring""" _lowerCamelCase : Tuple = Node(__lowerCAmelCase ) if not self.is_empty(): _lowerCamelCase : Optional[int] = self.top _lowerCamelCase : List[str] = node def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" if self.is_empty(): raise IndexError('''pop from empty stack''' ) assert isinstance(self.top , __lowerCAmelCase ) _lowerCamelCase : Any = self.top _lowerCamelCase : Any = self.top.next return pop_node.data def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" if self.is_empty(): raise IndexError('''peek from empty stack''' ) assert self.top is not None return self.top.data def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : List[str] = None if __name__ == "__main__": from doctest import testmod testmod()
83
0
import itertools import random import unittest import numpy as np from transformers import BatchFeature, SpeechTaFeatureExtractor from transformers.testing_utils import require_torch from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_torch_available(): import torch lowerCamelCase__ = random.Random() def A(__a: Union[str, Any] , __a: Union[str, Any]=1.0 , __a: Optional[Any]=None , __a: int=None ): if rng is None: lowerCAmelCase_ = global_rng lowerCAmelCase_ = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch class __magic_name__ (unittest.TestCase ): def __init__( self , _a , _a=7 , _a=400 , _a=2000 , _a=1 , _a=0.0 , _a=16000 , _a=True , _a=80 , _a=16 , _a=64 , _a="hann_window" , _a=80 , _a=7600 , _a=1E-10 , _a=True , ) -> int: lowerCAmelCase_ = parent lowerCAmelCase_ = batch_size lowerCAmelCase_ = min_seq_length lowerCAmelCase_ = max_seq_length lowerCAmelCase_ = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) lowerCAmelCase_ = feature_size lowerCAmelCase_ = padding_value lowerCAmelCase_ = sampling_rate lowerCAmelCase_ = do_normalize lowerCAmelCase_ = num_mel_bins lowerCAmelCase_ = hop_length lowerCAmelCase_ = win_length lowerCAmelCase_ = win_function lowerCAmelCase_ = fmin lowerCAmelCase_ = fmax lowerCAmelCase_ = mel_floor lowerCAmelCase_ = return_attention_mask def __a ( self ) -> List[str]: return { "feature_size": self.feature_size, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "do_normalize": self.do_normalize, "num_mel_bins": self.num_mel_bins, "hop_length": self.hop_length, "win_length": self.win_length, "win_function": self.win_function, "fmin": self.fmin, "fmax": self.fmax, "mel_floor": self.mel_floor, "return_attention_mask": self.return_attention_mask, } def __a ( self , _a=False , _a=False ) -> Dict: def _flatten(_a ): return list(itertools.chain(*__lowerCAmelCase ) ) if equal_length: lowerCAmelCase_ = floats_list((self.batch_size, self.max_seq_length) ) else: # make sure that inputs increase in size lowerCAmelCase_ = [ _flatten(floats_list((x, self.feature_size) ) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase_ = [np.asarray(__lowerCAmelCase ) for x in speech_inputs] return speech_inputs def __a ( self , _a=False , _a=False ) -> str: if equal_length: lowerCAmelCase_ = [floats_list((self.max_seq_length, self.num_mel_bins) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size lowerCAmelCase_ = [ floats_list((x, self.num_mel_bins) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: lowerCAmelCase_ = [np.asarray(__lowerCAmelCase ) for x in speech_inputs] return speech_inputs @require_torch class __magic_name__ (_lowercase , unittest.TestCase ): lowerCamelCase__ = SpeechTaFeatureExtractor def __a ( self ) -> int: lowerCAmelCase_ = SpeechTaFeatureExtractionTester(self ) def __a ( self , _a ) -> List[Any]: self.assertTrue(np.all(np.mean(__lowerCAmelCase , axis=0 ) < 1E-3 ) ) self.assertTrue(np.all(np.abs(np.var(__lowerCAmelCase , axis=0 ) - 1 ) < 1E-3 ) ) def __a ( self ) -> int: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = [np.asarray(__lowerCAmelCase ) for speech_input in speech_inputs] # Test not batched input lowerCAmelCase_ = feat_extract(speech_inputs[0] , return_tensors="np" ).input_values lowerCAmelCase_ = feat_extract(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test batched lowerCAmelCase_ = feat_extract(__lowerCAmelCase , return_tensors="np" ).input_values lowerCAmelCase_ = feat_extract(__lowerCAmelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) def __a ( self ) -> Dict: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase_ = [None, 1600, None] for max_length, padding in zip(__lowerCAmelCase , __lowerCAmelCase ): lowerCAmelCase_ = feat_extract(__lowerCAmelCase , padding=__lowerCAmelCase , max_length=__lowerCAmelCase , return_tensors="np" ) lowerCAmelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self.assertTrue(input_values[0][800:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self.assertTrue(input_values[0][1000:].sum() < 1E-6 ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Tuple: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ = range(800 , 1400 , 200 ) lowerCAmelCase_ = [floats_list((1, x) )[0] for x in lengths] lowerCAmelCase_ = ['''longest''', '''max_length''', '''do_not_pad'''] lowerCAmelCase_ = [None, 1600, None] for max_length, padding in zip(__lowerCAmelCase , __lowerCAmelCase ): lowerCAmelCase_ = feat_extract(__lowerCAmelCase , max_length=__lowerCAmelCase , padding=__lowerCAmelCase ) lowerCAmelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0][:800] ) self._check_zero_mean_unit_variance(input_values[1][:1000] ) self._check_zero_mean_unit_variance(input_values[2][:1200] ) def __a ( self ) -> Optional[int]: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = feat_extract( __lowerCAmelCase , truncation=__lowerCAmelCase , max_length=1000 , padding="max_length" , return_tensors="np" ) lowerCAmelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1] ) self._check_zero_mean_unit_variance(input_values[2] ) def __a ( self ) -> Dict: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = feat_extract( __lowerCAmelCase , truncation=__lowerCAmelCase , max_length=1000 , padding="longest" , return_tensors="np" ) lowerCAmelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length < longest -> then pad to max_length self.assertTrue(input_values.shape == (3, 1000) ) lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = feat_extract( __lowerCAmelCase , truncation=__lowerCAmelCase , max_length=2000 , padding="longest" , return_tensors="np" ) lowerCAmelCase_ = processed.input_values self._check_zero_mean_unit_variance(input_values[0, :800] ) self._check_zero_mean_unit_variance(input_values[1, :1000] ) self._check_zero_mean_unit_variance(input_values[2] ) # make sure that if max_length > longest -> then pad to longest self.assertTrue(input_values.shape == (3, 1200) ) def __a ( self ) -> int: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) lowerCAmelCase_ = np.random.rand(100 ).astype(np.floataa ) lowerCAmelCase_ = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: lowerCAmelCase_ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="np" ) self.assertTrue(np_processed.input_values.dtype == np.floataa ) lowerCAmelCase_ = feature_extractor.pad([{"input_values": inputs}] , return_tensors="pt" ) self.assertTrue(pt_processed.input_values.dtype == torch.floataa ) def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 lowerCAmelCase_ = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] lowerCAmelCase_ = [np.asarray(__lowerCAmelCase ) for speech_input in speech_inputs] # Test feature size lowerCAmelCase_ = feature_extractor(audio_target=__lowerCAmelCase , padding=__lowerCAmelCase , return_tensors="np" ).input_values self.assertTrue(input_values.ndim == 3 ) self.assertTrue(input_values.shape[-1] == feature_extractor.num_mel_bins ) # Test not batched input lowerCAmelCase_ = feature_extractor(speech_inputs[0] , return_tensors="np" ).input_values lowerCAmelCase_ = feature_extractor(np_speech_inputs[0] , return_tensors="np" ).input_values self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test batched lowerCAmelCase_ = feature_extractor(__lowerCAmelCase , return_tensors="np" ).input_values lowerCAmelCase_ = feature_extractor(__lowerCAmelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) # Test 2-D numpy arrays are batched. lowerCAmelCase_ = [floats_list((1, x) )[0] for x in (800, 800, 800)] lowerCAmelCase_ = np.asarray(__lowerCAmelCase ) lowerCAmelCase_ = feature_extractor(__lowerCAmelCase , return_tensors="np" ).input_values lowerCAmelCase_ = feature_extractor(__lowerCAmelCase , return_tensors="np" ).input_values for enc_seq_a, enc_seq_a in zip(__lowerCAmelCase , __lowerCAmelCase ): self.assertTrue(np.allclose(__lowerCAmelCase , __lowerCAmelCase , atol=1E-3 ) ) def __a ( self ) -> Union[str, Any]: lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target() lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase_ = feat_extract.model_input_names[0] lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} ) self.assertTrue(all(len(__lowerCAmelCase ) == len(__lowerCAmelCase ) for x, y in zip(__lowerCAmelCase , processed_features[input_name] ) ) ) lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target(equal_length=__lowerCAmelCase ) lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} , tensor_type="np" ) lowerCAmelCase_ = processed_features[input_name] if len(batch_features_input.shape ) < 3: lowerCAmelCase_ = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target(equal_length=__lowerCAmelCase ) lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase_ = feat_extract.model_input_names[0] lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} , tensor_type="pt" ) lowerCAmelCase_ = processed_features[input_name] if len(batch_features_input.shape ) < 3: lowerCAmelCase_ = batch_features_input[:, :, None] self.assertTrue( batch_features_input.shape == (self.feat_extract_tester.batch_size, len(speech_inputs[0] ), self.feat_extract_tester.num_mel_bins) ) @require_torch def __a ( self ) -> Optional[Any]: lowerCAmelCase_ = self.feature_extraction_class(**self.feat_extract_dict ) lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target() lowerCAmelCase_ = feat_extract.model_input_names[0] lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase_ = feat_extract.num_mel_bins # hack! lowerCAmelCase_ = feat_extract.pad(__lowerCAmelCase , padding="longest" , return_tensors="np" )[input_name] lowerCAmelCase_ = feat_extract.pad(__lowerCAmelCase , padding="longest" , return_tensors="pt" )[input_name] self.assertTrue(abs(input_np.astype(np.floataa ).sum() - input_pt.numpy().astype(np.floataa ).sum() ) < 1E-2 ) def __a ( self ) -> Optional[Any]: lowerCAmelCase_ = self.feat_extract_dict lowerCAmelCase_ = True lowerCAmelCase_ = self.feature_extraction_class(**__lowerCAmelCase ) lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target() lowerCAmelCase_ = [len(__lowerCAmelCase ) for x in speech_inputs] lowerCAmelCase_ = feat_extract.model_input_names[0] lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase_ = feat_extract.num_mel_bins # hack! lowerCAmelCase_ = feat_extract.pad(__lowerCAmelCase , padding="longest" , return_tensors="np" ) self.assertIn("attention_mask" , __lowerCAmelCase ) self.assertListEqual(list(processed.attention_mask.shape ) , list(processed[input_name].shape[:2] ) ) self.assertListEqual(processed.attention_mask.sum(-1 ).tolist() , __lowerCAmelCase ) def __a ( self ) -> str: lowerCAmelCase_ = self.feat_extract_dict lowerCAmelCase_ = True lowerCAmelCase_ = self.feature_extraction_class(**__lowerCAmelCase ) lowerCAmelCase_ = self.feat_extract_tester.prepare_inputs_for_target() lowerCAmelCase_ = [len(__lowerCAmelCase ) for x in speech_inputs] lowerCAmelCase_ = feat_extract.model_input_names[0] lowerCAmelCase_ = BatchFeature({input_name: speech_inputs} ) lowerCAmelCase_ = min(__lowerCAmelCase ) lowerCAmelCase_ = feat_extract.num_mel_bins # hack! lowerCAmelCase_ = feat_extract.pad( __lowerCAmelCase , padding="max_length" , max_length=__lowerCAmelCase , truncation=__lowerCAmelCase , return_tensors="np" ) self.assertIn("attention_mask" , __lowerCAmelCase ) self.assertListEqual( list(processed_pad.attention_mask.shape ) , [processed_pad[input_name].shape[0], max_length] ) self.assertListEqual( processed_pad.attention_mask[:, :max_length].sum(-1 ).tolist() , [max_length for x in speech_inputs] ) def __a ( self , _a ) -> str: from datasets import load_dataset lowerCAmelCase_ = load_dataset("hf-internal-testing/librispeech_asr_dummy" , "clean" , split="validation" ) # automatic decoding with librispeech lowerCAmelCase_ = ds.sort("id" ).select(range(__lowerCAmelCase ) )[:num_samples]['''audio'''] return [x["array"] for x in speech_samples] def __a ( self ) -> str: lowerCAmelCase_ = torch.tensor( [2.3804E-03, 2.0752E-03, 1.9836E-03, 2.1057E-03, 1.6174E-03, 3.0518E-04, 9.1553E-05, 3.3569E-04, 9.7656E-04, 1.8311E-03, 2.0142E-03, 2.1057E-03, 1.7395E-03, 4.5776E-04, -3.9673E-04, 4.5776E-04, 1.0071E-03, 9.1553E-05, 4.8828E-04, 1.1597E-03, 7.3242E-04, 9.4604E-04, 1.8005E-03, 1.8311E-03, 8.8501E-04, 4.2725E-04, 4.8828E-04, 7.3242E-04, 1.0986E-03, 2.1057E-03] ) # fmt: on lowerCAmelCase_ = self._load_datasamples(1 ) lowerCAmelCase_ = SpeechTaFeatureExtractor() lowerCAmelCase_ = feature_extractor(__lowerCAmelCase , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 93680) ) self.assertTrue(torch.allclose(input_values[0, :30] , __lowerCAmelCase , atol=1E-6 ) ) def __a ( self ) -> Optional[int]: lowerCAmelCase_ = torch.tensor( [-2.6_8_7_0, -3.0_1_0_4, -3.1_3_5_6, -3.5_3_5_2, -3.0_0_4_4, -3.0_3_5_3, -3.4_7_1_9, -3.6_7_7_7, -3.1_5_2_0, -2.9_4_3_5, -2.6_5_5_3, -2.8_7_9_5, -2.9_9_4_4, -2.5_9_2_1, -3.0_2_7_9, -3.0_3_8_6, -3.0_8_6_4, -3.1_2_9_1, -3.2_3_5_3, -2.7_4_4_4, -2.6_8_3_1, -2.7_2_8_7, -3.1_7_6_1, -3.1_5_7_1, -3.2_7_2_6, -3.0_5_8_2, -3.1_0_0_7, -3.4_5_3_3, -3.4_6_9_5, -3.0_9_9_8] ) # fmt: on lowerCAmelCase_ = self._load_datasamples(1 ) lowerCAmelCase_ = SpeechTaFeatureExtractor() lowerCAmelCase_ = feature_extractor(audio_target=__lowerCAmelCase , return_tensors="pt" ).input_values self.assertEquals(input_values.shape , (1, 366, 80) ) self.assertTrue(torch.allclose(input_values[0, 0, :30] , __lowerCAmelCase , atol=1E-4 ) )
122
"""simple docstring""" import importlib import inspect import os import re # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py lowerCAmelCase__ = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. lowerCAmelCase__ = importlib.util.spec_from_file_location( '''transformers''', os.path.join(PATH_TO_TRANSFORMERS, '''__init__.py'''), submodule_search_locations=[PATH_TO_TRANSFORMERS], ) lowerCAmelCase__ = spec.loader.load_module() lowerCAmelCase__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING # Regex pattern used to find the checkpoint mentioned in the docstring of `config_class`. # For example, `[bert-base-uncased](https://huggingface.co/bert-base-uncased)` lowerCAmelCase__ = re.compile('''\[(.+?)\]\((https://huggingface\.co/.+?)\)''') lowerCAmelCase__ = { '''CLIPConfigMixin''', '''DecisionTransformerConfigMixin''', '''EncoderDecoderConfigMixin''', '''RagConfigMixin''', '''SpeechEncoderDecoderConfigMixin''', '''VisionEncoderDecoderConfigMixin''', '''VisionTextDualEncoderConfigMixin''', } def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Any = [] for config_class in list(CONFIG_MAPPING.values() ): _lowerCamelCase : Tuple = False # source code of `config_class` _lowerCamelCase : int = inspect.getsource(A_ ) _lowerCamelCase : str = _re_checkpoint.findall(A_ ) for checkpoint in checkpoints: # Each `checkpoint` is a tuple of a checkpoint name and a checkpoint link. # For example, `('bert-base-uncased', 'https://huggingface.co/bert-base-uncased')` _lowerCamelCase , _lowerCamelCase : Tuple = checkpoint # verify the checkpoint name corresponds to the checkpoint link _lowerCamelCase : Tuple = F'''https://huggingface.co/{ckpt_name}''' if ckpt_link == ckpt_link_from_name: _lowerCamelCase : Union[str, Any] = True break _lowerCamelCase : Tuple = config_class.__name__ if not checkpoint_found and name not in CONFIG_CLASSES_TO_IGNORE_FOR_DOCSTRING_CHECKPOINT_CHECK: configs_without_checkpoint.append(A_ ) if len(A_ ) > 0: _lowerCamelCase : Union[str, Any] = '''\n'''.join(sorted(A_ ) ) raise ValueError(F'''The following configurations don\'t contain any valid checkpoint:\n{message}''' ) if __name__ == "__main__": check_config_docstrings_have_checkpoints()
83
0
"""simple docstring""" import inspect import os import unittest import torch import accelerate from accelerate import debug_launcher from accelerate.test_utils import ( execute_subprocess_async, require_cpu, require_huggingface_suite, require_multi_gpu, require_single_gpu, ) from accelerate.utils import patch_environment @require_huggingface_suite class __UpperCamelCase ( unittest.TestCase ): def __lowerCamelCase ( self ): '''simple docstring''' _lowerCAmelCase : Any = inspect.getfile(accelerate.test_utils ) _lowerCAmelCase : Union[str, Any] = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ['scripts', 'external_deps', 'test_metrics.py'] ) from accelerate.test_utils.scripts.external_deps import test_metrics # noqa: F401 _lowerCAmelCase : Optional[Any] = test_metrics @require_cpu def __lowerCamelCase ( self ): '''simple docstring''' debug_launcher(self.test_metrics.main ,num_processes=1 ) @require_cpu def __lowerCamelCase ( self ): '''simple docstring''' debug_launcher(self.test_metrics.main ) @require_single_gpu def __lowerCamelCase ( self ): '''simple docstring''' self.test_metrics.main() @require_multi_gpu def __lowerCamelCase ( self ): '''simple docstring''' print(F"""Found {torch.cuda.device_count()} devices.""" ) _lowerCAmelCase : str = ['''torchrun''', F"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__lowerCAmelCase ,env=os.environ.copy() )
259
"""simple docstring""" import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCAmelCase__ = False class __snake_case ( unittest.TestCase): pass @nightly @require_torch_gpu class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : int = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : str = '''A painting of a squirrel eating a burger ''' _lowerCamelCase : Dict = torch.manual_seed(0 ) _lowerCamelCase : List[Any] = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(__lowerCAmelCase ) _lowerCamelCase : Tuple = VersatileDiffusionTextToImagePipeline.from_pretrained(__lowerCAmelCase ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : int = generator.manual_seed(0 ) _lowerCamelCase : List[str] = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Optional[int] = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(__lowerCAmelCase ) pipe.set_progress_bar_config(disable=__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = '''A painting of a squirrel eating a burger ''' _lowerCamelCase : Optional[int] = torch.manual_seed(0 ) _lowerCamelCase : int = pipe( prompt=__lowerCAmelCase , generator=__lowerCAmelCase , guidance_scale=7.5 , num_inference_steps=5_0 , output_type='''numpy''' ).images _lowerCamelCase : List[str] = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) _lowerCamelCase : Dict = np.array([0.33_67, 0.31_69, 0.26_56, 0.38_70, 0.47_90, 0.37_96, 0.40_09, 0.48_78, 0.47_78] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
83
0
import math import random from typing import Any from .hill_climbing import SearchProblem def __lowerCAmelCase ( _UpperCamelCase : Dict , _UpperCamelCase : bool = True , _UpperCamelCase : float = math.inf , _UpperCamelCase : float = -math.inf , _UpperCamelCase : float = math.inf , _UpperCamelCase : float = -math.inf , _UpperCamelCase : bool = False , _UpperCamelCase : float = 1_00 , _UpperCamelCase : float = 0.01 , _UpperCamelCase : float = 1 , ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE = False SCREAMING_SNAKE_CASE = search_prob SCREAMING_SNAKE_CASE = start_temperate SCREAMING_SNAKE_CASE = [] SCREAMING_SNAKE_CASE = 0 SCREAMING_SNAKE_CASE = None while not search_end: SCREAMING_SNAKE_CASE = current_state.score() if best_state is None or current_score > best_state.score(): SCREAMING_SNAKE_CASE = current_state scores.append(A_ ) iterations += 1 SCREAMING_SNAKE_CASE = None SCREAMING_SNAKE_CASE = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to SCREAMING_SNAKE_CASE = random.randint(0 , len(A_ ) - 1 ) # picking a random neighbor SCREAMING_SNAKE_CASE = neighbors.pop(A_ ) SCREAMING_SNAKE_CASE = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: SCREAMING_SNAKE_CASE = change * -1 # in case we are finding minimum if change > 0: # improves the solution SCREAMING_SNAKE_CASE = picked_neighbor else: SCREAMING_SNAKE_CASE = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability SCREAMING_SNAKE_CASE = picked_neighbor SCREAMING_SNAKE_CASE = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor SCREAMING_SNAKE_CASE = True else: SCREAMING_SNAKE_CASE = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(A_ ) , A_ ) plt.xlabel('Iterations' ) plt.ylabel('Function values' ) plt.show() return best_state if __name__ == "__main__": def __lowerCAmelCase ( _UpperCamelCase : int , _UpperCamelCase : Tuple ) -> Optional[Any]: '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) a_ : Any = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) a_ : int = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( "The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 " F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) a_ : Union[str, Any] = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) a_ : int = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( "The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 " F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def __lowerCAmelCase ( _UpperCamelCase : Optional[int] , _UpperCamelCase : List[Any] ) -> Optional[int]: '''simple docstring''' return (3 * x**2) - (6 * y) a_ : List[Any] = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) a_ : str = simulated_annealing(prob, find_max=False, visualization=True) print( "The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: " F"""{local_min.score()}""" ) a_ : Tuple = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) a_ : Optional[int] = simulated_annealing(prob, find_max=True, visualization=True) print( "The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: " F"""{local_min.score()}""" )
439
"""simple docstring""" import os import time import numpy as np import onnxruntime as ort lowerCAmelCase__ = '''1''' lowerCAmelCase__ = '''0''' lowerCAmelCase__ = '''1''' lowerCAmelCase__ = ort.SessionOptions() lowerCAmelCase__ = ort.GraphOptimizationLevel.ORT_DISABLE_ALL print('''Create inference session...''') lowerCAmelCase__ = ['''TensorrtExecutionProvider''', '''CUDAExecutionProvider'''] lowerCAmelCase__ = ort.InferenceSession('''model.onnx''', sess_options=sess_opt, providers=execution_provider) lowerCAmelCase__ = ort.RunOptions() lowerCAmelCase__ = 128 lowerCAmelCase__ = 1 lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) lowerCAmelCase__ = np.ones((batch, sequence), dtype=np.intaa) print('''Warm up phase...''') sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Start inference...''') lowerCAmelCase__ = time.time() lowerCAmelCase__ = 2000 lowerCAmelCase__ = {} for iter in range(max_iters): lowerCAmelCase__ = sess.run( None, { sess.get_inputs()[0].name: input_ids, sess.get_inputs()[1].name: attention_mask, sess.get_inputs()[2].name: token_type_ids, }, run_options=run_opt, ) print('''Average Inference Time = {:.3f} ms'''.format((time.time() - start_time) * 1000 / max_iters))
83
0
import os import tempfile import unittest import uuid from pathlib import Path from transformers.testing_utils import get_tests_dir, require_soundfile, require_torch, require_vision from transformers.tools.agent_types import AgentAudio, AgentImage, AgentText from transformers.utils import is_soundfile_availble, is_torch_available, is_vision_available if is_torch_available(): import torch if is_soundfile_availble(): import soundfile as sf if is_vision_available(): from PIL import Image def lowercase_( SCREAMING_SNAKE_CASE_="" ): '''simple docstring''' lowerCamelCase : Union[str, Any] = tempfile.mkdtemp() return os.path.join(A_ , str(uuid.uuida() ) + suffix ) @require_soundfile @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[int] = torch.rand(12 , dtype=torch.floataa ) - 0.5 lowerCamelCase : List[str] = AgentAudio(__lowerCAmelCase ) lowerCamelCase : Dict = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(__lowerCAmelCase , agent_type.to_raw() , atol=1e-4 ) ) del agent_type # Ensure the path remains even after the object deletion self.assertTrue(os.path.exists(__lowerCAmelCase ) ) # Ensure that the file contains the same value as the original tensor lowerCamelCase : Union[str, Any] = sf.read(__lowerCAmelCase ) self.assertTrue(torch.allclose(__lowerCAmelCase , torch.tensor(__lowerCAmelCase ) , atol=1e-4 ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : Tuple = torch.rand(12 , dtype=torch.floataa ) - 0.5 lowerCamelCase : Union[str, Any] = get_new_path(suffix=".wav" ) sf.write(__lowerCAmelCase , __lowerCAmelCase , 1_6000 ) lowerCamelCase : Optional[Any] = AgentAudio(__lowerCAmelCase ) self.assertTrue(torch.allclose(__lowerCAmelCase , agent_type.to_raw() , atol=1e-4 ) ) self.assertEqual(agent_type.to_string() , __lowerCAmelCase ) @require_vision @require_torch class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" lowerCamelCase : Dict = torch.randint(0 , 256 , (64, 64, 3) ) lowerCamelCase : Tuple = AgentImage(__lowerCAmelCase ) lowerCamelCase : Union[str, Any] = str(agent_type.to_string() ) # Ensure that the tensor and the agent_type's tensor are the same self.assertTrue(torch.allclose(__lowerCAmelCase , agent_type._tensor , atol=1e-4 ) ) self.assertIsInstance(agent_type.to_raw() , Image.Image ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(__lowerCAmelCase ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : str = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / '''000000039769.png''' lowerCamelCase : Optional[int] = Image.open(__lowerCAmelCase ) lowerCamelCase : List[str] = AgentImage(__lowerCAmelCase ) self.assertTrue(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(__lowerCAmelCase ) ) def _snake_case ( self ): """simple docstring""" lowerCamelCase : List[str] = Path(get_tests_dir("fixtures/tests_samples/COCO" ) ) / '''000000039769.png''' lowerCamelCase : Dict = Image.open(__lowerCAmelCase ) lowerCamelCase : Optional[Any] = AgentImage(__lowerCAmelCase ) self.assertFalse(path.samefile(agent_type.to_string() ) ) self.assertTrue(image == agent_type.to_raw() ) # Ensure the path remains even after the object deletion del agent_type self.assertTrue(os.path.exists(__lowerCAmelCase ) ) class UpperCAmelCase_ ( unittest.TestCase ): '''simple docstring''' def _snake_case ( self ): """simple docstring""" lowerCamelCase : Optional[int] = '''Hey!''' lowerCamelCase : int = AgentText(__lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , agent_type.to_string() ) self.assertEqual(__lowerCAmelCase , agent_type.to_raw() ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase )
340
"""simple docstring""" import colorsys from PIL import Image # type: ignore def snake_case_ ( A_ : float, A_ : float, A_ : int ): '''simple docstring''' _lowerCamelCase : List[Any] = x _lowerCamelCase : List[Any] = y for step in range(A_ ): # noqa: B007 _lowerCamelCase : Dict = a * a - b * b + x _lowerCamelCase : List[str] = 2 * a * b + y _lowerCamelCase : Any = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def snake_case_ ( A_ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (2_55, 2_55, 2_55) def snake_case_ ( A_ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 2_55 ) for i in colorsys.hsv_to_rgb(A_, 1, 1 ) ) def snake_case_ ( A_ : int = 8_00, A_ : int = 6_00, A_ : float = -0.6, A_ : float = 0, A_ : float = 3.2, A_ : int = 50, A_ : bool = True, ): '''simple docstring''' _lowerCamelCase : Tuple = Image.new('''RGB''', (image_width, image_height) ) _lowerCamelCase : int = img.load() # loop through the image-coordinates for image_x in range(A_ ): for image_y in range(A_ ): # determine the figure-coordinates based on the image-coordinates _lowerCamelCase : Optional[Any] = figure_width / image_width * image_height _lowerCamelCase : List[Any] = figure_center_x + (image_x / image_width - 0.5) * figure_width _lowerCamelCase : Optional[Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height _lowerCamelCase : str = get_distance(A_, A_, A_ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _lowerCamelCase : Dict = get_color_coded_rgb(A_ ) else: _lowerCamelCase : str = get_black_and_white_rgb(A_ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure lowerCAmelCase__ = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
83
0
import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType SCREAMING_SNAKE_CASE__ = logging.get_logger(__name__) class __lowerCAmelCase ( _lowercase ): """simple docstring""" A__ : Optional[int] = "vision-encoder-decoder" A__ : str = True def __init__( self : Tuple , **_snake_case : Optional[int] ): """simple docstring""" super().__init__(**__lowerCAmelCase ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F'''A configuraton of type {self.model_type} cannot be instantiated because ''' F'''not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}''' ) A__ = kwargs.pop('encoder' ) A__ = encoder_config.pop('model_type' ) A__ = kwargs.pop('decoder' ) A__ = decoder_config.pop('model_type' ) A__ = AutoConfig.for_model(__lowerCAmelCase , **__lowerCAmelCase ) A__ = AutoConfig.for_model(__lowerCAmelCase , **__lowerCAmelCase ) A__ = True @classmethod def _a ( cls : Any , _snake_case : PretrainedConfig , _snake_case : PretrainedConfig , **_snake_case : Optional[Any] ): """simple docstring""" logger.info('Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config' ) A__ = True A__ = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **__lowerCAmelCase ) def _a ( self : Union[str, Any] ): """simple docstring""" A__ = copy.deepcopy(self.__dict__ ) A__ = self.encoder.to_dict() A__ = self.decoder.to_dict() A__ = self.__class__.model_type return output class __lowerCAmelCase ( _lowercase ): """simple docstring""" A__ : Dict = version.parse("1.11" ) @property def _a ( self : Dict ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) @property def _a ( self : Optional[int] ): """simple docstring""" return 1E-4 @property def _a ( self : List[str] ): """simple docstring""" return OrderedDict({'last_hidden_state': {0: 'batch', 1: 'encoder_sequence'}} ) class __lowerCAmelCase ( _lowercase ): """simple docstring""" @property def _a ( self : Optional[int] ): """simple docstring""" A__ = OrderedDict() A__ = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} A__ = {0: '''batch''', 1: '''past_decoder_sequence + sequence'''} A__ = {0: '''batch''', 1: '''encoder_sequence'''} return common_inputs def _a ( self : str , _snake_case : "PreTrainedTokenizerBase" , _snake_case : int = -1 , _snake_case : int = -1 , _snake_case : bool = False , _snake_case : Optional["TensorType"] = None , ): """simple docstring""" import torch A__ = OrderedDict() A__ = super().generate_dummy_inputs( __lowerCAmelCase , batch_size=__lowerCAmelCase , seq_length=__lowerCAmelCase , is_pair=__lowerCAmelCase , framework=__lowerCAmelCase ) A__ = dummy_input['''input_ids'''].shape A__ = (batch, encoder_sequence, self._config.encoder_hidden_size) A__ = dummy_input.pop('input_ids' ) A__ = dummy_input.pop('attention_mask' ) A__ = torch.zeros(__lowerCAmelCase ) return common_inputs class __lowerCAmelCase ( _lowercase ): """simple docstring""" @property def _a ( self : Dict ): """simple docstring""" pass def _a ( self : List[str] , _snake_case : PretrainedConfig ): """simple docstring""" return VisionEncoderDecoderEncoderOnnxConfig(__lowerCAmelCase ) def _a ( self : int , _snake_case : PretrainedConfig , _snake_case : PretrainedConfig , _snake_case : str = "default" ): """simple docstring""" A__ = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(__lowerCAmelCase , __lowerCAmelCase )
9
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def snake_case_ ( A_ : Tuple, A_ : List[str], A_ : Optional[Any], A_ : Dict, A_ : Dict=True, A_ : int="pt" ): '''simple docstring''' _lowerCamelCase : str = {'''add_prefix_space''': True} if isinstance(A_, A_ ) and not line.startswith(''' ''' ) else {} _lowerCamelCase : Union[str, Any] = padding_side return tokenizer( [line], max_length=A_, padding='''max_length''' if pad_to_max_length else None, truncation=A_, return_tensors=A_, add_special_tokens=A_, **A_, ) def snake_case_ ( A_ : Any, A_ : Optional[int], A_ : List[Any]=None, ): '''simple docstring''' _lowerCamelCase : Optional[int] = input_ids.ne(A_ ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class __snake_case ( _lowercase): def __init__( self : Tuple , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Tuple="train" , __lowerCAmelCase : List[str]=None , __lowerCAmelCase : Optional[int]=None , __lowerCAmelCase : Any=None , __lowerCAmelCase : Union[str, Any]="" , ): """simple docstring""" super().__init__() _lowerCamelCase : Optional[int] = Path(__lowerCAmelCase ).joinpath(type_path + '''.source''' ) _lowerCamelCase : List[str] = Path(__lowerCAmelCase ).joinpath(type_path + '''.target''' ) _lowerCamelCase : List[Any] = self.get_char_lens(self.src_file ) _lowerCamelCase : Optional[int] = max_source_length _lowerCamelCase : Optional[Any] = max_target_length assert min(self.src_lens ) > 0, f'''found empty line in {self.src_file}''' _lowerCamelCase : List[Any] = tokenizer _lowerCamelCase : List[Any] = prefix if n_obs is not None: _lowerCamelCase : List[str] = self.src_lens[:n_obs] _lowerCamelCase : int = src_lang _lowerCamelCase : Union[str, Any] = tgt_lang def __len__( self : int ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : Dict , __lowerCAmelCase : Optional[Any] ): """simple docstring""" _lowerCamelCase : str = index + 1 # linecache starts at 1 _lowerCamelCase : Union[str, Any] = self.prefix + linecache.getline(str(self.src_file ) , __lowerCAmelCase ).rstrip('''\n''' ) _lowerCamelCase : Optional[Any] = linecache.getline(str(self.tgt_file ) , __lowerCAmelCase ).rstrip('''\n''' ) assert source_line, f'''empty source line for index {index}''' assert tgt_line, f'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer , __lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right _lowerCamelCase : Optional[int] = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer ) _lowerCamelCase : Union[str, Any] = self.tokenizer.generator if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_source_length , '''right''' ) _lowerCamelCase : List[str] = encode_line(__lowerCAmelCase , __lowerCAmelCase , self.max_target_length , '''right''' ) _lowerCamelCase : Optional[Any] = source_inputs['''input_ids'''].squeeze() _lowerCamelCase : Union[str, Any] = target_inputs['''input_ids'''].squeeze() _lowerCamelCase : Any = source_inputs['''attention_mask'''].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : str ): """simple docstring""" return [len(__lowerCAmelCase ) for x in Path(__lowerCAmelCase ).open().readlines()] def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[Any] = torch.stack([x['''input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = torch.stack([x['''attention_mask'''] for x in batch] ) _lowerCamelCase : Union[str, Any] = torch.stack([x['''decoder_input_ids'''] for x in batch] ) _lowerCamelCase : Tuple = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Tuple = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , __lowerCAmelCase ) else self.tokenizer.pad_token_id ) _lowerCamelCase : Union[str, Any] = trim_batch(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase , _lowerCamelCase : List[str] = trim_batch(__lowerCAmelCase , __lowerCAmelCase , attention_mask=__lowerCAmelCase ) _lowerCamelCase : Optional[int] = { '''input_ids''': source_ids, '''attention_mask''': source_mask, '''decoder_input_ids''': y, } return batch lowerCAmelCase__ = getLogger(__name__) def snake_case_ ( A_ : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(A_ ) ) def snake_case_ ( A_ : str ): '''simple docstring''' _lowerCamelCase : Dict = get_git_info() save_json(A_, os.path.join(A_, '''git_log.json''' ) ) def snake_case_ ( A_ : str, A_ : Union[str, Any], A_ : int=4, **A_ : Optional[int] ): '''simple docstring''' with open(A_, '''w''' ) as f: json.dump(A_, A_, indent=A_, **A_ ) def snake_case_ ( A_ : Any ): '''simple docstring''' with open(A_ ) as f: return json.load(A_ ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = git.Repo(search_parent_directories=A_ ) _lowerCamelCase : str = { '''repo_id''': str(A_ ), '''repo_sha''': str(repo.head.object.hexsha ), '''repo_branch''': str(repo.active_branch ), '''hostname''': str(socket.gethostname() ), } return repo_infos def snake_case_ ( A_ : Callable, A_ : Iterable ): '''simple docstring''' return list(map(A_, A_ ) ) def snake_case_ ( A_ : str, A_ : Tuple ): '''simple docstring''' with open(A_, '''wb''' ) as f: return pickle.dump(A_, A_ ) def snake_case_ ( A_ : List[str] ): '''simple docstring''' def remove_articles(A_ : str ): return re.sub(R'''\b(a|an|the)\b''', ''' ''', A_ ) def white_space_fix(A_ : Any ): return " ".join(text.split() ) def remove_punc(A_ : List[Any] ): _lowerCamelCase : Any = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(A_ : Optional[int] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(A_ ) ) ) ) def snake_case_ ( A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : str = normalize_answer(A_ ).split() _lowerCamelCase : int = normalize_answer(A_ ).split() _lowerCamelCase : str = Counter(A_ ) & Counter(A_ ) _lowerCamelCase : Any = sum(common.values() ) if num_same == 0: return 0 _lowerCamelCase : int = 1.0 * num_same / len(A_ ) _lowerCamelCase : str = 1.0 * num_same / len(A_ ) _lowerCamelCase : List[Any] = (2 * precision * recall) / (precision + recall) return fa def snake_case_ ( A_ : Dict, A_ : str ): '''simple docstring''' return normalize_answer(A_ ) == normalize_answer(A_ ) def snake_case_ ( A_ : List[str], A_ : List[str] ): '''simple docstring''' assert len(A_ ) == len(A_ ) _lowerCamelCase : Optional[Any] = 0 for hypo, pred in zip(A_, A_ ): em += exact_match_score(A_, A_ ) if len(A_ ) > 0: em /= len(A_ ) return {"em": em} def snake_case_ ( A_ : Optional[int] ): '''simple docstring''' return model_prefix.startswith('''rag''' ) def snake_case_ ( A_ : Dict, A_ : int, A_ : List[Any] ): '''simple docstring''' _lowerCamelCase : Dict = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead _lowerCamelCase : Tuple = '''dropout_rate''' for p in extra_params: if getattr(A_, A_, A_ ): if not hasattr(A_, A_ ) and not hasattr(A_, equivalent_param[p] ): logger.info('''config doesn\'t have a `{}` attribute'''.format(A_ ) ) delattr(A_, A_ ) continue _lowerCamelCase : Union[str, Any] = p if hasattr(A_, A_ ) else equivalent_param[p] setattr(A_, A_, getattr(A_, A_ ) ) delattr(A_, A_ ) return hparams, config
83
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCamelCase__ : Union[str, Any] = {"""configuration_encoder_decoder""": ["""EncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : Union[str, Any] = ["""EncoderDecoderModel"""] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : str = ["""TFEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase__ : List[str] = ["""FlaxEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys lowerCamelCase__ : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
33
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''camembert-base''': '''https://huggingface.co/camembert-base/resolve/main/config.json''', '''umberto-commoncrawl-cased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json''' ), '''umberto-wikipedia-uncased-v1''': ( '''https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json''' ), } class __snake_case ( _lowercase): snake_case__ : Optional[Any] = "camembert" def __init__( self : Optional[Any] , __lowerCAmelCase : Any=3_0_5_2_2 , __lowerCAmelCase : List[str]=7_6_8 , __lowerCAmelCase : List[str]=1_2 , __lowerCAmelCase : Optional[int]=1_2 , __lowerCAmelCase : List[Any]=3_0_7_2 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Union[str, Any]=0.1 , __lowerCAmelCase : Optional[int]=0.1 , __lowerCAmelCase : Optional[int]=5_1_2 , __lowerCAmelCase : str=2 , __lowerCAmelCase : int=0.02 , __lowerCAmelCase : List[Any]=1E-12 , __lowerCAmelCase : Union[str, Any]=1 , __lowerCAmelCase : Optional[Any]=0 , __lowerCAmelCase : List[Any]=2 , __lowerCAmelCase : str="absolute" , __lowerCAmelCase : Any=True , __lowerCAmelCase : Optional[int]=None , **__lowerCAmelCase : Optional[int] , ): """simple docstring""" super().__init__(pad_token_id=__lowerCAmelCase , bos_token_id=__lowerCAmelCase , eos_token_id=__lowerCAmelCase , **__lowerCAmelCase ) _lowerCamelCase : Tuple = vocab_size _lowerCamelCase : str = hidden_size _lowerCamelCase : Union[str, Any] = num_hidden_layers _lowerCamelCase : Any = num_attention_heads _lowerCamelCase : Optional[Any] = hidden_act _lowerCamelCase : List[str] = intermediate_size _lowerCamelCase : Optional[Any] = hidden_dropout_prob _lowerCamelCase : List[Any] = attention_probs_dropout_prob _lowerCamelCase : Optional[Any] = max_position_embeddings _lowerCamelCase : Tuple = type_vocab_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Dict = layer_norm_eps _lowerCamelCase : List[Any] = position_embedding_type _lowerCamelCase : int = use_cache _lowerCamelCase : List[str] = classifier_dropout class __snake_case ( _lowercase): @property def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" if self.task == "multiple-choice": _lowerCamelCase : Optional[Any] = {0: '''batch''', 1: '''choice''', 2: '''sequence'''} else: _lowerCamelCase : Union[str, Any] = {0: '''batch''', 1: '''sequence'''} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
83
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer lowerCAmelCase__: Dict = logging.get_logger(__name__) lowerCAmelCase__: Optional[Any] = {"vocab_file": "vocab.txt"} lowerCAmelCase__: Tuple = { "vocab_file": { "YituTech/conv-bert-base": "https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt", "YituTech/conv-bert-medium-small": ( "https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt" ), "YituTech/conv-bert-small": "https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt", } } lowerCAmelCase__: str = { "YituTech/conv-bert-base": 512, "YituTech/conv-bert-medium-small": 512, "YituTech/conv-bert-small": 512, } lowerCAmelCase__: str = { "YituTech/conv-bert-base": {"do_lower_case": True}, "YituTech/conv-bert-medium-small": {"do_lower_case": True}, "YituTech/conv-bert-small": {"do_lower_case": True}, } class snake_case_ ( _lowercase ): __lowerCamelCase : List[str] = VOCAB_FILES_NAMES __lowerCamelCase : int = PRETRAINED_VOCAB_FILES_MAP __lowerCamelCase : List[str] = PRETRAINED_INIT_CONFIGURATION __lowerCamelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowerCamelCase : Optional[int] = ConvBertTokenizer def __init__( self , __lowerCAmelCase=None , __lowerCAmelCase=None , __lowerCAmelCase=True , __lowerCAmelCase="[UNK]" , __lowerCAmelCase="[SEP]" , __lowerCAmelCase="[PAD]" , __lowerCAmelCase="[CLS]" , __lowerCAmelCase="[MASK]" , __lowerCAmelCase=True , __lowerCAmelCase=None , **__lowerCAmelCase , ): super().__init__( __lowerCAmelCase , tokenizer_file=__lowerCAmelCase , do_lower_case=__lowerCAmelCase , unk_token=__lowerCAmelCase , sep_token=__lowerCAmelCase , pad_token=__lowerCAmelCase , cls_token=__lowerCAmelCase , mask_token=__lowerCAmelCase , tokenize_chinese_chars=__lowerCAmelCase , strip_accents=__lowerCAmelCase , **__lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , __lowerCAmelCase ) != do_lower_case or normalizer_state.get('strip_accents' , __lowerCAmelCase ) != strip_accents or normalizer_state.get('handle_chinese_chars' , __lowerCAmelCase ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE_ : int = getattr(__lowerCAmelCase , normalizer_state.pop('type' ) ) SCREAMING_SNAKE_CASE_ : Tuple = do_lower_case SCREAMING_SNAKE_CASE_ : str = strip_accents SCREAMING_SNAKE_CASE_ : List[str] = tokenize_chinese_chars SCREAMING_SNAKE_CASE_ : Dict = normalizer_class(**__lowerCAmelCase ) SCREAMING_SNAKE_CASE_ : str = do_lower_case def __A ( self , __lowerCAmelCase , __lowerCAmelCase=None ): SCREAMING_SNAKE_CASE_ : List[Any] = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __A ( self , __lowerCAmelCase , __lowerCAmelCase = None ): SCREAMING_SNAKE_CASE_ : Optional[int] = [self.sep_token_id] SCREAMING_SNAKE_CASE_ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __A ( self , __lowerCAmelCase , __lowerCAmelCase = None ): SCREAMING_SNAKE_CASE_ : List[Any] = self._tokenizer.model.save(__lowerCAmelCase , name=__lowerCAmelCase ) return tuple(__lowerCAmelCase )
345
"""simple docstring""" import os import re import shutil from argparse import ArgumentParser, Namespace from datasets.commands import BaseDatasetsCLICommand from datasets.utils.logging import get_logger lowerCAmelCase__ = '''<<<<<<< This should probably be modified because it mentions: ''' lowerCAmelCase__ = '''======= >>>>>>> ''' lowerCAmelCase__ = [ '''TextEncoderConfig''', '''ByteTextEncoder''', '''SubwordTextEncoder''', '''encoder_config''', '''maybe_build_from_corpus''', '''manual_dir''', ] lowerCAmelCase__ = [ # (pattern, replacement) # Order is important here for some replacements (R'''tfds\.core''', R'''datasets'''), (R'''tf\.io\.gfile\.GFile''', R'''open'''), (R'''tf\.([\w\d]+)''', R'''datasets.Value(\'\1\')'''), (R'''tfds\.features\.Text\(\)''', R'''datasets.Value(\'string\')'''), (R'''tfds\.features\.Text\(''', R'''datasets.Value(\'string\'),'''), (R'''features\s*=\s*tfds.features.FeaturesDict\(''', R'''features=datasets.Features('''), (R'''tfds\.features\.FeaturesDict\(''', R'''dict('''), (R'''The TensorFlow Datasets Authors''', R'''The TensorFlow Datasets Authors and the HuggingFace Datasets Authors'''), (R'''tfds\.''', R'''datasets.'''), (R'''dl_manager\.manual_dir''', R'''self.config.data_dir'''), (R'''self\.builder_config''', R'''self.config'''), ] def snake_case_ ( A_ : Namespace ): '''simple docstring''' return ConvertCommand(args.tfds_path, args.datasets_directory ) class __snake_case ( _lowercase): @staticmethod def SCREAMING_SNAKE_CASE ( __lowerCAmelCase : ArgumentParser ): """simple docstring""" _lowerCamelCase : List[str] = parser.add_parser( '''convert''' , help='''Convert a TensorFlow Datasets dataset to a HuggingFace Datasets dataset.''' , ) train_parser.add_argument( '''--tfds_path''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to a TensorFlow Datasets folder to convert or a single tfds file to convert.''' , ) train_parser.add_argument( '''--datasets_directory''' , type=__lowerCAmelCase , required=__lowerCAmelCase , help='''Path to the HuggingFace Datasets folder.''' ) train_parser.set_defaults(func=__lowerCAmelCase ) def __init__( self : str , __lowerCAmelCase : str , __lowerCAmelCase : str , *__lowerCAmelCase : int ): """simple docstring""" _lowerCamelCase : List[str] = get_logger('''datasets-cli/converting''' ) _lowerCamelCase : int = tfds_path _lowerCamelCase : Dict = datasets_directory def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" if os.path.isdir(self._tfds_path ): _lowerCamelCase : Union[str, Any] = os.path.abspath(self._tfds_path ) elif os.path.isfile(self._tfds_path ): _lowerCamelCase : Dict = os.path.dirname(self._tfds_path ) else: raise ValueError('''--tfds_path is neither a directory nor a file. Please check path.''' ) _lowerCamelCase : int = os.path.abspath(self._datasets_directory ) self._logger.info(f'''Converting datasets from {abs_tfds_path} to {abs_datasets_path}''' ) _lowerCamelCase : str = [] _lowerCamelCase : Union[str, Any] = [] _lowerCamelCase : Union[str, Any] = {} if os.path.isdir(self._tfds_path ): _lowerCamelCase : List[str] = os.listdir(__lowerCAmelCase ) else: _lowerCamelCase : Optional[Any] = [os.path.basename(self._tfds_path )] for f_name in file_names: self._logger.info(f'''Looking at file {f_name}''' ) _lowerCamelCase : Union[str, Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : Optional[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) if not os.path.isfile(__lowerCAmelCase ) or "__init__" in f_name or "_test" in f_name or ".py" not in f_name: self._logger.info('''Skipping file''' ) continue with open(__lowerCAmelCase , encoding='''utf-8''' ) as f: _lowerCamelCase : Tuple = f.readlines() _lowerCamelCase : Optional[int] = [] _lowerCamelCase : Union[str, Any] = False _lowerCamelCase : int = False _lowerCamelCase : Tuple = [] for line in lines: _lowerCamelCase : Optional[int] = line # Convert imports if "import tensorflow.compat.v2 as tf" in out_line: continue elif "@tfds.core" in out_line: continue elif "builder=self" in out_line: continue elif "import tensorflow_datasets.public_api as tfds" in out_line: _lowerCamelCase : Union[str, Any] = '''import datasets\n''' elif "import tensorflow" in out_line: # order is important here _lowerCamelCase : List[str] = '''''' continue elif "from absl import logging" in out_line: _lowerCamelCase : str = '''from datasets import logging\n''' elif "getLogger" in out_line: _lowerCamelCase : Union[str, Any] = out_line.replace('''getLogger''' , '''get_logger''' ) elif any(expression in out_line for expression in TO_HIGHLIGHT ): _lowerCamelCase : Dict = True _lowerCamelCase : Optional[int] = list(filter(lambda __lowerCAmelCase : e in out_line , __lowerCAmelCase ) ) out_lines.append(HIGHLIGHT_MESSAGE_PRE + str(__lowerCAmelCase ) + '''\n''' ) out_lines.append(__lowerCAmelCase ) out_lines.append(__lowerCAmelCase ) continue else: for pattern, replacement in TO_CONVERT: _lowerCamelCase : str = re.sub(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) # Take care of saving utilities (to later move them together with main script) if "tensorflow_datasets" in out_line: _lowerCamelCase : Dict = re.match(R'''from\stensorflow_datasets.*import\s([^\.\r\n]+)''' , __lowerCAmelCase ) tfds_imports.extend(imp.strip() for imp in match.group(1 ).split(''',''' ) ) _lowerCamelCase : Union[str, Any] = '''from . import ''' + match.group(1 ) # Check we have not forget anything if "tf." in out_line or "tfds." in out_line or "tensorflow_datasets" in out_line: raise ValueError(f'''Error converting {out_line.strip()}''' ) if "GeneratorBasedBuilder" in out_line or "BeamBasedBuilder" in out_line: _lowerCamelCase : Any = True out_lines.append(__lowerCAmelCase ) if is_builder or "wmt" in f_name: # We create a new directory for each dataset _lowerCamelCase : Union[str, Any] = f_name.replace('''.py''' , '''''' ) _lowerCamelCase : List[str] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = os.path.join(__lowerCAmelCase , __lowerCAmelCase ) os.makedirs(__lowerCAmelCase , exist_ok=__lowerCAmelCase ) self._logger.info(f'''Adding directory {output_dir}''' ) imports_to_builder_map.update({imp: output_dir for imp in tfds_imports} ) else: # Utilities will be moved at the end utils_files.append(__lowerCAmelCase ) if needs_manual_update: with_manual_update.append(__lowerCAmelCase ) with open(__lowerCAmelCase , '''w''' , encoding='''utf-8''' ) as f: f.writelines(__lowerCAmelCase ) self._logger.info(f'''Converted in {output_file}''' ) for utils_file in utils_files: try: _lowerCamelCase : Optional[int] = os.path.basename(__lowerCAmelCase ) _lowerCamelCase : Union[str, Any] = imports_to_builder_map[f_name.replace('''.py''' , '''''' )] self._logger.info(f'''Moving {dest_folder} to {utils_file}''' ) shutil.copy(__lowerCAmelCase , __lowerCAmelCase ) except KeyError: self._logger.error(f'''Cannot find destination folder for {utils_file}. Please copy manually.''' ) if with_manual_update: for file_path in with_manual_update: self._logger.warning( f'''You need to manually update file {file_path} to remove configurations using \'TextEncoderConfig\'.''' )
83
0
import random def _UpperCAmelCase ( A , A , A ): '''simple docstring''' UpperCAmelCase__ =a[left_index] UpperCAmelCase__ =left_index + 1 for j in range(left_index + 1 , A_ ): if a[j] < pivot: UpperCAmelCase__ =a[i], a[j] i += 1 UpperCAmelCase__ =a[i - 1], a[left_index] return i - 1 def _UpperCAmelCase ( A , A , A ): '''simple docstring''' if left < right: UpperCAmelCase__ =random.randint(A_ , right - 1 ) UpperCAmelCase__ =( a[left], a[pivot], ) # switches the pivot with the left most bound UpperCAmelCase__ =partition(A_ , A_ , A_ ) quick_sort_random( A_ , A_ , A_ ) # recursive quicksort to the left of the pivot point quick_sort_random( A_ , pivot_index + 1 , A_ ) # recursive quicksort to the right of the pivot point def _UpperCAmelCase ( ): '''simple docstring''' UpperCAmelCase__ =input("Enter numbers separated by a comma:\n" ).strip() UpperCAmelCase__ =[int(A_ ) for item in user_input.split("," )] quick_sort_random(A_ , 0 , len(A_ ) ) print(A_ ) if __name__ == "__main__": main()
625
"""simple docstring""" def snake_case_ ( A_ : list, A_ : list, A_ : int, A_ : int, A_ : int ): '''simple docstring''' if index == number_of_items: return 0 _lowerCamelCase : int = 0 _lowerCamelCase : str = 0 _lowerCamelCase : Dict = knapsack(A_, A_, A_, A_, index + 1 ) if weights[index] <= max_weight: _lowerCamelCase : Tuple = values[index] + knapsack( A_, A_, A_, max_weight - weights[index], index + 1 ) return max(A_, A_ ) if __name__ == "__main__": import doctest doctest.testmod()
83
0
from math import ceil from typing import List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import BatchFeature, SequenceFeatureExtractor from ...utils import TensorType, logging lowerCamelCase = logging.get_logger(__name__) class A ( _lowercase ): UpperCamelCase__ : Dict =["audio_values", "audio_mask"] def __init__( self : List[str] , lowercase_ : Optional[Any]=2048 , lowercase_ : Optional[int]=1 , lowercase_ : Dict=[16, 16] , lowercase_ : Optional[Any]=128 , lowercase_ : Optional[int]=4_4100 , lowercase_ : Optional[Any]=86 , lowercase_ : Dict=2048 , lowercase_ : Tuple=0.0 , **lowercase_ : Dict , ) -> Optional[int]: """simple docstring""" super().__init__( feature_size=__lowerCAmelCase , sampling_rate=__lowerCAmelCase , padding_value=__lowerCAmelCase , **__lowerCAmelCase , ) _lowerCamelCase : str =spectrogram_length _lowerCamelCase : List[str] =num_channels _lowerCamelCase : List[str] =patch_size _lowerCamelCase : Optional[Any] =feature_size // self.patch_size[1] _lowerCamelCase : Any =n_fft _lowerCamelCase : int =sampling_rate // hop_length_to_sampling_rate _lowerCamelCase : Optional[int] =sampling_rate _lowerCamelCase : Optional[Any] =padding_value _lowerCamelCase : str =mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__lowerCAmelCase , min_frequency=0.0 , max_frequency=2_2050.0 , sampling_rate=__lowerCAmelCase , norm='slaney' , mel_scale='slaney' , ).T def lowerCamelCase ( self : Optional[Any] , lowercase_ : np.array ) -> Union[str, Any]: """simple docstring""" _lowerCamelCase : Tuple =spectrogram( __lowerCAmelCase , window_function(self.n_fft , 'hann' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters.T , log_mel='dB' , db_range=80.0 , ) _lowerCamelCase : Union[str, Any] =log_spec[:, :-1] _lowerCamelCase : Optional[Any] =log_spec - 20.0 _lowerCamelCase : List[Any] =np.clip(log_spec / 40.0 , -2.0 , 0.0 ) + 1.0 return log_spec def __call__( self : Optional[Any] , lowercase_ : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , lowercase_ : Optional[Union[str, TensorType]] = None , lowercase_ : Optional[bool] = True , lowercase_ : Optional[int] = None , lowercase_ : bool = False , lowercase_ : bool = False , **lowercase_ : List[str] , ) -> Dict: """simple docstring""" if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( 'This feature extractor is set to support sampling rate' F''' of {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled''' F''' with {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) _lowerCamelCase : Union[str, Any] =isinstance(__lowerCAmelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'''Only mono-channel audio is supported for input to {self}''' ) _lowerCamelCase : List[str] =is_batched_numpy or ( isinstance(__lowerCAmelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _lowerCamelCase : str =[np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__lowerCAmelCase , np.ndarray ): _lowerCamelCase : Any =np.asarray(__lowerCAmelCase , dtype=np.floataa ) elif isinstance(__lowerCAmelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _lowerCamelCase : Union[str, Any] =raw_speech.astype(np.floataa ) # always return batch if not is_batched: _lowerCamelCase : Tuple =[np.asarray([raw_speech] ).T] # Convert audio signals to log mel spectrograms, truncate by time axis _lowerCamelCase : int =[ self._np_extract_fbank_features(waveform.squeeze() ).T[: self.spectrogram_length] for waveform in raw_speech ] if isinstance(audio_features[0] , __lowerCAmelCase ): _lowerCamelCase : Union[str, Any] =[np.asarray(__lowerCAmelCase , dtype=np.floataa ) for feature in audio_features] # Create audio attention mask _lowerCamelCase : List[Any] =max( [ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len for feature in audio_features] ) # The maximum number of audio patches in a batch if return_attention_mask: _lowerCamelCase : Any =[ (ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [1] + (max_patch_len - ceil(feature.shape[0] / self.patch_size[0] ) * self.freq_len) * [0] for feature in audio_features ] _lowerCamelCase : str =np.array(__lowerCAmelCase ).astype(np.floataa ) # convert into correct format for padding _lowerCamelCase : List[str] =max_patch_len // self.freq_len * self.patch_size[0] # The maximum audio size in a batch _lowerCamelCase : Optional[int] =np.ones([len(__lowerCAmelCase ), 1, max_time_len, self.feature_size] ).astype(np.floataa ) _lowerCamelCase : int =padded_audio_features * self.padding_value for i in range(len(__lowerCAmelCase ) ): _lowerCamelCase : List[str] =audio_features[i] _lowerCamelCase : Optional[Any] =feature # return as BatchFeature if return_attention_mask: _lowerCamelCase : Union[str, Any] ={'''audio_values''': padded_audio_features, '''audio_mask''': audio_mask} else: _lowerCamelCase : Any ={'''audio_values''': padded_audio_features} _lowerCamelCase : Optional[int] =BatchFeature(data=__lowerCAmelCase , tensor_type=__lowerCAmelCase ) return encoded_inputs
464
"""simple docstring""" from ..utils import DummyObject, requires_backends class __snake_case ( metaclass=_lowercase): snake_case__ : Optional[Any] = ["transformers", "torch", "note_seq"] def __init__( self : Union[str, Any] , *__lowerCAmelCase : Optional[Any] , **__lowerCAmelCase : List[Any] ): """simple docstring""" requires_backends(self , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : List[Any] , *__lowerCAmelCase : str , **__lowerCAmelCase : Union[str, Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] ) @classmethod def SCREAMING_SNAKE_CASE ( cls : int , *__lowerCAmelCase : List[str] , **__lowerCAmelCase : Optional[Any] ): """simple docstring""" requires_backends(cls , ['''transformers''', '''torch''', '''note_seq'''] )
83
0
'''simple docstring''' import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class _snake_case ( _lowercase ): lowerCAmelCase :Any = (DDPMScheduler,) def snake_case__ ( self , **_lowerCamelCase): UpperCAmelCase__ : Optional[int] = { '''num_train_timesteps''': 1000, '''beta_start''': 0.0001, '''beta_end''': 0.02, '''beta_schedule''': '''linear''', '''variance_type''': '''fixed_small''', '''clip_sample''': True, } config.update(**__lowerCAmelCase) return config def snake_case__ ( self): for timesteps in [1, 5, 100, 1000]: self.check_over_configs(num_train_timesteps=__lowerCAmelCase) def snake_case__ ( self): for beta_start, beta_end in zip([0.0001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2]): self.check_over_configs(beta_start=__lowerCAmelCase , beta_end=__lowerCAmelCase) def snake_case__ ( self): for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=__lowerCAmelCase) def snake_case__ ( self): for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=__lowerCAmelCase) def snake_case__ ( self): for clip_sample in [True, False]: self.check_over_configs(clip_sample=__lowerCAmelCase) def snake_case__ ( self): self.check_over_configs(thresholding=__lowerCAmelCase) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=__lowerCAmelCase , prediction_type=__lowerCAmelCase , sample_max_value=__lowerCAmelCase , ) def snake_case__ ( self): for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=__lowerCAmelCase) def snake_case__ ( self): for t in [0, 500, 999]: self.check_over_forward(time_step=__lowerCAmelCase) def snake_case__ ( self): UpperCAmelCase__ : Tuple = self.scheduler_classes[0] UpperCAmelCase__ : int = self.get_scheduler_config() UpperCAmelCase__ : Any = scheduler_class(**__lowerCAmelCase) assert torch.sum(torch.abs(scheduler._get_variance(0) - 0.0)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(487) - 0.00979)) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(999) - 0.02)) < 1e-5 def snake_case__ ( self): UpperCAmelCase__ : str = self.scheduler_classes[0] UpperCAmelCase__ : List[Any] = self.get_scheduler_config() UpperCAmelCase__ : Optional[int] = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : Tuple = len(__lowerCAmelCase) UpperCAmelCase__ : Dict = self.dummy_model() UpperCAmelCase__ : str = self.dummy_sample_deter UpperCAmelCase__ : Any = torch.manual_seed(0) for t in reversed(range(__lowerCAmelCase)): # 1. predict noise residual UpperCAmelCase__ : Optional[int] = model(__lowerCAmelCase , __lowerCAmelCase) # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ : Dict = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , generator=__lowerCAmelCase).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance UpperCAmelCase__ : Dict = pred_prev_sample UpperCAmelCase__ : str = torch.sum(torch.abs(__lowerCAmelCase)) UpperCAmelCase__ : str = torch.mean(torch.abs(__lowerCAmelCase)) assert abs(result_sum.item() - 258.9606) < 1e-2 assert abs(result_mean.item() - 0.3372) < 1e-3 def snake_case__ ( self): UpperCAmelCase__ : List[str] = self.scheduler_classes[0] UpperCAmelCase__ : Union[str, Any] = self.get_scheduler_config(prediction_type="""v_prediction""") UpperCAmelCase__ : Union[str, Any] = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : Dict = len(__lowerCAmelCase) UpperCAmelCase__ : str = self.dummy_model() UpperCAmelCase__ : Optional[Any] = self.dummy_sample_deter UpperCAmelCase__ : Dict = torch.manual_seed(0) for t in reversed(range(__lowerCAmelCase)): # 1. predict noise residual UpperCAmelCase__ : Union[str, Any] = model(__lowerCAmelCase , __lowerCAmelCase) # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ : List[Any] = scheduler.step(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , generator=__lowerCAmelCase).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance UpperCAmelCase__ : Any = pred_prev_sample UpperCAmelCase__ : List[str] = torch.sum(torch.abs(__lowerCAmelCase)) UpperCAmelCase__ : str = torch.mean(torch.abs(__lowerCAmelCase)) assert abs(result_sum.item() - 202.0296) < 1e-2 assert abs(result_mean.item() - 0.2631) < 1e-3 def snake_case__ ( self): UpperCAmelCase__ : List[str] = self.scheduler_classes[0] UpperCAmelCase__ : List[str] = self.get_scheduler_config() UpperCAmelCase__ : Dict = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : List[str] = [100, 87, 50, 1, 0] scheduler.set_timesteps(timesteps=__lowerCAmelCase) UpperCAmelCase__ : Union[str, Any] = scheduler.timesteps for i, timestep in enumerate(__lowerCAmelCase): if i == len(__lowerCAmelCase) - 1: UpperCAmelCase__ : str = -1 else: UpperCAmelCase__ : List[Any] = timesteps[i + 1] UpperCAmelCase__ : Optional[int] = scheduler.previous_timestep(__lowerCAmelCase) UpperCAmelCase__ : Tuple = prev_t.item() self.assertEqual(__lowerCAmelCase , __lowerCAmelCase) def snake_case__ ( self): UpperCAmelCase__ : str = self.scheduler_classes[0] UpperCAmelCase__ : Optional[int] = self.get_scheduler_config() UpperCAmelCase__ : str = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : Any = [100, 87, 50, 51, 0] with self.assertRaises(__lowerCAmelCase , msg="""`custom_timesteps` must be in descending order."""): scheduler.set_timesteps(timesteps=__lowerCAmelCase) def snake_case__ ( self): UpperCAmelCase__ : Tuple = self.scheduler_classes[0] UpperCAmelCase__ : Optional[Any] = self.get_scheduler_config() UpperCAmelCase__ : List[str] = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : List[Any] = [100, 87, 50, 1, 0] UpperCAmelCase__ : List[str] = len(__lowerCAmelCase) with self.assertRaises(__lowerCAmelCase , msg="""Can only pass one of `num_inference_steps` or `custom_timesteps`."""): scheduler.set_timesteps(num_inference_steps=__lowerCAmelCase , timesteps=__lowerCAmelCase) def snake_case__ ( self): UpperCAmelCase__ : Optional[Any] = self.scheduler_classes[0] UpperCAmelCase__ : str = self.get_scheduler_config() UpperCAmelCase__ : Optional[Any] = scheduler_class(**__lowerCAmelCase) UpperCAmelCase__ : int = [scheduler.config.num_train_timesteps] with self.assertRaises( __lowerCAmelCase , msg="""`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}""" , ): scheduler.set_timesteps(timesteps=__lowerCAmelCase)
407
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class __snake_case ( unittest.TestCase): @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : Any = FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Any = AutoTokenizer.from_pretrained('''google/mt5-small''' ) _lowerCamelCase : Union[str, Any] = tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids _lowerCamelCase : Optional[int] = tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids _lowerCamelCase : List[Any] = shift_tokens_right(__lowerCAmelCase , model.config.pad_token_id , model.config.decoder_start_token_id ) _lowerCamelCase : int = model(__lowerCAmelCase , decoder_input_ids=__lowerCAmelCase ).logits _lowerCamelCase : Optional[Any] = optax.softmax_cross_entropy(__lowerCAmelCase , onehot(__lowerCAmelCase , logits.shape[-1] ) ).mean() _lowerCamelCase : Dict = -(labels.shape[-1] * loss.item()) _lowerCamelCase : Dict = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
83
0
'''simple docstring''' import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class UpperCAmelCase ( unittest.TestCase): """simple docstring""" @slow def UpperCamelCase__ ( self : Optional[int] ) -> Any: _UpperCamelCase =FlaxMTaForConditionalGeneration.from_pretrained('''google/mt5-small''' ) _UpperCamelCase =AutoTokenizer.from_pretrained('''google/mt5-small''' ) _UpperCamelCase =tokenizer('''Hello there''' , return_tensors='''np''' ).input_ids _UpperCamelCase =tokenizer('''Hi I am''' , return_tensors='''np''' ).input_ids _UpperCamelCase =shift_tokens_right(__lowerCAmelCase , model.config.pad_token_id , model.config.decoder_start_token_id ) _UpperCamelCase =model(__lowerCAmelCase , decoder_input_ids=__lowerCAmelCase ).logits _UpperCamelCase =optax.softmax_cross_entropy(__lowerCAmelCase , onehot(__lowerCAmelCase , logits.shape[-1] ) ).mean() _UpperCamelCase =-(labels.shape[-1] * loss.item()) _UpperCamelCase =-84.9127 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
404
"""simple docstring""" import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( '''files''', [ ['''full:README.md''', '''dataset_infos.json'''], ['''empty:README.md''', '''dataset_infos.json'''], ['''dataset_infos.json'''], ['''full:README.md'''], ], ) def snake_case_ ( A_ : Dict, A_ : List[str] ): '''simple docstring''' _lowerCamelCase : int = tmp_path_factory.mktemp('''dset_infos_dir''' ) if "full:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''---\ndataset_info:\n dataset_size: 42\n---''' ) if "empty:README.md" in files: with open(dataset_infos_dir / '''README.md''', '''w''' ) as f: f.write('''''' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / '''dataset_infos.json''', '''w''' ) as f: f.write('''{"default": {"dataset_size": 42}}''' ) _lowerCamelCase : str = DatasetInfosDict.from_directory(A_ ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( '''dataset_info''', [ DatasetInfo(), DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ), ], ) def snake_case_ ( A_ : str, A_ : DatasetInfo ): '''simple docstring''' _lowerCamelCase : Optional[Any] = str(A_ ) dataset_info.write_to_directory(A_ ) _lowerCamelCase : str = DatasetInfo.from_directory(A_ ) assert dataset_info == reloaded assert os.path.exists(os.path.join(A_, '''dataset_info.json''' ) ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : Union[str, Any] = DatasetInfo( description='''foo''', citation='''bar''', homepage='''https://foo.bar''', license='''CC0''', features=Features({'''a''': Value('''int32''' )} ), post_processed={}, supervised_keys=(), task_templates=[], builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train''', '''num_examples''': 42}], download_checksums={}, download_size=13_37, post_processing_size=4_42, dataset_size=12_34, size_in_bytes=13_37 + 4_42 + 12_34, ) _lowerCamelCase : Optional[Any] = dataset_info._to_yaml_dict() assert sorted(A_ ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key], (list, dict, int, str) ) _lowerCamelCase : str = yaml.safe_dump(A_ ) _lowerCamelCase : Tuple = yaml.safe_load(A_ ) assert dataset_info_yaml_dict == reloaded def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : int = DatasetInfo() _lowerCamelCase : Dict = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( '''dataset_infos_dict''', [ DatasetInfosDict(), DatasetInfosDict({'''default''': DatasetInfo()} ), DatasetInfosDict({'''my_config_name''': DatasetInfo()} ), DatasetInfosDict( { '''default''': DatasetInfo( description='''foo''', features=Features({'''a''': Value('''int32''' )} ), builder_name='''builder''', config_name='''config''', version='''1.0.0''', splits=[{'''name''': '''train'''}], download_size=42, ) } ), DatasetInfosDict( { '''v1''': DatasetInfo(dataset_size=42 ), '''v2''': DatasetInfo(dataset_size=13_37 ), } ), ], ) def snake_case_ ( A_ : Optional[Any], A_ : DatasetInfosDict ): '''simple docstring''' _lowerCamelCase : List[str] = str(A_ ) dataset_infos_dict.write_to_directory(A_ ) _lowerCamelCase : List[Any] = DatasetInfosDict.from_directory(A_ ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): _lowerCamelCase : str = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml _lowerCamelCase : Any = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(A_, '''README.md''' ) )
83
0
import argparse import os import torch from diffusers import ( CMStochasticIterativeScheduler, ConsistencyModelPipeline, UNetaDModel, ) lowerCamelCase__ = { '''sample_size''': 32, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': 10_00, '''block_out_channels''': [32, 64], '''attention_head_dim''': 8, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCamelCase__ = { '''sample_size''': 64, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 3, '''num_class_embeds''': 10_00, '''block_out_channels''': [1_92, 1_92 * 2, 1_92 * 3, 1_92 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''scale_shift''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCamelCase__ = { '''sample_size''': 2_56, '''in_channels''': 3, '''out_channels''': 3, '''layers_per_block''': 2, '''num_class_embeds''': None, '''block_out_channels''': [2_56, 2_56, 2_56 * 2, 2_56 * 2, 2_56 * 4, 2_56 * 4], '''attention_head_dim''': 64, '''down_block_types''': [ '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''ResnetDownsampleBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', '''AttnDownBlock2D''', ], '''up_block_types''': [ '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''AttnUpBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', '''ResnetUpsampleBlock2D''', ], '''resnet_time_scale_shift''': '''default''', '''upsample_type''': '''resnet''', '''downsample_type''': '''resnet''', } lowerCamelCase__ = { '''num_train_timesteps''': 40, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowerCamelCase__ = { '''num_train_timesteps''': 2_01, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } lowerCamelCase__ = { '''num_train_timesteps''': 1_51, '''sigma_min''': 0.002, '''sigma_max''': 80.0, } def A(__a: List[str] ): if isinstance(A_ , A_ ): return v if v.lower() in ("yes", "true", "t", "y", "1"): return True elif v.lower() in ("no", "false", "f", "n", "0"): return False else: raise argparse.ArgumentTypeError("boolean value expected" ) def A(__a: Union[str, Any] , __a: Optional[Any] , __a: Tuple , __a: Dict , __a: Optional[Any]=False ): lowerCAmelCase_ = checkpoint[F"{old_prefix}.in_layers.0.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.in_layers.0.bias"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.in_layers.2.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.in_layers.2.bias"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.emb_layers.1.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.emb_layers.1.bias"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.out_layers.0.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.out_layers.0.bias"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.out_layers.3.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.out_layers.3.bias"] if has_skip: lowerCAmelCase_ = checkpoint[F"{old_prefix}.skip_connection.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.skip_connection.bias"] return new_checkpoint def A(__a: Optional[int] , __a: Tuple , __a: Union[str, Any] , __a: Tuple , __a: str=None ): lowerCAmelCase_ = checkpoint[F"{old_prefix}.qkv.weight"].chunk(3 , dim=0 ) lowerCAmelCase_ = checkpoint[F"{old_prefix}.qkv.bias"].chunk(3 , dim=0 ) lowerCAmelCase_ = checkpoint[F"{old_prefix}.norm.weight"] lowerCAmelCase_ = checkpoint[F"{old_prefix}.norm.bias"] lowerCAmelCase_ = weight_q.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = bias_q.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = weight_k.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = bias_k.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = weight_v.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = bias_v.squeeze(-1 ).squeeze(-1 ) lowerCAmelCase_ = ( checkpoint[F"{old_prefix}.proj_out.weight"].squeeze(-1 ).squeeze(-1 ) ) lowerCAmelCase_ = checkpoint[F"{old_prefix}.proj_out.bias"].squeeze(-1 ).squeeze(-1 ) return new_checkpoint def A(__a: str , __a: Union[str, Any] ): lowerCAmelCase_ = torch.load(A_ , map_location="cpu" ) lowerCAmelCase_ = {} lowerCAmelCase_ = checkpoint['''time_embed.0.weight'''] lowerCAmelCase_ = checkpoint['''time_embed.0.bias'''] lowerCAmelCase_ = checkpoint['''time_embed.2.weight'''] lowerCAmelCase_ = checkpoint['''time_embed.2.bias'''] if unet_config["num_class_embeds"] is not None: lowerCAmelCase_ = checkpoint['''label_emb.weight'''] lowerCAmelCase_ = checkpoint['''input_blocks.0.0.weight'''] lowerCAmelCase_ = checkpoint['''input_blocks.0.0.bias'''] lowerCAmelCase_ = unet_config['''down_block_types'''] lowerCAmelCase_ = unet_config['''layers_per_block'''] lowerCAmelCase_ = unet_config['''attention_head_dim'''] lowerCAmelCase_ = unet_config['''block_out_channels'''] lowerCAmelCase_ = 1 lowerCAmelCase_ = channels_list[0] for i, layer_type in enumerate(A_ ): lowerCAmelCase_ = channels_list[i] lowerCAmelCase_ = current_channels != prev_channels if layer_type == "ResnetDownsampleBlock2D": for j in range(A_ ): lowerCAmelCase_ = F"down_blocks.{i}.resnets.{j}" lowerCAmelCase_ = F"input_blocks.{current_layer}.0" lowerCAmelCase_ = True if j == 0 and downsample_block_has_skip else False lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ , has_skip=A_ ) current_layer += 1 elif layer_type == "AttnDownBlock2D": for j in range(A_ ): lowerCAmelCase_ = F"down_blocks.{i}.resnets.{j}" lowerCAmelCase_ = F"input_blocks.{current_layer}.0" lowerCAmelCase_ = True if j == 0 and downsample_block_has_skip else False lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ , has_skip=A_ ) lowerCAmelCase_ = F"down_blocks.{i}.attentions.{j}" lowerCAmelCase_ = F"input_blocks.{current_layer}.1" lowerCAmelCase_ = convert_attention( A_ , A_ , A_ , A_ , A_ ) current_layer += 1 if i != len(A_ ) - 1: lowerCAmelCase_ = F"down_blocks.{i}.downsamplers.0" lowerCAmelCase_ = F"input_blocks.{current_layer}.0" lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ ) current_layer += 1 lowerCAmelCase_ = current_channels # hardcoded the mid-block for now lowerCAmelCase_ = '''mid_block.resnets.0''' lowerCAmelCase_ = '''middle_block.0''' lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ ) lowerCAmelCase_ = '''mid_block.attentions.0''' lowerCAmelCase_ = '''middle_block.1''' lowerCAmelCase_ = convert_attention(A_ , A_ , A_ , A_ , A_ ) lowerCAmelCase_ = '''mid_block.resnets.1''' lowerCAmelCase_ = '''middle_block.2''' lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ ) lowerCAmelCase_ = 0 lowerCAmelCase_ = unet_config['''up_block_types'''] for i, layer_type in enumerate(A_ ): if layer_type == "ResnetUpsampleBlock2D": for j in range(layers_per_block + 1 ): lowerCAmelCase_ = F"up_blocks.{i}.resnets.{j}" lowerCAmelCase_ = F"output_blocks.{current_layer}.0" lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ , has_skip=A_ ) current_layer += 1 if i != len(A_ ) - 1: lowerCAmelCase_ = F"up_blocks.{i}.upsamplers.0" lowerCAmelCase_ = F"output_blocks.{current_layer-1}.1" lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ ) elif layer_type == "AttnUpBlock2D": for j in range(layers_per_block + 1 ): lowerCAmelCase_ = F"up_blocks.{i}.resnets.{j}" lowerCAmelCase_ = F"output_blocks.{current_layer}.0" lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ , has_skip=A_ ) lowerCAmelCase_ = F"up_blocks.{i}.attentions.{j}" lowerCAmelCase_ = F"output_blocks.{current_layer}.1" lowerCAmelCase_ = convert_attention( A_ , A_ , A_ , A_ , A_ ) current_layer += 1 if i != len(A_ ) - 1: lowerCAmelCase_ = F"up_blocks.{i}.upsamplers.0" lowerCAmelCase_ = F"output_blocks.{current_layer-1}.2" lowerCAmelCase_ = convert_resnet(A_ , A_ , A_ , A_ ) lowerCAmelCase_ = checkpoint['''out.0.weight'''] lowerCAmelCase_ = checkpoint['''out.0.bias'''] lowerCAmelCase_ = checkpoint['''out.2.weight'''] lowerCAmelCase_ = checkpoint['''out.2.bias'''] return new_checkpoint if __name__ == "__main__": lowerCamelCase__ = argparse.ArgumentParser() parser.add_argument('''--unet_path''', default=None, type=str, required=True, help='''Path to the unet.pt to convert.''') parser.add_argument( '''--dump_path''', default=None, type=str, required=True, help='''Path to output the converted UNet model.''' ) parser.add_argument('''--class_cond''', default=True, type=str, help='''Whether the model is class-conditional.''') lowerCamelCase__ = parser.parse_args() lowerCamelCase__ = strabool(args.class_cond) lowerCamelCase__ = os.path.basename(args.unet_path) print(F'''Checkpoint: {ckpt_name}''') # Get U-Net config if "imagenet64" in ckpt_name: lowerCamelCase__ = IMAGENET_64_UNET_CONFIG elif "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCamelCase__ = LSUN_256_UNET_CONFIG elif "test" in ckpt_name: lowerCamelCase__ = TEST_UNET_CONFIG else: raise ValueError(F'''Checkpoint type {ckpt_name} is not currently supported.''') if not args.class_cond: lowerCamelCase__ = None lowerCamelCase__ = con_pt_to_diffuser(args.unet_path, unet_config) lowerCamelCase__ = UNetaDModel(**unet_config) image_unet.load_state_dict(converted_unet_ckpt) # Get scheduler config if "cd" in ckpt_name or "test" in ckpt_name: lowerCamelCase__ = CD_SCHEDULER_CONFIG elif "ct" in ckpt_name and "imagenet64" in ckpt_name: lowerCamelCase__ = CT_IMAGENET_64_SCHEDULER_CONFIG elif "ct" in ckpt_name and "256" in ckpt_name and (("bedroom" in ckpt_name) or ("cat" in ckpt_name)): lowerCamelCase__ = CT_LSUN_256_SCHEDULER_CONFIG else: raise ValueError(F'''Checkpoint type {ckpt_name} is not currently supported.''') lowerCamelCase__ = CMStochasticIterativeScheduler(**scheduler_config) lowerCamelCase__ = ConsistencyModelPipeline(unet=image_unet, scheduler=cm_scheduler) consistency_model.save_pretrained(args.dump_path)
122
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __snake_case : def __init__( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple=1_3 , __lowerCAmelCase : Any=2 , __lowerCAmelCase : List[str]=2_4 , __lowerCAmelCase : str=1_6 , __lowerCAmelCase : List[Any]=True , __lowerCAmelCase : Union[str, Any]=True , __lowerCAmelCase : Optional[Any]=3_2 , __lowerCAmelCase : List[Any]=5 , __lowerCAmelCase : int=4 , __lowerCAmelCase : int=3_7 , __lowerCAmelCase : Union[str, Any]="gelu" , __lowerCAmelCase : Dict=0.1 , __lowerCAmelCase : str=0.1 , __lowerCAmelCase : int=1_0 , __lowerCAmelCase : List[Any]=0.02 , __lowerCAmelCase : str=None , __lowerCAmelCase : List[str]=2 , __lowerCAmelCase : Union[str, Any]=2 , ): """simple docstring""" _lowerCamelCase : List[str] = parent _lowerCamelCase : str = batch_size _lowerCamelCase : Tuple = patch_size _lowerCamelCase : Optional[int] = max_length _lowerCamelCase : List[Any] = num_mel_bins _lowerCamelCase : int = is_training _lowerCamelCase : Union[str, Any] = use_labels _lowerCamelCase : Dict = hidden_size _lowerCamelCase : Tuple = num_hidden_layers _lowerCamelCase : int = num_attention_heads _lowerCamelCase : Tuple = intermediate_size _lowerCamelCase : List[str] = hidden_act _lowerCamelCase : Dict = hidden_dropout_prob _lowerCamelCase : int = attention_probs_dropout_prob _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : List[str] = scope _lowerCamelCase : Optional[int] = frequency_stride _lowerCamelCase : List[Any] = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) _lowerCamelCase : Optional[int] = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 _lowerCamelCase : Union[str, Any] = (self.max_length - self.patch_size) // self.time_stride + 1 _lowerCamelCase : Any = frequency_out_dimension * time_out_dimension _lowerCamelCase : List[Any] = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : int = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) _lowerCamelCase : str = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Optional[int] = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__lowerCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : List[Any] = ASTModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : List[Any] = model(__lowerCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" _lowerCamelCase : int = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Optional[Any] = config_and_inputs _lowerCamelCase : int = {'''input_values''': input_values} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) snake_case__ : Tuple = ( {"audio-classification": ASTForAudioClassification, "feature-extraction": ASTModel} if is_torch_available() else {} ) snake_case__ : Any = False snake_case__ : List[Any] = False snake_case__ : Optional[Any] = False snake_case__ : Optional[Any] = False def SCREAMING_SNAKE_CASE ( self : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : List[str] , __lowerCAmelCase : Optional[Any] ): """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[int] = ASTModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , has_text_modality=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" pass def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : Dict = model_class(__lowerCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) _lowerCamelCase : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__lowerCAmelCase , nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase , _lowerCamelCase : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _lowerCamelCase : str = model_class(__lowerCAmelCase ) _lowerCamelCase : Tuple = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _lowerCamelCase : Any = [*signature.parameters.keys()] _lowerCamelCase : str = ['''input_values'''] self.assertListEqual(arg_names[:1] , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Union[str, Any] = ASTModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) def snake_case_ ( ): '''simple docstring''' _lowerCamelCase : List[str] = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''', filename='''sample_audio.flac''', repo_type='''dataset''' ) _lowerCamelCase , _lowerCamelCase : str = torchaudio.load(A_ ) return audio, sampling_rate @require_torch @require_torchaudio class __snake_case ( unittest.TestCase): @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" _lowerCamelCase : int = self.default_feature_extractor _lowerCamelCase : Union[str, Any] = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(__lowerCAmelCase ) _lowerCamelCase : List[Any] = self.default_feature_extractor _lowerCamelCase , _lowerCamelCase : List[Any] = prepare_audio() _lowerCamelCase : Dict = audio.squeeze().numpy() _lowerCamelCase : Tuple = feature_extractor(__lowerCAmelCase , sampling_rate=__lowerCAmelCase , return_tensors='''pt''' ).to(__lowerCAmelCase ) # forward pass with torch.no_grad(): _lowerCamelCase : Tuple = model(**__lowerCAmelCase ) # verify the logits _lowerCamelCase : Tuple = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , __lowerCAmelCase ) _lowerCamelCase : Optional[int] = torch.tensor([-0.87_60, -7.00_42, -8.66_02] ).to(__lowerCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __lowerCAmelCase , atol=1E-4 ) )
83
0
"""simple docstring""" import argparse from collections import defaultdict import yaml _lowerCAmelCase = """docs/source/en/_toctree.yml""" def lowerCamelCase__ ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : int = defaultdict(A_ ) for doc in model_doc: counts[doc["local"]] += 1 _lowerCAmelCase : List[Any] = [key for key, value in counts.items() if value > 1] _lowerCAmelCase : List[str] = [] for duplicate_key in duplicates: _lowerCAmelCase : Dict = list({doc['title'] for doc in model_doc if doc['local'] == duplicate_key} ) if len(A_ ) > 1: raise ValueError( f"""{duplicate_key} is present several times in the documentation table of content at """ '`docs/source/en/_toctree.yml` with different *Title* values. Choose one of those and remove the ' 'others.' ) # Only add this once new_doc.append({'local': duplicate_key, 'title': titles[0]} ) # Add none duplicate-keys new_doc.extend([doc for doc in model_doc if counts[doc['local']] == 1] ) # Sort return sorted(A_ , key=lambda _lowerCamelCase : s["title"].lower() ) def lowerCamelCase__ ( _lowerCamelCase=False ): '''simple docstring''' with open(A_ , encoding='utf-8' ) as f: _lowerCAmelCase : Dict = yaml.safe_load(f.read() ) # Get to the API doc _lowerCAmelCase : List[str] = 0 while content[api_idx]["title"] != "API": api_idx += 1 _lowerCAmelCase : Any = content[api_idx]['''sections'''] # Then to the model doc _lowerCAmelCase : str = 0 while api_doc[model_idx]["title"] != "Models": model_idx += 1 _lowerCAmelCase : List[str] = api_doc[model_idx]['''sections'''] _lowerCAmelCase : Tuple = [(idx, section) for idx, section in enumerate(A_ ) if '''sections''' in section] _lowerCAmelCase : Any = False for idx, modality_doc in modalities_docs: _lowerCAmelCase : str = modality_doc['''sections'''] _lowerCAmelCase : List[str] = clean_model_doc_toc(A_ ) if old_modality_doc != new_modality_doc: _lowerCAmelCase : Any = True if overwrite: _lowerCAmelCase : Optional[int] = new_modality_doc if diff: if overwrite: _lowerCAmelCase : Any = model_doc _lowerCAmelCase : Optional[Any] = api_doc with open(A_ , 'w' , encoding='utf-8' ) as f: f.write(yaml.dump(A_ , allow_unicode=A_ ) ) else: raise ValueError( 'The model doc part of the table of content is not properly sorted, run `make style` to fix this.' ) if __name__ == "__main__": _lowerCAmelCase = argparse.ArgumentParser() parser.add_argument("""--fix_and_overwrite""", action="""store_true""", help="""Whether to fix inconsistencies.""") _lowerCAmelCase = parser.parse_args() check_model_doc(args.fix_and_overwrite)
259
"""simple docstring""" import math import random from typing import Any from .hill_climbing import SearchProblem def snake_case_ ( A_ : Dict, A_ : bool = True, A_ : float = math.inf, A_ : float = -math.inf, A_ : float = math.inf, A_ : float = -math.inf, A_ : bool = False, A_ : float = 1_00, A_ : float = 0.01, A_ : float = 1, ): '''simple docstring''' _lowerCamelCase : Optional[int] = False _lowerCamelCase : str = search_prob _lowerCamelCase : str = start_temperate _lowerCamelCase : Optional[Any] = [] _lowerCamelCase : int = 0 _lowerCamelCase : Any = None while not search_end: _lowerCamelCase : Dict = current_state.score() if best_state is None or current_score > best_state.score(): _lowerCamelCase : Tuple = current_state scores.append(A_ ) iterations += 1 _lowerCamelCase : List[Any] = None _lowerCamelCase : Optional[int] = current_state.get_neighbors() while ( next_state is None and neighbors ): # till we do not find a neighbor that we can move to _lowerCamelCase : List[Any] = random.randint(0, len(A_ ) - 1 ) # picking a random neighbor _lowerCamelCase : Dict = neighbors.pop(A_ ) _lowerCamelCase : Union[str, Any] = picked_neighbor.score() - current_score if ( picked_neighbor.x > max_x or picked_neighbor.x < min_x or picked_neighbor.y > max_y or picked_neighbor.y < min_y ): continue # neighbor outside our bounds if not find_max: _lowerCamelCase : str = change * -1 # in case we are finding minimum if change > 0: # improves the solution _lowerCamelCase : Optional[Any] = picked_neighbor else: _lowerCamelCase : Optional[int] = (math.e) ** ( change / current_temp ) # probability generation function if random.random() < probability: # random number within probability _lowerCamelCase : Union[str, Any] = picked_neighbor _lowerCamelCase : List[str] = current_temp - (current_temp * rate_of_decrease) if current_temp < threshold_temp or next_state is None: # temperature below threshold, or could not find a suitable neighbor _lowerCamelCase : Tuple = True else: _lowerCamelCase : Optional[Any] = next_state if visualization: from matplotlib import pyplot as plt plt.plot(range(A_ ), A_ ) plt.xlabel('''Iterations''' ) plt.ylabel('''Function values''' ) plt.show() return best_state if __name__ == "__main__": def snake_case_ ( A_ : int, A_ : Tuple ): '''simple docstring''' return (x**2) + (y**2) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=False, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The minimum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) # starting the problem with initial coordinates (12, 47) lowerCAmelCase__ = SearchProblem(x=12, y=47, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing( prob, find_max=True, max_x=100, min_x=5, max_y=50, min_y=-5, visualization=True ) print( '''The maximum score for f(x, y) = x^2 + y^2 with the domain 100 > x > 5 ''' F"""and 50 > y > - 5 found via hill climbing: {local_min.score()}""" ) def snake_case_ ( A_ : Optional[int], A_ : List[Any] ): '''simple docstring''' return (3 * x**2) - (6 * y) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=False, visualization=True) print( '''The minimum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" ) lowerCAmelCase__ = SearchProblem(x=3, y=4, step_size=1, function_to_optimize=test_fa) lowerCAmelCase__ = simulated_annealing(prob, find_max=True, visualization=True) print( '''The maximum score for f(x, y) = 3*x^2 - 6*y found via hill climbing: ''' F"""{local_min.score()}""" )
83
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING a_ : List[Any] = logging.get_logger(__name__) a_ : Optional[Any] = { "microsoft/table-transformer-detection": ( "https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json" ), } class UpperCamelCase ( _lowercase ): __UpperCamelCase ="table-transformer" __UpperCamelCase =["past_key_values"] __UpperCamelCase ={ "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self : List[Any] , snake_case__ : Optional[int]=True , snake_case__ : List[Any]=None , snake_case__ : List[Any]=3 , snake_case__ : Any=1_0_0 , snake_case__ : Optional[Any]=6 , snake_case__ : Optional[Any]=2_0_4_8 , snake_case__ : Any=8 , snake_case__ : List[str]=6 , snake_case__ : int=2_0_4_8 , snake_case__ : Optional[int]=8 , snake_case__ : Any=0.0 , snake_case__ : str=0.0 , snake_case__ : List[Any]=True , snake_case__ : List[str]="relu" , snake_case__ : List[str]=2_5_6 , snake_case__ : int=0.1 , snake_case__ : Tuple=0.0 , snake_case__ : Optional[int]=0.0 , snake_case__ : Optional[int]=0.02 , snake_case__ : List[Any]=1.0 , snake_case__ : Tuple=False , snake_case__ : List[str]="sine" , snake_case__ : Any="resnet50" , snake_case__ : Optional[Any]=True , snake_case__ : Union[str, Any]=False , snake_case__ : Any=1 , snake_case__ : Any=5 , snake_case__ : Optional[int]=2 , snake_case__ : List[Any]=1 , snake_case__ : Tuple=1 , snake_case__ : List[Any]=5 , snake_case__ : int=2 , snake_case__ : Dict=0.1 , **snake_case__ : Tuple , ): """simple docstring""" if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) SCREAMING_SNAKE_CASE = CONFIG_MAPPING['''resnet'''](out_features=['stage4'] ) elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): SCREAMING_SNAKE_CASE = backbone_config.get('model_type' ) SCREAMING_SNAKE_CASE = CONFIG_MAPPING[backbone_model_type] SCREAMING_SNAKE_CASE = config_class.from_dict(__lowerCAmelCase ) # set timm attributes to None SCREAMING_SNAKE_CASE = None, None, None SCREAMING_SNAKE_CASE = use_timm_backbone SCREAMING_SNAKE_CASE = backbone_config SCREAMING_SNAKE_CASE = num_channels SCREAMING_SNAKE_CASE = num_queries SCREAMING_SNAKE_CASE = d_model SCREAMING_SNAKE_CASE = encoder_ffn_dim SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = encoder_attention_heads SCREAMING_SNAKE_CASE = decoder_ffn_dim SCREAMING_SNAKE_CASE = decoder_layers SCREAMING_SNAKE_CASE = decoder_attention_heads SCREAMING_SNAKE_CASE = dropout SCREAMING_SNAKE_CASE = attention_dropout SCREAMING_SNAKE_CASE = activation_dropout SCREAMING_SNAKE_CASE = activation_function SCREAMING_SNAKE_CASE = init_std SCREAMING_SNAKE_CASE = init_xavier_std SCREAMING_SNAKE_CASE = encoder_layerdrop SCREAMING_SNAKE_CASE = decoder_layerdrop SCREAMING_SNAKE_CASE = encoder_layers SCREAMING_SNAKE_CASE = auxiliary_loss SCREAMING_SNAKE_CASE = position_embedding_type SCREAMING_SNAKE_CASE = backbone SCREAMING_SNAKE_CASE = use_pretrained_backbone SCREAMING_SNAKE_CASE = dilation # Hungarian matcher SCREAMING_SNAKE_CASE = class_cost SCREAMING_SNAKE_CASE = bbox_cost SCREAMING_SNAKE_CASE = giou_cost # Loss coefficients SCREAMING_SNAKE_CASE = mask_loss_coefficient SCREAMING_SNAKE_CASE = dice_loss_coefficient SCREAMING_SNAKE_CASE = bbox_loss_coefficient SCREAMING_SNAKE_CASE = giou_loss_coefficient SCREAMING_SNAKE_CASE = eos_coefficient super().__init__(is_encoder_decoder=__lowerCAmelCase , **__lowerCAmelCase ) @property def UpperCamelCase ( self : Tuple ): """simple docstring""" return self.encoder_attention_heads @property def UpperCamelCase ( self : Tuple ): """simple docstring""" return self.d_model class UpperCamelCase ( _lowercase ): __UpperCamelCase =version.parse("1.11" ) @property def UpperCamelCase ( self : str ): """simple docstring""" return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def UpperCamelCase ( self : Optional[int] ): """simple docstring""" return 1E-5 @property def UpperCamelCase ( self : Tuple ): """simple docstring""" return 1_2
439
"""simple docstring""" from collections import namedtuple lowerCAmelCase__ = namedtuple('''from_to''', '''from_ to''') lowerCAmelCase__ = { '''cubicmeter''': from_to(1, 1), '''litre''': from_to(0.0_0_1, 1000), '''kilolitre''': from_to(1, 1), '''gallon''': from_to(0.0_0_4_5_4, 2_6_4.1_7_2), '''cubicyard''': from_to(0.7_6_4_5_5, 1.3_0_7_9_5), '''cubicfoot''': from_to(0.0_2_8, 3_5.3_1_4_7), '''cup''': from_to(0.0_0_0_2_3_6_5_8_8, 4_2_2_6.7_5), } def snake_case_ ( A_ : float, A_ : str, A_ : str ): '''simple docstring''' if from_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'from_type\' value: {from_type!r} Supported values are:\n''' + ''', '''.join(A_ ) ) if to_type not in METRIC_CONVERSION: raise ValueError( F'''Invalid \'to_type\' value: {to_type!r}. Supported values are:\n''' + ''', '''.join(A_ ) ) return value * METRIC_CONVERSION[from_type].from_ * METRIC_CONVERSION[to_type].to if __name__ == "__main__": import doctest doctest.testmod()
83
0
import math import time from transformers import Trainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class UpperCAmelCase_ ( _lowercase ): '''simple docstring''' def __init__( self , *__A , __A=None , __A=None , **__A ): """simple docstring""" super().__init__(*__lowerCAmelCase , **__lowerCAmelCase ) lowerCamelCase : Union[str, Any] = eval_examples lowerCamelCase : Optional[int] = post_process_function def _snake_case ( self , __A=None , __A=None , __A=None , __A = "eval" ): """simple docstring""" lowerCamelCase : Union[str, Any] = self.eval_dataset if eval_dataset is None else eval_dataset lowerCamelCase : Any = self.get_eval_dataloader(__lowerCAmelCase ) lowerCamelCase : Optional[int] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowerCamelCase : Dict = self.compute_metrics lowerCamelCase : str = None lowerCamelCase : Union[str, Any] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop lowerCamelCase : List[Any] = time.time() try: lowerCamelCase : Union[str, Any] = eval_loop( __lowerCAmelCase , description="Evaluation" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=__lowerCAmelCase , metric_key_prefix=__lowerCAmelCase , ) finally: lowerCamelCase : int = compute_metrics lowerCamelCase : List[str] = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( __lowerCAmelCase , __lowerCAmelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowerCamelCase : Optional[Any] = self.post_process_function(__lowerCAmelCase , __lowerCAmelCase , output.predictions ) lowerCamelCase : Optional[int] = self.compute_metrics(__lowerCAmelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): lowerCamelCase : Tuple = metrics.pop(__lowerCAmelCase ) metrics.update(output.metrics ) else: lowerCamelCase : Tuple = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(__lowerCAmelCase ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) lowerCamelCase : str = self.callback_handler.on_evaluate(self.args , self.state , self.control , __lowerCAmelCase ) return metrics def _snake_case ( self , __A , __A , __A=None , __A = "test" ): """simple docstring""" lowerCamelCase : Optional[int] = self.get_test_dataloader(__lowerCAmelCase ) # Temporarily disable metric computation, we will do it in the loop here. lowerCamelCase : List[str] = self.compute_metrics lowerCamelCase : List[Any] = None lowerCamelCase : List[str] = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop lowerCamelCase : Any = time.time() try: lowerCamelCase : str = eval_loop( __lowerCAmelCase , description="Prediction" , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=__lowerCAmelCase , metric_key_prefix=__lowerCAmelCase , ) finally: lowerCamelCase : str = compute_metrics lowerCamelCase : str = self.args.eval_batch_size * self.args.world_size if F"""{metric_key_prefix}_jit_compilation_time""" in output.metrics: start_time += output.metrics[F"""{metric_key_prefix}_jit_compilation_time"""] output.metrics.update( speed_metrics( __lowerCAmelCase , __lowerCAmelCase , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output lowerCamelCase : str = self.post_process_function(__lowerCAmelCase , __lowerCAmelCase , output.predictions , "predict" ) lowerCamelCase : Tuple = self.compute_metrics(__lowerCAmelCase ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F"""{metric_key_prefix}_""" ): lowerCamelCase : Tuple = metrics.pop(__lowerCAmelCase ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=__lowerCAmelCase )
340
"""simple docstring""" import unittest from transformers import DebertaVaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaVaForMaskedLM, DebertaVaForMultipleChoice, DebertaVaForQuestionAnswering, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaModel, ) from transformers.models.deberta_va.modeling_deberta_va import DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST class __snake_case ( _lowercase): def __init__( self : Dict , __lowerCAmelCase : Any , __lowerCAmelCase : int=1_3 , __lowerCAmelCase : Optional[int]=7 , __lowerCAmelCase : Optional[Any]=True , __lowerCAmelCase : str=True , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : int=True , __lowerCAmelCase : List[str]=9_9 , __lowerCAmelCase : List[Any]=3_2 , __lowerCAmelCase : Union[str, Any]=5 , __lowerCAmelCase : List[str]=4 , __lowerCAmelCase : Optional[int]=3_7 , __lowerCAmelCase : List[Any]="gelu" , __lowerCAmelCase : Tuple=0.1 , __lowerCAmelCase : List[Any]=0.1 , __lowerCAmelCase : int=5_1_2 , __lowerCAmelCase : Tuple=1_6 , __lowerCAmelCase : Tuple=2 , __lowerCAmelCase : Tuple=0.02 , __lowerCAmelCase : List[str]=False , __lowerCAmelCase : Tuple=True , __lowerCAmelCase : Any="None" , __lowerCAmelCase : str=3 , __lowerCAmelCase : Dict=4 , __lowerCAmelCase : Optional[Any]=None , ): """simple docstring""" _lowerCamelCase : Dict = parent _lowerCamelCase : Union[str, Any] = batch_size _lowerCamelCase : Optional[Any] = seq_length _lowerCamelCase : Optional[Any] = is_training _lowerCamelCase : Dict = use_input_mask _lowerCamelCase : Tuple = use_token_type_ids _lowerCamelCase : Optional[Any] = use_labels _lowerCamelCase : List[str] = vocab_size _lowerCamelCase : Any = hidden_size _lowerCamelCase : int = num_hidden_layers _lowerCamelCase : Optional[Any] = num_attention_heads _lowerCamelCase : int = intermediate_size _lowerCamelCase : Optional[int] = hidden_act _lowerCamelCase : int = hidden_dropout_prob _lowerCamelCase : Dict = attention_probs_dropout_prob _lowerCamelCase : List[Any] = max_position_embeddings _lowerCamelCase : str = type_vocab_size _lowerCamelCase : List[Any] = type_sequence_label_size _lowerCamelCase : List[Any] = initializer_range _lowerCamelCase : Optional[int] = num_labels _lowerCamelCase : Any = num_choices _lowerCamelCase : int = relative_attention _lowerCamelCase : Union[str, Any] = position_biased_input _lowerCamelCase : str = pos_att_type _lowerCamelCase : Tuple = scope def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _lowerCamelCase : List[Any] = None if self.use_input_mask: _lowerCamelCase : Any = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) _lowerCamelCase : Any = None if self.use_token_type_ids: _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _lowerCamelCase : Any = None _lowerCamelCase : int = None _lowerCamelCase : Union[str, Any] = None if self.use_labels: _lowerCamelCase : Tuple = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _lowerCamelCase : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _lowerCamelCase : List[str] = ids_tensor([self.batch_size] , self.num_choices ) _lowerCamelCase : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" return DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def SCREAMING_SNAKE_CASE ( self : List[str] , __lowerCAmelCase : str ): """simple docstring""" self.parent.assertListEqual(list(result.loss.size() ) , [] ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple , __lowerCAmelCase : Any , __lowerCAmelCase : Tuple , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModel(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : str = model(__lowerCAmelCase , token_type_ids=__lowerCAmelCase )[0] _lowerCamelCase : List[Any] = model(__lowerCAmelCase )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] , __lowerCAmelCase : Any , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : str , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaForMaskedLM(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[int] = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = self.num_labels _lowerCamelCase : Dict = DebertaVaForSequenceClassification(__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Dict = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : List[str] , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Union[str, Any] , __lowerCAmelCase : Optional[Any] , __lowerCAmelCase : Optional[int] ): """simple docstring""" _lowerCamelCase : Optional[int] = self.num_labels _lowerCamelCase : Tuple = DebertaVaForTokenClassification(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Any = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : int , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Dict , __lowerCAmelCase : List[Any] , __lowerCAmelCase : List[Any] , __lowerCAmelCase : Tuple ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaForQuestionAnswering(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Tuple = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , start_positions=__lowerCAmelCase , end_positions=__lowerCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : str , __lowerCAmelCase : str , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[int] , __lowerCAmelCase : str , __lowerCAmelCase : Dict , __lowerCAmelCase : Dict ): """simple docstring""" _lowerCamelCase : Optional[int] = DebertaVaForMultipleChoice(config=__lowerCAmelCase ) model.to(__lowerCAmelCase ) model.eval() _lowerCamelCase : Optional[Any] = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Optional[Any] = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : Union[str, Any] = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() _lowerCamelCase : List[Any] = model( __lowerCAmelCase , attention_mask=__lowerCAmelCase , token_type_ids=__lowerCAmelCase , labels=__lowerCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" _lowerCamelCase : Any = self.prepare_config_and_inputs() ( ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ( _lowerCamelCase ) , ) : Union[str, Any] = config_and_inputs _lowerCamelCase : Optional[int] = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class __snake_case ( _lowercase , _lowercase , unittest.TestCase): snake_case__ : int = ( ( DebertaVaModel, DebertaVaForMaskedLM, DebertaVaForSequenceClassification, DebertaVaForTokenClassification, DebertaVaForQuestionAnswering, DebertaVaForMultipleChoice, ) if is_torch_available() else () ) snake_case__ : int = ( { "feature-extraction": DebertaVaModel, "fill-mask": DebertaVaForMaskedLM, "question-answering": DebertaVaForQuestionAnswering, "text-classification": DebertaVaForSequenceClassification, "token-classification": DebertaVaForTokenClassification, "zero-shot": DebertaVaForSequenceClassification, } if is_torch_available() else {} ) snake_case__ : List[str] = True snake_case__ : List[Any] = False snake_case__ : int = False snake_case__ : Optional[Any] = False snake_case__ : str = False def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[str] = DebertaVaModelTester(self ) _lowerCamelCase : Any = ConfigTester(self , config_class=__lowerCAmelCase , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : str ): """simple docstring""" _lowerCamelCase : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : int ): """simple docstring""" _lowerCamelCase : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[str] ): """simple docstring""" _lowerCamelCase : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_multiple_choice(*__lowerCAmelCase ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" for model_name in DEBERTA_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained(__lowerCAmelCase ) self.assertIsNotNone(__lowerCAmelCase ) @require_torch @require_sentencepiece @require_tokenizers class __snake_case ( unittest.TestCase): @unittest.skip(reason='''Model not available yet''' ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): """simple docstring""" pass @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): """simple docstring""" _lowerCamelCase : Tuple = DebertaVaModel.from_pretrained('''microsoft/deberta-v2-xlarge''' ) _lowerCamelCase : List[str] = torch.tensor([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) _lowerCamelCase : Any = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): _lowerCamelCase : Tuple = model(__lowerCAmelCase , attention_mask=__lowerCAmelCase )[0] # compare the actual values for a slice. _lowerCamelCase : Union[str, Any] = torch.tensor( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , __lowerCAmelCase , atol=1E-4 ) , f'''{output[:, 1:4, 1:4]}''' )
83
0
import inspect import os import re from transformers.configuration_utils import PretrainedConfig from transformers.utils import direct_transformers_import # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_config_docstrings.py SCREAMING_SNAKE_CASE__ = '''src/transformers''' # This is to make sure the transformers module imported is the one in the repo. SCREAMING_SNAKE_CASE__ = direct_transformers_import(PATH_TO_TRANSFORMERS) SCREAMING_SNAKE_CASE__ = transformers.models.auto.configuration_auto.CONFIG_MAPPING SCREAMING_SNAKE_CASE__ = { # used to compute the property `self.chunk_length` '''EncodecConfig''': ['''overlap'''], # used as `self.bert_model = BertModel(config, ...)` '''DPRConfig''': True, # not used in modeling files, but it's an important information '''FSMTConfig''': ['''langs'''], # used internally in the configuration class file '''GPTNeoConfig''': ['''attention_types'''], # used internally in the configuration class file '''EsmConfig''': ['''is_folding_model'''], # used during training (despite we don't have training script for these models yet) '''Mask2FormerConfig''': ['''ignore_value'''], # `ignore_value` used during training (despite we don't have training script for these models yet) # `norm` used in conversion script (despite not using in the modeling file) '''OneFormerConfig''': ['''ignore_value''', '''norm'''], # used during preprocessing and collation, see `collating_graphormer.py` '''GraphormerConfig''': ['''spatial_pos_max'''], # used internally in the configuration class file '''T5Config''': ['''feed_forward_proj'''], # used internally in the configuration class file # `tokenizer_class` get default value `T5Tokenizer` intentionally '''MT5Config''': ['''feed_forward_proj''', '''tokenizer_class'''], '''UMT5Config''': ['''feed_forward_proj''', '''tokenizer_class'''], # used internally in the configuration class file '''LongT5Config''': ['''feed_forward_proj'''], # used internally in the configuration class file '''SwitchTransformersConfig''': ['''feed_forward_proj'''], # having default values other than `1e-5` - we can't fix them without breaking '''BioGptConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''GLPNConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''SegformerConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''CvtConfig''': ['''layer_norm_eps'''], # having default values other than `1e-5` - we can't fix them without breaking '''PerceiverConfig''': ['''layer_norm_eps'''], # used internally to calculate the feature size '''InformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate the feature size '''TimeSeriesTransformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate the feature size '''AutoformerConfig''': ['''num_static_real_features''', '''num_time_features'''], # used internally to calculate `mlp_dim` '''SamVisionConfig''': ['''mlp_ratio'''], # For (head) training, but so far not implemented '''ClapAudioConfig''': ['''num_classes'''], # Not used, but providing useful information to users '''SpeechT5HifiGanConfig''': ['''sampling_rate'''], } # TODO (ydshieh): Check the failing cases, try to fix them or move some cases to the above block once we are sure SPECIAL_CASES_TO_ALLOW.update( { '''CLIPSegConfig''': True, '''DeformableDetrConfig''': True, '''DetaConfig''': True, '''DinatConfig''': True, '''DonutSwinConfig''': True, '''EfficientFormerConfig''': True, '''FSMTConfig''': True, '''JukeboxConfig''': True, '''LayoutLMv2Config''': True, '''MaskFormerSwinConfig''': True, '''MT5Config''': True, '''NatConfig''': True, '''OneFormerConfig''': True, '''PerceiverConfig''': True, '''RagConfig''': True, '''SpeechT5Config''': True, '''SwinConfig''': True, '''Swin2SRConfig''': True, '''Swinv2Config''': True, '''SwitchTransformersConfig''': True, '''TableTransformerConfig''': True, '''TapasConfig''': True, '''TransfoXLConfig''': True, '''UniSpeechConfig''': True, '''UniSpeechSatConfig''': True, '''WavLMConfig''': True, '''WhisperConfig''': True, # TODO: @Arthur (for `alignment_head` and `alignment_layer`) '''JukeboxPriorConfig''': True, # TODO: @Younes (for `is_decoder`) '''Pix2StructTextConfig''': True, } ) def A ( __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> str: A__ = False for attribute in attributes: for modeling_source in source_strings: # check if we can find `config.xxx`, `getattr(config, "xxx", ...)` or `getattr(self.config, "xxx", ...)` if ( f'''config.{attribute}''' in modeling_source or f'''getattr(config, "{attribute}"''' in modeling_source or f'''getattr(self.config, "{attribute}"''' in modeling_source ): A__ = True # Deal with multi-line cases elif ( re.search( rf'''getattr[ \t\v\n\r\f]*\([ \t\v\n\r\f]*(self\.)?config,[ \t\v\n\r\f]*"{attribute}"''' , A_ , ) is not None ): A__ = True # `SequenceSummary` is called with `SequenceSummary(config)` elif attribute in [ "summary_type", "summary_use_proj", "summary_activation", "summary_last_dropout", "summary_proj_to_labels", "summary_first_dropout", ]: if "SequenceSummary" in modeling_source: A__ = True if attribute_used: break if attribute_used: break # common and important attributes, even if they do not always appear in the modeling files A__ = [ '''bos_index''', '''eos_index''', '''pad_index''', '''unk_index''', '''mask_index''', '''image_size''', '''use_cache''', '''out_features''', '''out_indices''', ] A__ = ['''encoder_no_repeat_ngram_size'''] # Special cases to be allowed A__ = True if not attribute_used: A__ = False for attribute in attributes: # Allow if the default value in the configuration class is different from the one in `PretrainedConfig` if attribute in ["is_encoder_decoder"] and default_value is True: A__ = True elif attribute in ["tie_word_embeddings"] and default_value is False: A__ = True # Allow cases without checking the default value in the configuration class elif attribute in attributes_to_allow + attributes_used_in_generation: A__ = True elif attribute.endswith('_token_id' ): A__ = True # configuration class specific cases if not case_allowed: A__ = SPECIAL_CASES_TO_ALLOW.get(config_class.__name__ , [] ) A__ = allowed_cases is True or attribute in allowed_cases return attribute_used or case_allowed def A ( __UpperCamelCase ) -> Any: A__ = dict(inspect.signature(config_class.__init__ ).parameters ) A__ = [x for x in list(signature.keys() ) if x not in ['''self''', '''kwargs''']] A__ = [signature[param].default for param in parameter_names] # If `attribute_map` exists, an attribute can have different names to be used in the modeling files, and as long # as one variant is used, the test should pass A__ = {} if len(config_class.attribute_map ) > 0: A__ = {v: k for k, v in config_class.attribute_map.items()} # Get the path to modeling source files A__ = inspect.getsourcefile(A_ ) A__ = os.path.dirname(A_ ) # Let's check against all frameworks: as long as one framework uses an attribute, we are good. A__ = [os.path.join(A_ , A_ ) for fn in os.listdir(A_ ) if fn.startswith('modeling_' )] # Get the source code strings A__ = [] for path in modeling_paths: if os.path.isfile(A_ ): with open(A_ ) as fp: modeling_sources.append(fp.read() ) A__ = [] for config_param, default_value in zip(A_ , A_ ): # `attributes` here is all the variant names for `config_param` A__ = [config_param] # some configuration classes have non-empty `attribute_map`, and both names could be used in the # corresponding modeling files. As long as one of them appears, it is fine. if config_param in reversed_attribute_map: attributes.append(reversed_attribute_map[config_param] ) if not check_attribute_being_used(A_ , A_ , A_ , A_ ): unused_attributes.append(attributes[0] ) return sorted(A_ ) def A ( ) -> Union[str, Any]: A__ = {} for _config_class in list(CONFIG_MAPPING.values() ): # Skip deprecated models if "models.deprecated" in _config_class.__module__: continue # Some config classes are not in `CONFIG_MAPPING` (e.g. `CLIPVisionConfig`, `Blip2VisionConfig`, etc.) A__ = [ cls for name, cls in inspect.getmembers( inspect.getmodule(_config_class ) , lambda __UpperCamelCase : inspect.isclass(A_ ) and issubclass(A_ , A_ ) and inspect.getmodule(A_ ) == inspect.getmodule(_config_class ) , ) ] for config_class in config_classes_in_module: A__ = check_config_attributes_being_used(A_ ) if len(A_ ) > 0: A__ = unused_attributes if len(A_ ) > 0: A__ = '''The following configuration classes contain unused attributes in the corresponding modeling files:\n''' for name, attributes in configs_with_unused_attributes.items(): error += f'''{name}: {attributes}\n''' raise ValueError(A_ ) if __name__ == "__main__": check_config_attributes()
9
"""simple docstring""" import json import os import tempfile import unittest import unittest.mock as mock from pathlib import Path from requests.exceptions import HTTPError from transformers.utils import ( CONFIG_NAME, FLAX_WEIGHTS_NAME, TF2_WEIGHTS_NAME, TRANSFORMERS_CACHE, WEIGHTS_NAME, cached_file, get_file_from_repo, has_file, ) lowerCAmelCase__ = '''hf-internal-testing/tiny-random-bert''' lowerCAmelCase__ = os.path.join(TRANSFORMERS_CACHE, '''models--hf-internal-testing--tiny-random-bert''') lowerCAmelCase__ = '''9b8c223d42b2188cb49d29af482996f9d0f3e5a6''' class __snake_case ( unittest.TestCase): def SCREAMING_SNAKE_CASE ( self : Any ): """simple docstring""" _lowerCamelCase : Optional[Any] = cached_file(__lowerCAmelCase , __lowerCAmelCase ) # Should have downloaded the file in here self.assertTrue(os.path.isdir(__lowerCAmelCase ) ) # Cache should contain at least those three subfolders: for subfolder in ["blobs", "refs", "snapshots"]: self.assertTrue(os.path.isdir(os.path.join(__lowerCAmelCase , __lowerCAmelCase ) ) ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : Optional[int] = f.read() self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) self.assertTrue(os.path.isfile(__lowerCAmelCase ) ) # File is cached at the same place the second time. _lowerCamelCase : Tuple = cached_file(__lowerCAmelCase , __lowerCAmelCase ) self.assertEqual(__lowerCAmelCase , __lowerCAmelCase ) # Using a specific revision to test the full commit hash. _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''9b8c223''' ) self.assertEqual(__lowerCAmelCase , os.path.join(__lowerCAmelCase , '''snapshots''' , __lowerCAmelCase , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): _lowerCamelCase : Optional[int] = cached_file('''tiny-random-bert''' , __lowerCAmelCase ) with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): _lowerCamelCase : str = cached_file(__lowerCAmelCase , __lowerCAmelCase , revision='''aaaa''' ) with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : int = cached_file(__lowerCAmelCase , '''conf''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): """simple docstring""" with self.assertRaisesRegex(__lowerCAmelCase , '''does not appear to have a file named''' ): _lowerCamelCase : Dict = cached_file(__lowerCAmelCase , '''conf''' ) with open(os.path.join(__lowerCAmelCase , '''refs''' , '''main''' ) ) as f: _lowerCamelCase : List[Any] = f.read() self.assertTrue(os.path.isfile(os.path.join(__lowerCAmelCase , '''.no_exist''' , __lowerCAmelCase , '''conf''' ) ) ) _lowerCamelCase : str = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Optional[int] = cached_file(__lowerCAmelCase , '''conf''' , local_files_only=__lowerCAmelCase , _raise_exceptions_for_missing_entries=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) _lowerCamelCase : Any = mock.Mock() _lowerCamelCase : Optional[Any] = 5_0_0 _lowerCamelCase : Dict = {} _lowerCamelCase : List[Any] = HTTPError _lowerCamelCase : int = {} # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch('''requests.Session.request''' , return_value=__lowerCAmelCase ) as mock_head: _lowerCamelCase : Union[str, Any] = cached_file(__lowerCAmelCase , '''conf''' , _raise_exceptions_for_connection_errors=__lowerCAmelCase ) self.assertIsNone(__lowerCAmelCase ) # This check we did call the fake head request mock_head.assert_called() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertTrue(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) self.assertFalse(has_file('''hf-internal-testing/tiny-bert-pt-only''' , __lowerCAmelCase ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" self.assertIsNone(get_file_from_repo('''bert-base-cased''' , '''ahah.txt''' ) ) # The function raises if the repository does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid model identifier''' ): get_file_from_repo('''bert-base-case''' , __lowerCAmelCase ) # The function raises if the revision does not exist. with self.assertRaisesRegex(__lowerCAmelCase , '''is not a valid git identifier''' ): get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase , revision='''ahaha''' ) _lowerCamelCase : Dict = get_file_from_repo('''bert-base-cased''' , __lowerCAmelCase ) # The name is the cached name which is not very easy to test, so instead we load the content. _lowerCamelCase : Dict = json.loads(open(__lowerCAmelCase , '''r''' ).read() ) self.assertEqual(config['''hidden_size'''] , 7_6_8 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: _lowerCamelCase : Any = Path(__lowerCAmelCase ) / '''a.txt''' filename.touch() self.assertEqual(get_file_from_repo(__lowerCAmelCase , '''a.txt''' ) , str(__lowerCAmelCase ) ) self.assertIsNone(get_file_from_repo(__lowerCAmelCase , '''b.txt''' ) )
83
0
def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=None ) -> Any: snake_case__ = (path or []) + [u] for v in graph[u]: if visited_edge[u][v] is False: snake_case__ = True, True snake_case__ = dfs(A_ , A_ , A_ , A_ ) return path def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Any: snake_case__ = 0 snake_case__ = -1 for i in range(A_ ): if i not in graph.keys(): continue if len(graph[i] ) % 2 == 1: odd_degree_nodes += 1 snake_case__ = i if odd_degree_nodes == 0: return 1, odd_node if odd_degree_nodes == 2: return 2, odd_node return 3, odd_node def SCREAMING_SNAKE_CASE ( __lowerCAmelCase , __lowerCAmelCase ) -> Tuple: snake_case__ = [[False for _ in range(max_node + 1 )] for _ in range(max_node + 1 )] snake_case__ = check_circuit_or_path(A_ , A_ ) if check == 3: print('''graph is not Eulerian''' ) print('''no path''' ) return snake_case__ = 1 if check == 2: snake_case__ = odd_node print('''graph has a Euler path''' ) if check == 1: print('''graph has a Euler cycle''' ) snake_case__ = dfs(A_ , A_ , A_ ) print(A_ ) def SCREAMING_SNAKE_CASE ( ) -> List[str]: snake_case__ = {1: [2, 3, 4], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [4]} snake_case__ = {1: [2, 3, 4, 5], 2: [1, 3], 3: [1, 2], 4: [1, 5], 5: [1, 4]} snake_case__ = {1: [2, 3, 4], 2: [1, 3, 4], 3: [1, 2], 4: [1, 2, 5], 5: [4]} snake_case__ = {1: [2, 3], 2: [1, 3], 3: [1, 2]} snake_case__ = { 1: [], 2: [] # all degree is zero } snake_case__ = 10 check_euler(A_ , A_ ) check_euler(A_ , A_ ) check_euler(A_ , A_ ) check_euler(A_ , A_ ) check_euler(A_ , A_ ) if __name__ == "__main__": main()
33
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class __snake_case ( _lowercase): snake_case__ : List[str] = "cvt" def __init__( self : Any , __lowerCAmelCase : Union[str, Any]=3 , __lowerCAmelCase : List[str]=[7, 3, 3] , __lowerCAmelCase : int=[4, 2, 2] , __lowerCAmelCase : int=[2, 1, 1] , __lowerCAmelCase : str=[6_4, 1_9_2, 3_8_4] , __lowerCAmelCase : Dict=[1, 3, 6] , __lowerCAmelCase : Optional[Any]=[1, 2, 1_0] , __lowerCAmelCase : Dict=[4.0, 4.0, 4.0] , __lowerCAmelCase : Dict=[0.0, 0.0, 0.0] , __lowerCAmelCase : Union[str, Any]=[0.0, 0.0, 0.0] , __lowerCAmelCase : int=[0.0, 0.0, 0.1] , __lowerCAmelCase : Union[str, Any]=[True, True, True] , __lowerCAmelCase : str=[False, False, True] , __lowerCAmelCase : List[str]=["dw_bn", "dw_bn", "dw_bn"] , __lowerCAmelCase : List[Any]=[3, 3, 3] , __lowerCAmelCase : Dict=[1, 1, 1] , __lowerCAmelCase : str=[2, 2, 2] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : Optional[Any]=[1, 1, 1] , __lowerCAmelCase : str=0.02 , __lowerCAmelCase : Any=1E-12 , **__lowerCAmelCase : int , ): """simple docstring""" super().__init__(**__lowerCAmelCase ) _lowerCamelCase : Optional[int] = num_channels _lowerCamelCase : int = patch_sizes _lowerCamelCase : Optional[Any] = patch_stride _lowerCamelCase : str = patch_padding _lowerCamelCase : Any = embed_dim _lowerCamelCase : Optional[Any] = num_heads _lowerCamelCase : Dict = depth _lowerCamelCase : Optional[int] = mlp_ratio _lowerCamelCase : Any = attention_drop_rate _lowerCamelCase : Any = drop_rate _lowerCamelCase : Dict = drop_path_rate _lowerCamelCase : Optional[int] = qkv_bias _lowerCamelCase : int = cls_token _lowerCamelCase : int = qkv_projection_method _lowerCamelCase : Optional[Any] = kernel_qkv _lowerCamelCase : List[str] = padding_kv _lowerCamelCase : Tuple = stride_kv _lowerCamelCase : Union[str, Any] = padding_q _lowerCamelCase : Optional[Any] = stride_q _lowerCamelCase : Tuple = initializer_range _lowerCamelCase : Optional[int] = layer_norm_eps
83
0
import time import warnings from abc import ABC from copy import deepcopy from typing import Optional import torch from ..utils import add_start_docstrings, logging lowerCAmelCase__: List[Any] = logging.get_logger(__name__) lowerCAmelCase__: Optional[int] = R"\n Args:\n input_ids (`torch.LongTensor` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`AutoTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`torch.FloatTensor` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be scores for each vocabulary token before SoftMax\n or scores for each vocabulary token after SoftMax.\n kwargs (`Dict[str, Any]`, *optional*):\n Additional stopping criteria specific kwargs.\n\n Return:\n `bool`. `False` indicates we should continue, `True` indicates we should stop.\n\n" class snake_case_ ( _lowercase ): @add_start_docstrings(__lowerCAmelCase ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): raise NotImplementedError('StoppingCriteria needs to be subclassed' ) class snake_case_ ( _lowercase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase = None ): SCREAMING_SNAKE_CASE_ : Dict = max_length SCREAMING_SNAKE_CASE_ : int = max_position_embeddings @add_start_docstrings(__lowerCAmelCase ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): SCREAMING_SNAKE_CASE_ : Union[str, Any] = input_ids.shape[-1] SCREAMING_SNAKE_CASE_ : Union[str, Any] = cur_len >= self.max_length if self.max_position_embeddings is not None and not is_done and cur_len >= self.max_position_embeddings: logger.warning_once( 'This is a friendly reminder - the current text generation call will exceed the model\'s predefined ' F'maximum length ({self.max_position_embeddings}). Depending on the model, you may observe ' 'exceptions, performance degradation, or nothing at all.' ) return is_done class snake_case_ ( _lowercase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase ): warnings.warn( 'The class `MaxNewTokensCriteria` is deprecated. ' F'Please use `MaxLengthCriteria(max_length={start_length + max_new_tokens})` ' 'with `max_length = start_length + max_new_tokens` instead.' , __lowerCAmelCase , ) SCREAMING_SNAKE_CASE_ : Dict = start_length SCREAMING_SNAKE_CASE_ : str = max_new_tokens SCREAMING_SNAKE_CASE_ : str = start_length + max_new_tokens @add_start_docstrings(__lowerCAmelCase ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): return input_ids.shape[-1] >= self.max_length class snake_case_ ( _lowercase ): def __init__( self , __lowerCAmelCase , __lowerCAmelCase = None ): SCREAMING_SNAKE_CASE_ : Tuple = max_time SCREAMING_SNAKE_CASE_ : Any = time.time() if initial_timestamp is None else initial_timestamp @add_start_docstrings(__lowerCAmelCase ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): return time.time() - self.initial_timestamp > self.max_time class snake_case_ ( _lowercase ): @add_start_docstrings(__lowerCAmelCase ) def __call__( self , __lowerCAmelCase , __lowerCAmelCase , **__lowerCAmelCase ): return any(criteria(__lowerCAmelCase , __lowerCAmelCase ) for criteria in self ) @property def __A ( self ): for stopping_criterium in self: if isinstance(__lowerCAmelCase , __lowerCAmelCase ): return stopping_criterium.max_length elif isinstance(__lowerCAmelCase , __lowerCAmelCase ): return stopping_criterium.max_length return None def __SCREAMING_SNAKE_CASE ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: SCREAMING_SNAKE_CASE_ : Tuple = stopping_criteria.max_length SCREAMING_SNAKE_CASE_ : str = deepcopy(A_ ) if stopping_max_length is not None and stopping_max_length != max_length: warnings.warn('You set different `max_length` for stopping criteria and `max_length` parameter' , A_ ) elif stopping_max_length is None: new_stopping_criteria.append(MaxLengthCriteria(max_length=A_ ) ) return new_stopping_criteria
345
"""simple docstring""" # DISCLAIMER: This file is strongly influenced by https://github.com/yang-song/score_sde_pytorch import math from dataclasses import dataclass from typing import Optional, Tuple, Union import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin, SchedulerOutput @dataclass class __snake_case ( _lowercase): snake_case__ : torch.FloatTensor snake_case__ : torch.FloatTensor class __snake_case ( _lowercase , _lowercase): snake_case__ : int = 1 @register_to_config def __init__( self : str , __lowerCAmelCase : int = 2_0_0_0 , __lowerCAmelCase : float = 0.15 , __lowerCAmelCase : float = 0.01 , __lowerCAmelCase : float = 13_48.0 , __lowerCAmelCase : float = 1E-5 , __lowerCAmelCase : int = 1 , ): """simple docstring""" _lowerCamelCase : Optional[int] = sigma_max # setable values _lowerCamelCase : Dict = None self.set_sigmas(__lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[int] = None ): """simple docstring""" return sample def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : Union[str, torch.device] = None ): """simple docstring""" _lowerCamelCase : Tuple = sampling_eps if sampling_eps is not None else self.config.sampling_eps _lowerCamelCase : Optional[int] = torch.linspace(1 , __lowerCAmelCase , __lowerCAmelCase , device=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] , __lowerCAmelCase : int , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None , __lowerCAmelCase : float = None ): """simple docstring""" _lowerCamelCase : List[str] = sigma_min if sigma_min is not None else self.config.sigma_min _lowerCamelCase : int = sigma_max if sigma_max is not None else self.config.sigma_max _lowerCamelCase : Any = sampling_eps if sampling_eps is not None else self.config.sampling_eps if self.timesteps is None: self.set_timesteps(__lowerCAmelCase , __lowerCAmelCase ) _lowerCamelCase : List[Any] = sigma_min * (sigma_max / sigma_min) ** (self.timesteps / sampling_eps) _lowerCamelCase : Optional[int] = torch.exp(torch.linspace(math.log(__lowerCAmelCase ) , math.log(__lowerCAmelCase ) , __lowerCAmelCase ) ) _lowerCamelCase : Tuple = torch.tensor([sigma_min * (sigma_max / sigma_min) ** t for t in self.timesteps] ) def SCREAMING_SNAKE_CASE ( self : Tuple , __lowerCAmelCase : Tuple , __lowerCAmelCase : Optional[Any] ): """simple docstring""" return torch.where( timesteps == 0 , torch.zeros_like(t.to(timesteps.device ) ) , self.discrete_sigmas[timesteps - 1].to(timesteps.device ) , ) def SCREAMING_SNAKE_CASE ( self : Optional[int] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : int , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) _lowerCamelCase : Tuple = timestep * torch.ones( sample.shape[0] , device=sample.device ) # torch.repeat_interleave(timestep, sample.shape[0]) _lowerCamelCase : Dict = (timestep * (len(self.timesteps ) - 1)).long() # mps requires indices to be in the same device, so we use cpu as is the default with cuda _lowerCamelCase : Optional[int] = timesteps.to(self.discrete_sigmas.device ) _lowerCamelCase : Any = self.discrete_sigmas[timesteps].to(sample.device ) _lowerCamelCase : int = self.get_adjacent_sigma(__lowerCAmelCase , __lowerCAmelCase ).to(sample.device ) _lowerCamelCase : Any = torch.zeros_like(__lowerCAmelCase ) _lowerCamelCase : Any = (sigma**2 - adjacent_sigma**2) ** 0.5 # equation 6 in the paper: the model_output modeled by the network is grad_x log pt(x) # also equation 47 shows the analog from SDE models to ancestral sampling methods _lowerCamelCase : Union[str, Any] = diffusion.flatten() while len(diffusion.shape ) < len(sample.shape ): _lowerCamelCase : List[Any] = diffusion.unsqueeze(-1 ) _lowerCamelCase : int = drift - diffusion**2 * model_output # equation 6: sample noise for the diffusion term of _lowerCamelCase : List[str] = randn_tensor( sample.shape , layout=sample.layout , generator=__lowerCAmelCase , device=sample.device , dtype=sample.dtype ) _lowerCamelCase : List[Any] = sample - drift # subtract because `dt` is a small negative timestep # TODO is the variable diffusion the correct scaling term for the noise? _lowerCamelCase : int = prev_sample_mean + diffusion * noise # add impact of diffusion field g if not return_dict: return (prev_sample, prev_sample_mean) return SdeVeOutput(prev_sample=__lowerCAmelCase , prev_sample_mean=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : Optional[torch.Generator] = None , __lowerCAmelCase : bool = True , ): """simple docstring""" if self.timesteps is None: raise ValueError( '''`self.timesteps` is not set, you need to run \'set_timesteps\' after creating the scheduler''' ) # For small batch sizes, the paper "suggest replacing norm(z) with sqrt(d), where d is the dim. of z" # sample noise for correction _lowerCamelCase : Union[str, Any] = randn_tensor(sample.shape , layout=sample.layout , generator=__lowerCAmelCase ).to(sample.device ) # compute step size from the model_output, the noise, and the snr _lowerCamelCase : Union[str, Any] = torch.norm(model_output.reshape(model_output.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : Tuple = torch.norm(noise.reshape(noise.shape[0] , -1 ) , dim=-1 ).mean() _lowerCamelCase : str = (self.config.snr * noise_norm / grad_norm) ** 2 * 2 _lowerCamelCase : Tuple = step_size * torch.ones(sample.shape[0] ).to(sample.device ) # self.repeat_scalar(step_size, sample.shape[0]) # compute corrected sample: model_output term and noise term _lowerCamelCase : Union[str, Any] = step_size.flatten() while len(step_size.shape ) < len(sample.shape ): _lowerCamelCase : str = step_size.unsqueeze(-1 ) _lowerCamelCase : Any = sample + step_size * model_output _lowerCamelCase : int = prev_sample_mean + ((step_size * 2) ** 0.5) * noise if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__lowerCAmelCase ) def SCREAMING_SNAKE_CASE ( self : Any , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , __lowerCAmelCase : torch.FloatTensor , ): """simple docstring""" _lowerCamelCase : Dict = timesteps.to(original_samples.device ) _lowerCamelCase : Union[str, Any] = self.discrete_sigmas.to(original_samples.device )[timesteps] _lowerCamelCase : Union[str, Any] = ( noise * sigmas[:, None, None, None] if noise is not None else torch.randn_like(__lowerCAmelCase ) * sigmas[:, None, None, None] ) _lowerCamelCase : int = noise + original_samples return noisy_samples def __len__( self : Optional[int] ): """simple docstring""" return self.config.num_train_timesteps
83
0
from typing import Dict import numpy as np from ..utils import add_end_docstrings, is_tf_available, is_torch_available, logging from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline, PipelineException if is_tf_available(): import tensorflow as tf from ..tf_utils import stable_softmax if is_torch_available(): import torch UpperCamelCase_ = logging.get_logger(__name__) @add_end_docstrings( _lowercase, R'\n top_k (`int`, defaults to 5):\n The number of predictions to return.\n targets (`str` or `List[str]`, *optional*):\n When passed, the model will limit the scores to the passed targets instead of looking up in the whole\n vocab. If the provided targets are not in the model vocab, they will be tokenized and the first resulting\n token will be used (with a warning, and that might be slower).\n\n ', ) class snake_case_ ( _lowercase ): '''simple docstring''' def __UpperCAmelCase ( self, A_ ) -> Dict: if self.framework == "tf": UpperCAmelCase__ =tf.where(input_ids == self.tokenizer.mask_token_id ).numpy() elif self.framework == "pt": UpperCAmelCase__ =torch.nonzero(input_ids == self.tokenizer.mask_token_id, as_tuple=__lowerCAmelCase ) else: raise ValueError("Unsupported framework" ) return masked_index def __UpperCAmelCase ( self, A_ ) -> Optional[int]: UpperCAmelCase__ =self.get_masked_index(__lowerCAmelCase ) UpperCAmelCase__ =np.prod(masked_index.shape ) if numel < 1: raise PipelineException( "fill-mask", self.model.base_model_prefix, f"""No mask_token ({self.tokenizer.mask_token}) found on the input""", ) def __UpperCAmelCase ( self, A_ ) -> Dict: if isinstance(__lowerCAmelCase, __lowerCAmelCase ): for model_input in model_inputs: self._ensure_exactly_one_mask_token(model_input["input_ids"][0] ) else: for input_ids in model_inputs["input_ids"]: self._ensure_exactly_one_mask_token(__lowerCAmelCase ) def __UpperCAmelCase ( self, A_, A_=None, **A_ ) -> Tuple: if return_tensors is None: UpperCAmelCase__ =self.framework UpperCAmelCase__ =self.tokenizer(__lowerCAmelCase, return_tensors=__lowerCAmelCase ) self.ensure_exactly_one_mask_token(__lowerCAmelCase ) return model_inputs def __UpperCAmelCase ( self, A_ ) -> List[Any]: UpperCAmelCase__ =self.model(**__lowerCAmelCase ) UpperCAmelCase__ =model_inputs['''input_ids'''] return model_outputs def __UpperCAmelCase ( self, A_, A_=5, A_=None ) -> Optional[int]: if target_ids is not None and target_ids.shape[0] < top_k: UpperCAmelCase__ =target_ids.shape[0] UpperCAmelCase__ =model_outputs['''input_ids'''][0] UpperCAmelCase__ =model_outputs['''logits'''] if self.framework == "tf": UpperCAmelCase__ =tf.where(input_ids == self.tokenizer.mask_token_id ).numpy()[:, 0] UpperCAmelCase__ =outputs.numpy() UpperCAmelCase__ =outputs[0, masked_index, :] UpperCAmelCase__ =stable_softmax(__lowerCAmelCase, axis=-1 ) if target_ids is not None: UpperCAmelCase__ =tf.gather_nd(tf.squeeze(__lowerCAmelCase, 0 ), target_ids.reshape(-1, 1 ) ) UpperCAmelCase__ =tf.expand_dims(__lowerCAmelCase, 0 ) UpperCAmelCase__ =tf.math.top_k(__lowerCAmelCase, k=__lowerCAmelCase ) UpperCAmelCase__ =topk.values.numpy(), topk.indices.numpy() else: UpperCAmelCase__ =torch.nonzero(input_ids == self.tokenizer.mask_token_id, as_tuple=__lowerCAmelCase ).squeeze(-1 ) # Fill mask pipeline supports only one ${mask_token} per sample UpperCAmelCase__ =outputs[0, masked_index, :] UpperCAmelCase__ =logits.softmax(dim=-1 ) if target_ids is not None: UpperCAmelCase__ =probs[..., target_ids] UpperCAmelCase__ =probs.topk(__lowerCAmelCase ) UpperCAmelCase__ =[] UpperCAmelCase__ =values.shape[0] == 1 for i, (_values, _predictions) in enumerate(zip(values.tolist(), predictions.tolist() ) ): UpperCAmelCase__ =[] for v, p in zip(_values, _predictions ): # Copy is important since we're going to modify this array in place UpperCAmelCase__ =input_ids.numpy().copy() if target_ids is not None: UpperCAmelCase__ =target_ids[p].tolist() UpperCAmelCase__ =p # Filter padding out: UpperCAmelCase__ =tokens[np.where(tokens != self.tokenizer.pad_token_id )] # Originally we skip special tokens to give readable output. # For multi masks though, the other [MASK] would be removed otherwise # making the output look odd, so we add them back UpperCAmelCase__ =self.tokenizer.decode(__lowerCAmelCase, skip_special_tokens=__lowerCAmelCase ) UpperCAmelCase__ ={'''score''': v, '''token''': p, '''token_str''': self.tokenizer.decode([p] ), '''sequence''': sequence} row.append(__lowerCAmelCase ) result.append(__lowerCAmelCase ) if single_mask: return result[0] return result def __UpperCAmelCase ( self, A_, A_=None ) -> List[str]: if isinstance(__lowerCAmelCase, __lowerCAmelCase ): UpperCAmelCase__ =[targets] try: UpperCAmelCase__ =self.tokenizer.get_vocab() except Exception: UpperCAmelCase__ ={} UpperCAmelCase__ =[] for target in targets: UpperCAmelCase__ =vocab.get(__lowerCAmelCase, __lowerCAmelCase ) if id_ is None: UpperCAmelCase__ =self.tokenizer( __lowerCAmelCase, add_special_tokens=__lowerCAmelCase, return_attention_mask=__lowerCAmelCase, return_token_type_ids=__lowerCAmelCase, max_length=1, truncation=__lowerCAmelCase, )['''input_ids'''] if len(__lowerCAmelCase ) == 0: logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ "We cannot replace it with anything meaningful, ignoring it" ) continue UpperCAmelCase__ =input_ids[0] # XXX: If users encounter this pass # it becomes pretty slow, so let's make sure # The warning enables them to fix the input to # get faster performance. logger.warning( f"""The specified target token `{target}` does not exist in the model vocabulary. """ f"""Replacing with `{self.tokenizer.convert_ids_to_tokens(id_ )}`.""" ) target_ids.append(id_ ) UpperCAmelCase__ =list(set(__lowerCAmelCase ) ) if len(__lowerCAmelCase ) == 0: raise ValueError("At least one target must be provided when passed." ) UpperCAmelCase__ =np.array(__lowerCAmelCase ) return target_ids def __UpperCAmelCase ( self, A_=None, A_=None ) -> Any: UpperCAmelCase__ ={} if targets is not None: UpperCAmelCase__ =self.get_target_ids(__lowerCAmelCase, __lowerCAmelCase ) UpperCAmelCase__ =target_ids if top_k is not None: UpperCAmelCase__ =top_k if self.tokenizer.mask_token_id is None: raise PipelineException( "fill-mask", self.model.base_model_prefix, "The tokenizer does not define a `mask_token`." ) return {}, {}, postprocess_params def __call__( self, A_, *A_, **A_ ) -> str: UpperCAmelCase__ =super().__call__(__lowerCAmelCase, **__lowerCAmelCase ) if isinstance(__lowerCAmelCase, __lowerCAmelCase ) and len(__lowerCAmelCase ) == 1: return outputs[0] return outputs
625
"""simple docstring""" from torch import nn def snake_case_ ( A_ : int ): '''simple docstring''' if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'''Unsupported activation function: {act_fn}''' )
83
0
import colorsys from PIL import Image # type: ignore def a_ ( SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : float , SCREAMING_SNAKE_CASE__ : int ): '''simple docstring''' _lowerCamelCase : List[Any] =x _lowerCamelCase : List[Any] =y for step in range(A_ ): # noqa: B007 _lowerCamelCase : Dict =a * a - b * b + x _lowerCamelCase : List[str] =2 * a * b + y _lowerCamelCase : Any =a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def a_ ( SCREAMING_SNAKE_CASE__ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def a_ ( SCREAMING_SNAKE_CASE__ : float ): '''simple docstring''' if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255 ) for i in colorsys.hsv_to_rgb(A_ , 1 , 1 ) ) def a_ ( SCREAMING_SNAKE_CASE__ : int = 800 , SCREAMING_SNAKE_CASE__ : int = 600 , SCREAMING_SNAKE_CASE__ : float = -0.6 , SCREAMING_SNAKE_CASE__ : float = 0 , SCREAMING_SNAKE_CASE__ : float = 3.2 , SCREAMING_SNAKE_CASE__ : int = 50 , SCREAMING_SNAKE_CASE__ : bool = True , ): '''simple docstring''' _lowerCamelCase : Tuple =Image.new('RGB' , (image_width, image_height) ) _lowerCamelCase : int =img.load() # loop through the image-coordinates for image_x in range(A_ ): for image_y in range(A_ ): # determine the figure-coordinates based on the image-coordinates _lowerCamelCase : Optional[Any] =figure_width / image_width * image_height _lowerCamelCase : List[Any] =figure_center_x + (image_x / image_width - 0.5) * figure_width _lowerCamelCase : Optional[Any] =figure_center_y + (image_y / image_height - 0.5) * figure_height _lowerCamelCase : str =get_distance(A_ , A_ , A_ ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _lowerCamelCase : Dict =get_color_coded_rgb(A_ ) else: _lowerCamelCase : str =get_black_and_white_rgb(A_ ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure lowerCamelCase = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
464
"""simple docstring""" def snake_case_ ( A_ : int, A_ : int ): '''simple docstring''' return int(input_a == input_a == 0 ) def snake_case_ ( ): '''simple docstring''' print('''Truth Table of NOR Gate:''' ) print('''| Input 1 | Input 2 | Output |''' ) print(F'''| 0 | 0 | {nor_gate(0, 0 )} |''' ) print(F'''| 0 | 1 | {nor_gate(0, 1 )} |''' ) print(F'''| 1 | 0 | {nor_gate(1, 0 )} |''' ) print(F'''| 1 | 1 | {nor_gate(1, 1 )} |''' ) if __name__ == "__main__": import doctest doctest.testmod() main()
83
0
import gc import unittest import numpy as np import torch from diffusers import StableDiffusionKDiffusionPipeline from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() @slow @require_torch_gpu class A_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = StableDiffusionKDiffusionPipeline.from_pretrained('CompVis/stable-diffusion-v1-4' ) lowercase = sd_pipe.to(snake_case ) sd_pipe.set_progress_bar_config(disable=snake_case ) sd_pipe.set_scheduler('sample_euler' ) lowercase = 'A painting of a squirrel eating a burger' lowercase = torch.manual_seed(0 ) lowercase = sd_pipe([prompt] , generator=snake_case , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) lowercase = output.images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowercase = np.array([0.0_447, 0.0_492, 0.0_468, 0.0_408, 0.0_383, 0.0_408, 0.0_354, 0.0_380, 0.0_339] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) lowercase = sd_pipe.to(snake_case ) sd_pipe.set_progress_bar_config(disable=snake_case ) sd_pipe.set_scheduler('sample_euler' ) lowercase = 'A painting of a squirrel eating a burger' lowercase = torch.manual_seed(0 ) lowercase = sd_pipe([prompt] , generator=snake_case , guidance_scale=9.0 , num_inference_steps=20 , output_type='np' ) lowercase = output.images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowercase = np.array([0.1_237, 0.1_320, 0.1_438, 0.1_359, 0.1_390, 0.1_132, 0.1_277, 0.1_175, 0.1_112] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 5E-1 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = StableDiffusionKDiffusionPipeline.from_pretrained('stabilityai/stable-diffusion-2-1-base' ) lowercase = sd_pipe.to(snake_case ) sd_pipe.set_progress_bar_config(disable=snake_case ) sd_pipe.set_scheduler('sample_dpmpp_2m' ) lowercase = 'A painting of a squirrel eating a burger' lowercase = torch.manual_seed(0 ) lowercase = sd_pipe( [prompt] , generator=snake_case , guidance_scale=7.5 , num_inference_steps=15 , output_type='np' , use_karras_sigmas=snake_case , ) lowercase = output.images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 512, 512, 3) lowercase = np.array( [0.11_381_689, 0.12_112_921, 0.1_389_457, 0.12_549_606, 0.1_244_964, 0.10_831_517, 0.11_562_866, 0.10_867_816, 0.10_499_048] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
84
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } UpperCAmelCase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } UpperCAmelCase = { '''facebook/blenderbot_small-90M''': 512, } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : Dict = VOCAB_FILES_NAMES _UpperCamelCase : Any = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : str = BlenderbotSmallTokenizer def __init__( self , snake_case=None , snake_case=None , snake_case="<|endoftext|>" , snake_case="<|endoftext|>" , snake_case="<|endoftext|>" , snake_case=False , snake_case=True , **snake_case , ): super().__init__( ByteLevelBPETokenizer( vocab=snake_case , merges=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case , ) , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , **snake_case , ) lowercase = add_prefix_space def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=None ): lowercase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
84
1
from itertools import product def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = sides_number lowercase = max_face_number * dice_number lowercase = [0] * (max_total + 1) lowercase = 1 lowercase = range(__SCREAMING_SNAKE_CASE , max_face_number + 1 ) for dice_numbers in product(__SCREAMING_SNAKE_CASE , repeat=__SCREAMING_SNAKE_CASE ): lowercase = sum(__SCREAMING_SNAKE_CASE ) totals_frequencies[total] += 1 return totals_frequencies def UpperCAmelCase_ ( ): lowercase = total_frequency_distribution( sides_number=4 , dice_number=9 ) lowercase = total_frequency_distribution( sides_number=6 , dice_number=6 ) lowercase = 0 lowercase = 9 lowercase = 4 * 9 lowercase = 6 for peter_total in range(__SCREAMING_SNAKE_CASE , max_peter_total + 1 ): peter_wins_count += peter_totals_frequencies[peter_total] * sum( colin_totals_frequencies[min_colin_total:peter_total] ) lowercase = (4**9) * (6**6) lowercase = peter_wins_count / total_games_number lowercase = round(__SCREAMING_SNAKE_CASE , ndigits=7 ) return rounded_peter_win_probability if __name__ == "__main__": print(F"""{solution() = }""")
84
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = self.vocab_size - 1 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) lowercase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , head_mask=snake_case ) lowercase = model(snake_case , token_type_ids=snake_case ) lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTLMHeadModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTDoubleHeadsModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = self.num_labels lowercase = OpenAIGPTForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask, } return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _UpperCamelCase : Tuple = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _UpperCamelCase : str = ( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case=False ): lowercase = super()._prepare_for_class(snake_case , snake_case , return_labels=snake_case ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=snake_case , ) lowercase = inputs_dict['labels'] lowercase = inputs_dict['labels'] lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=snake_case , ) lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case ) return inputs_dict def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , n_embd=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = OpenAIGPTModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt' ) model.to(snake_case ) lowercase = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=snake_case ) # the president is lowercase = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 4_0477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the lowercase = model.generate(snake_case , do_sample=snake_case ) self.assertListEqual(output_ids[0].tolist() , snake_case )
84
1
from collections.abc import Callable def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = a lowercase = b if function(__SCREAMING_SNAKE_CASE ) == 0: # one of the a or b is a root for the function return a elif function(__SCREAMING_SNAKE_CASE ) == 0: return b elif ( function(__SCREAMING_SNAKE_CASE ) * function(__SCREAMING_SNAKE_CASE ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError('could not find root in given interval.' ) else: lowercase = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(__SCREAMING_SNAKE_CASE ) == 0: return mid elif function(__SCREAMING_SNAKE_CASE ) * function(__SCREAMING_SNAKE_CASE ) < 0: lowercase = mid else: lowercase = mid lowercase = start + (end - start) / 2.0 return mid def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1000)) import doctest doctest.testmod()
84
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = {'''configuration_vit_msn''': ['''VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMSNConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ '''VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMSNModel''', '''ViTMSNForImageClassification''', '''ViTMSNPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
1
import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint UpperCAmelCase = { '''169M''': 12, '''430M''': 24, '''1B5''': 24, '''3B''': 32, '''7B''': 32, '''14B''': 40, } UpperCAmelCase = { '''169M''': 768, '''430M''': 1024, '''1B5''': 2048, '''3B''': 2560, '''7B''': 4096, '''14B''': 5120, } def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = list(state_dict.keys() ) for name in state_dict_keys: lowercase = state_dict.pop(__SCREAMING_SNAKE_CASE ) # emb -> embedding if name.startswith('emb.' ): lowercase = name.replace('emb.' , 'embeddings.' ) # ln_0 -> pre_ln (only present at block 0) if name.startswith('blocks.0.ln0' ): lowercase = name.replace('blocks.0.ln0' , 'blocks.0.pre_ln' ) # att -> attention lowercase = re.sub(r'blocks\.(\d+)\.att' , r'blocks.\1.attention' , __SCREAMING_SNAKE_CASE ) # ffn -> feed_forward lowercase = re.sub(r'blocks\.(\d+)\.ffn' , r'blocks.\1.feed_forward' , __SCREAMING_SNAKE_CASE ) # time_mix_k -> time_mix_key and reshape if name.endswith('.time_mix_k' ): lowercase = name.replace('.time_mix_k' , '.time_mix_key' ) # time_mix_v -> time_mix_value and reshape if name.endswith('.time_mix_v' ): lowercase = name.replace('.time_mix_v' , '.time_mix_value' ) # time_mix_r -> time_mix_key and reshape if name.endswith('.time_mix_r' ): lowercase = name.replace('.time_mix_r' , '.time_mix_receptance' ) if name != "head.weight": lowercase = 'rwkv.' + name lowercase = weight return state_dict def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=None ): # 1. If possible, build the tokenizer. if tokenizer_file is None: print('No `--tokenizer_file` provided, we will use the default tokenizer.' ) lowercase = 5_0277 lowercase = AutoTokenizer.from_pretrained('EleutherAI/gpt-neox-20b' ) else: lowercase = PreTrainedTokenizerFast(tokenizer_file=__SCREAMING_SNAKE_CASE ) lowercase = len(__SCREAMING_SNAKE_CASE ) tokenizer.save_pretrained(__SCREAMING_SNAKE_CASE ) # 2. Build the config lowercase = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: lowercase = candidate break if size is None: raise ValueError('Could not infer the size, please provide it with the `--size` argument.' ) if size not in possible_sizes: raise ValueError(F'''`size` should be one of {possible_sizes}, got {size}.''' ) lowercase = RwkvConfig( vocab_size=__SCREAMING_SNAKE_CASE , num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size] , hidden_size=HIDEN_SIZE_MAPPING[size] , ) config.save_pretrained(__SCREAMING_SNAKE_CASE ) # 3. Download model file then convert state_dict lowercase = hf_hub_download(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase = torch.load(__SCREAMING_SNAKE_CASE , map_location='cpu' ) lowercase = convert_state_dict(__SCREAMING_SNAKE_CASE ) # 4. Split in shards and save lowercase , lowercase = shard_checkpoint(__SCREAMING_SNAKE_CASE ) for shard_file, shard in shards.items(): torch.save(__SCREAMING_SNAKE_CASE , os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if index is not None: lowercase = os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Save the index as well with open(__SCREAMING_SNAKE_CASE , 'w' , encoding='utf-8' ) as f: lowercase = json.dumps(__SCREAMING_SNAKE_CASE , indent=2 , sort_keys=__SCREAMING_SNAKE_CASE ) + '\n' f.write(__SCREAMING_SNAKE_CASE ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( 'Cleaning up shards. This may error with an OOM error, it this is the case don\'t worry you still have converted the model.' ) lowercase = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: lowercase = torch.load(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()} , os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError('Please provide a `model_name` to push the model to the Hub.' ) lowercase = AutoModelForCausalLM.from_pretrained(__SCREAMING_SNAKE_CASE ) model.push_to_hub(__SCREAMING_SNAKE_CASE , max_shard_size='2GB' ) tokenizer.push_to_hub(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": UpperCAmelCase = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--repo_id''', default=None, type=str, required=True, help='''Repo ID from which to pull the checkpoint.''' ) parser.add_argument( '''--checkpoint_file''', default=None, type=str, required=True, help='''Name of the checkpoint file in the repo.''' ) parser.add_argument( '''--output_dir''', default=None, type=str, required=True, help='''Where to save the converted model.''' ) parser.add_argument( '''--tokenizer_file''', default=None, type=str, help='''Path to the tokenizer file to use (if not provided, only the model is converted).''', ) parser.add_argument( '''--size''', default=None, type=str, help='''Size of the model. Will be inferred from the `checkpoint_file` if not passed.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Push to the Hub the converted model.''', ) parser.add_argument( '''--model_name''', default=None, type=str, help='''Name of the pushed model on the Hub, including the username / organization.''', ) UpperCAmelCase = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
84
import math def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = [True] * n lowercase = False lowercase = False lowercase = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): lowercase = i * 2 while index < n: lowercase = False lowercase = index + i lowercase = [2] for i in range(3 , __SCREAMING_SNAKE_CASE , 2 ): if is_prime[i]: primes.append(__SCREAMING_SNAKE_CASE ) return primes def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 9999_6666_3333 ): lowercase = math.floor(math.sqrt(__SCREAMING_SNAKE_CASE ) ) + 100 lowercase = prime_sieve(__SCREAMING_SNAKE_CASE ) lowercase = 0 lowercase = 0 lowercase = primes[prime_index] while (last_prime**2) <= limit: lowercase = primes[prime_index + 1] lowercase = last_prime**2 lowercase = next_prime**2 # Get numbers divisible by lps(current) lowercase = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) lowercase = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps lowercase = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair lowercase = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
84
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available UpperCAmelCase = {'''configuration_yolos''': ['''YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''YolosConfig''', '''YolosOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''YolosFeatureExtractor'''] UpperCAmelCase = ['''YolosImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ '''YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''YolosForObjectDetection''', '''YolosModel''', '''YolosPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
import collections import os import re from pathlib import Path UpperCAmelCase = '''src/transformers''' # Matches is_xxx_available() UpperCAmelCase = re.compile(R'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} UpperCAmelCase = re.compile(R'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] UpperCAmelCase = re.compile(R'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available UpperCAmelCase = re.compile(R'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") UpperCAmelCase = re.compile(R'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] UpperCAmelCase = re.compile(R'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", UpperCAmelCase = re.compile(R'''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], UpperCAmelCase = re.compile(R'''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo UpperCAmelCase = re.compile(R'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: UpperCAmelCase = re.compile(R'''^\s*try:''') # Catches a line with else: UpperCAmelCase = re.compile(R'''^\s*else:''') def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if _re_test_backend.search(__SCREAMING_SNAKE_CASE ) is None: return None lowercase = [b[0] for b in _re_backend.findall(__SCREAMING_SNAKE_CASE )] backends.sort() return "_and_".join(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): with open(__SCREAMING_SNAKE_CASE , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase = f.readlines() lowercase = 0 while line_index < len(__SCREAMING_SNAKE_CASE ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__SCREAMING_SNAKE_CASE ): return None # First grab the objects without a specific backend in _import_structure lowercase = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: lowercase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__SCREAMING_SNAKE_CASE ): lowercase = _re_one_line_import_struct.search(__SCREAMING_SNAKE_CASE ).groups()[0] lowercase = re.findall(r'\[([^\]]+)\]' , __SCREAMING_SNAKE_CASE ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue lowercase = _re_import_struct_key_value.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: lowercase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 lowercase = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): lowercase = lines[line_index] if _re_import_struct_add_one.search(__SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_import_struct_add_one.search(__SCREAMING_SNAKE_CASE ).groups()[0] ) elif _re_import_struct_add_many.search(__SCREAMING_SNAKE_CASE ) is not None: lowercase = _re_import_struct_add_many.search(__SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) lowercase = [obj[1:-1] for obj in imports if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif _re_between_brackets.search(__SCREAMING_SNAKE_CASE ) is not None: lowercase = _re_between_brackets.search(__SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) lowercase = [obj[1:-1] for obj in imports if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif _re_quote_object.search(__SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_quote_object.search(__SCREAMING_SNAKE_CASE ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 lowercase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend lowercase = [] while ( line_index < len(__SCREAMING_SNAKE_CASE ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): lowercase = lines[line_index] lowercase = _re_import.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 lowercase = {'none': objects} # Let's continue with backend-specific objects while line_index < len(__SCREAMING_SNAKE_CASE ): # If the line is an if is_backend_available, we grab all objects associated. lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): lowercase = lines[line_index] lowercase = _re_import.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 lowercase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): def find_duplicates(__SCREAMING_SNAKE_CASE ): return [k for k, v in collections.Counter(__SCREAMING_SNAKE_CASE ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] lowercase = [] for key in import_dict_objects.keys(): lowercase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) lowercase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): lowercase = 'base imports' if key == 'none' else F'''{key} backend''' errors.append(F'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def UpperCAmelCase_ ( ): lowercase = [] for root, _, files in os.walk(__SCREAMING_SNAKE_CASE ): if "__init__.py" in files: lowercase = os.path.join(__SCREAMING_SNAKE_CASE , '__init__.py' ) lowercase = parse_init(__SCREAMING_SNAKE_CASE ) if objects is not None: lowercase = analyze_results(*__SCREAMING_SNAKE_CASE ) if len(__SCREAMING_SNAKE_CASE ) > 0: lowercase = F'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('\n'.join(__SCREAMING_SNAKE_CASE ) ) if len(__SCREAMING_SNAKE_CASE ) > 0: raise ValueError('\n\n'.join(__SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase_ ( ): lowercase = [] for path, directories, files in os.walk(__SCREAMING_SNAKE_CASE ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(__SCREAMING_SNAKE_CASE ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__SCREAMING_SNAKE_CASE ) / folder).glob('*.py' ) ) ) == 0: continue lowercase = str((Path(__SCREAMING_SNAKE_CASE ) / folder).relative_to(__SCREAMING_SNAKE_CASE ) ) lowercase = short_path.replace(os.path.sep , '.' ) submodules.append(__SCREAMING_SNAKE_CASE ) for fname in files: if fname == "__init__.py": continue lowercase = str((Path(__SCREAMING_SNAKE_CASE ) / fname).relative_to(__SCREAMING_SNAKE_CASE ) ) lowercase = short_path.replace('.py' , '' ).replace(os.path.sep , '.' ) if len(submodule.split('.' ) ) == 1: submodules.append(__SCREAMING_SNAKE_CASE ) return submodules UpperCAmelCase = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', '''models.esm.openfold_utils''', ] def UpperCAmelCase_ ( ): # This is to make sure the transformers module imported is the one in the repo. from transformers.utils import direct_transformers_import lowercase = direct_transformers_import(__SCREAMING_SNAKE_CASE ) lowercase = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(__SCREAMING_SNAKE_CASE , '__init__.py' ) , 'r' ) as f: lowercase = f.read() import_structure_keys.update(set(re.findall(r'import_structure\[\"([^\"]*)\"\]' , __SCREAMING_SNAKE_CASE ) ) ) lowercase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(__SCREAMING_SNAKE_CASE ) > 0: lowercase = '\n'.join(F'''- {module}''' for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registed in the main init of Transformers:\n' F'''{list_of_modules}\n''' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
84
1
from __future__ import annotations class A_ : '''simple docstring''' def __init__( self , snake_case ): lowercase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(snake_case ) != 0: lowercase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(snake_case ) != cols: raise error for value in row: if not isinstance(snake_case , (int, float) ): raise error lowercase = rows else: lowercase = [] def SCREAMING_SNAKE_CASE__ ( self ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows ) @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows[0] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return (self.num_rows, self.num_columns) @property def SCREAMING_SNAKE_CASE__ ( self ): return self.order[0] == self.order[1] def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def SCREAMING_SNAKE_CASE__ ( self ): return bool(self.determinant() ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(snake_case ).determinant() def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): if (row + column) % 2 == 0: return self.get_minor(snake_case , snake_case ) return -1 * self.get_minor(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [self.get_minor(snake_case , snake_case ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self ): return str(self.rows ) def __str__( self ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(snake_case ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in row: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(snake_case ) else: lowercase = self.rows[0:position] + [row] + self.rows[position:] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in column: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: lowercase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: lowercase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self , snake_case ): if not isinstance(snake_case , snake_case ): return NotImplemented return self.rows == other.rows def __ne__( self , snake_case ): return not self == other def __neg__( self ): return self * -1 def __add__( self , snake_case ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self , snake_case ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self , snake_case ): if isinstance(snake_case , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(snake_case , snake_case ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(snake_case , snake_case ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self , snake_case ): if not isinstance(snake_case , snake_case ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) lowercase = self for _ in range(other - 1 ): result *= self return result @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , snake_case ): return sum(row[i] * column[i] for i in range(len(snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
84
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar UpperCAmelCase = TypeVar('''T''') class A_ ( Generic[T] ): '''simple docstring''' def __init__( self , snake_case ): lowercase = data lowercase = None def __str__( self ): return F'''{self.data}''' class A_ ( Generic[T] ): '''simple docstring''' def __init__( self ): lowercase = None def __iter__( self ): lowercase = self.top while node: yield node.data lowercase = node.next def __str__( self ): return "->".join([str(snake_case ) for item in self] ) def __len__( self ): return len(tuple(iter(self ) ) ) def SCREAMING_SNAKE_CASE__ ( self ): return self.top is None def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = Node(snake_case ) if not self.is_empty(): lowercase = self.top lowercase = node def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top , snake_case ) lowercase = self.top lowercase = self.top.next return pop_node.data def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def SCREAMING_SNAKE_CASE__ ( self ): lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
84
1
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class A_ ( unittest.TestCase ): '''simple docstring''' @property def SCREAMING_SNAKE_CASE__ ( self ): torch.manual_seed(0 ) lowercase = UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=('DownBlock2D', 'AttnDownBlock2D') , up_block_types=('AttnUpBlock2D', 'UpBlock2D') , ) return model def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.dummy_uncond_unet lowercase = ScoreSdeVeScheduler() lowercase = ScoreSdeVePipeline(unet=snake_case , scheduler=snake_case ) sde_ve.to(snake_case ) sde_ve.set_progress_bar_config(disable=snake_case ) lowercase = torch.manual_seed(0 ) lowercase = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=snake_case ).images lowercase = torch.manual_seed(0 ) lowercase = sde_ve(num_inference_steps=2 , output_type='numpy' , generator=snake_case , return_dict=snake_case )[ 0 ] lowercase = image[0, -3:, -3:, -1] lowercase = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowercase = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): lowercase = 'google/ncsnpp-church-256' lowercase = UNetaDModel.from_pretrained(snake_case ) lowercase = ScoreSdeVeScheduler.from_pretrained(snake_case ) lowercase = ScoreSdeVePipeline(unet=snake_case , scheduler=snake_case ) sde_ve.to(snake_case ) sde_ve.set_progress_bar_config(disable=snake_case ) lowercase = torch.manual_seed(0 ) lowercase = sde_ve(num_inference_steps=10 , output_type='numpy' , generator=snake_case ).images lowercase = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) lowercase = np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
84
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=False , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length] ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self ): return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = LlamaModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case ) lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = True lowercase = LlamaModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , ) lowercase = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = LlamaForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = True lowercase = True lowercase = LlamaForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , use_cache=snake_case , ) lowercase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowercase = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase = torch.cat([input_mask, next_mask] , dim=-1 ) lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , output_hidden_states=snake_case , )['hidden_states'][0] lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )['hidden_states'][0] # select random slice lowercase = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase = output_from_no_past[:, -3:, random_slice_idx].detach() lowercase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Tuple = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () _UpperCamelCase : List[Any] = (LlamaForCausalLM,) if is_torch_available() else () _UpperCamelCase : int = ( { """feature-extraction""": LlamaModel, """text-classification""": LlamaForSequenceClassification, """text-generation""": LlamaForCausalLM, """zero-shot""": LlamaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase : int = False _UpperCamelCase : int = False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = LlamaModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase = type self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = 'single_label_classification' lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = 'multi_label_classification' lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @parameterized.expand([('linear',), ('dynamic',)] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = ids_tensor([1, 10] , config.vocab_size ) lowercase = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase = LlamaModel(snake_case ) original_model.to(snake_case ) original_model.eval() lowercase = original_model(snake_case ).last_hidden_state lowercase = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase = {'type': scaling_type, 'factor': 10.0} lowercase = LlamaModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() lowercase = scaled_model(snake_case ).last_hidden_state lowercase = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) lowercase = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-6.6_550, -4.1_227, -4.9_859, -3.2_406, 0.8_262, -3.0_033, 1.2_964, -3.3_699]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-12.8_281, -7.4_453, -0.4_639, -8.0_625, -7.2_500, -8.0_000, -6.4_883, -7.7_695, -7.8_438, -7.0_312, -6.2_188, -7.1_328, -1.8_496, 1.9_961, -8.6_250, -6.7_227, -12.8_281, -6.9_492, -7.0_742, -7.7_852, -7.5_820, -7.9_062, -6.9_375, -7.9_805, -8.3_438, -8.1_562, -8.0_469, -7.6_250, -7.7_422, -7.3_398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-2.0_622, -1.2_794, -1.1_638, -0.9_788, -1.4_603, -1.0_238, -1.7_893, -1.4_411]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-8.1_406, -8.0_547, 2.7_461, -1.2_344, -0.1_448, -1.8_262, -1.0_020, -1.8_154, -1.6_895, -1.8_516, -2.3_574, -0.9_277, 3.7_598, 6.5_742, -1.2_998, -0.1_177, -8.1_406, -2.9_688, -2.9_199, -3.1_699, -3.5_254, -2.3_555, -2.7_988, -3.4_141, -2.8_262, -4.5_195, -3.3_379, -3.3_164, -2.7_832, -3.0_273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-0.8_562, -1.8_520, -0.7_551, -0.4_162, -1.5_161, -1.2_038, -2.4_823, -2.3_254]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-2.2_227, 4.8_828, 0.9_023, -0.4_578, -0.7_871, -0.1_033, -0.6_221, -0.5_786, -0.7_803, -1.0_674, -1.2_920, -0.1_570, 0.8_008, 2.0_723, -0.9_497, 0.2_771, -2.2_227, -0.7_612, -1.4_346, -1.2_061, -1.6_426, -0.3_000, -0.7_139, -1.1_934, -1.8_691, -1.6_973, -1.5_947, -1.2_705, -0.3_523, -0.5_513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) lowercase = torch.tensor( [[-4.2_327, -3.3_360, -4.6_665, -4.7_631, -1.8_180, -3.4_170, -1.4_211, -3.1_810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # fmt: off lowercase = torch.tensor([-9.4_922, -3.9_551, 1.7_998, -5.6_758, -5.1_055, -5.8_984, -4.8_320, -6.8_086, -6.5_391, -5.6_172, -5.5_820, -5.5_352, 1.7_881, 3.6_289, -6.5_117, -3.4_785, -9.5_000, -6.0_352, -6.8_125, -6.0_195, -6.6_836, -5.4_727, -6.2_812, -6.0_391, -7.3_398, -7.4_297, -7.4_844, -6.5_820, -5.8_789, -5.5_312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' lowercase = 'Simply put, the theory of relativity states that ' lowercase = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) lowercase = tokenizer.encode(snake_case , return_tensors='pt' ) lowercase = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=snake_case ) # greedy generation outputs lowercase = model.generate(snake_case , max_new_tokens=64 , top_p=snake_case , temperature=1 , do_sample=snake_case ) lowercase = tokenizer.decode(generated_ids[0] , skip_special_tokens=snake_case ) self.assertEqual(snake_case , snake_case )
84
1
UpperCAmelCase = { 0: '''0''', 1: '''1''', 2: '''2''', 3: '''3''', 4: '''4''', 5: '''5''', 6: '''6''', 7: '''7''', 8: '''8''', 9: '''9''', 10: '''a''', 11: '''b''', 12: '''c''', 13: '''d''', 14: '''e''', 15: '''f''', } def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): assert type(__SCREAMING_SNAKE_CASE ) in (int, float) and decimal == int(__SCREAMING_SNAKE_CASE ) lowercase = int(__SCREAMING_SNAKE_CASE ) lowercase = '' lowercase = False if decimal < 0: lowercase = True decimal *= -1 while decimal > 0: lowercase , lowercase = divmod(__SCREAMING_SNAKE_CASE , 16 ) lowercase = values[remainder] + hexadecimal lowercase = '0x' + hexadecimal if negative: lowercase = '-' + hexadecimal return hexadecimal if __name__ == "__main__": import doctest doctest.testmod()
84
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version UpperCAmelCase = get_logger(__name__) class A_ : '''simple docstring''' _UpperCamelCase : Dict = """dummy_data""" _UpperCamelCase : Optional[int] = """datasets""" _UpperCamelCase : Tuple = False def __init__( self , snake_case , snake_case , snake_case , snake_case = None , snake_case = False , snake_case = True , snake_case = None , ): lowercase = 0 lowercase = dataset_name lowercase = cache_dir lowercase = use_local_dummy_data lowercase = config # download_callbacks take a single url as input lowercase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root lowercase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general lowercase = str(snake_case ) # to be downloaded lowercase = None lowercase = None @property def SCREAMING_SNAKE_CASE__ ( self ): if self._dummy_file is None: lowercase = self.download_dummy_data() return self._dummy_file @property def SCREAMING_SNAKE_CASE__ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def SCREAMING_SNAKE_CASE__ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) lowercase = cached_path( snake_case , cache_dir=self.cache_dir , extract_compressed_file=snake_case , force_extract=snake_case ) return os.path.join(snake_case , self.dummy_file_name ) @property def SCREAMING_SNAKE_CASE__ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def SCREAMING_SNAKE_CASE__ ( self ): if self._bucket_url is None: lowercase = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def SCREAMING_SNAKE_CASE__ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case ): if self.load_existing_dummy_data: # dummy data is downloaded and tested lowercase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned lowercase = self.dummy_file_name # special case when data_url is a dict if isinstance(snake_case , snake_case ): return self.create_dummy_data_dict(snake_case , snake_case ) elif isinstance(snake_case , (list, tuple) ): return self.create_dummy_data_list(snake_case , snake_case ) else: return self.create_dummy_data_single(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case ): return self.download_and_extract(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): return self.download_and_extract(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case , **snake_case ): return path def SCREAMING_SNAKE_CASE__ ( self ): return {} def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(snake_case , snake_case ): for single_url in single_urls: download_callback(snake_case ) else: lowercase = single_urls download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(snake_case , snake_case ): lowercase = [os.path.join(snake_case , urllib.parse.quote_plus(Path(snake_case ).name ) ) for x in single_urls] else: lowercase = single_urls lowercase = os.path.join(snake_case , urllib.parse.quote_plus(Path(snake_case ).name ) ) lowercase = value # make sure that values are unique if all(isinstance(snake_case , snake_case ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique lowercase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one lowercase = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , snake_case ) ) for url in data_url ) lowercase = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): lowercase = [data_url[0]] * len(snake_case ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowercase = os.path.join(snake_case , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(snake_case ) return dummy_data_list def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): for download_callback in self.download_callbacks: download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowercase = os.path.join(snake_case , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(snake_case ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self , snake_case ): def _iter_archive_members(snake_case ): # this preserves the order of the members inside the ZIP archive lowercase = Path(self.dummy_file ).parent lowercase = path.relative_to(snake_case ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: lowercase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(snake_case ) lowercase = Path(snake_case ) lowercase = _iter_archive_members(snake_case ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(snake_case ).as_posix(), file_path.open('rb' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if not isinstance(snake_case , snake_case ): lowercase = [paths] for path in paths: if os.path.isfile(snake_case ): if os.path.basename(snake_case ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(snake_case ): if os.path.basename(snake_case ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(snake_case ): if filename.startswith(('.', '__') ): continue yield os.path.join(snake_case , snake_case )
84
1
import requests UpperCAmelCase = '''''' # <-- Put your OpenWeatherMap appid here! UpperCAmelCase = '''https://api.openweathermap.org/data/2.5/''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = "Chicago" , __SCREAMING_SNAKE_CASE = APPID ): return requests.get(URL_BASE + 'weather' , params=locals() ).json() def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = "Kolkata, India" , __SCREAMING_SNAKE_CASE = APPID ): return requests.get(URL_BASE + 'forecast' , params=locals() ).json() def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 55.68 , __SCREAMING_SNAKE_CASE = 12.57 , __SCREAMING_SNAKE_CASE = APPID ): return requests.get(URL_BASE + 'onecall' , params=locals() ).json() if __name__ == "__main__": from pprint import pprint while True: UpperCAmelCase = input('''Enter a location:''').strip() if location: pprint(current_weather(location)) else: break
84
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Tuple = OpenAIGPTTokenizer _UpperCamelCase : List[Any] = OpenAIGPTTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : List[Any] = False def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowercase = dict(zip(snake_case , range(len(snake_case ) ) ) ) lowercase = ['#version: 0.2', 'l o', 'lo w', 'e r</w>', ''] lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(snake_case ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(snake_case ) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): return "lower newer", "lower newer" def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) lowercase = 'lower' lowercase = ['low', 'er</w>'] lowercase = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , snake_case ) lowercase = tokens + ['<unk>'] lowercase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase = self.rust_tokenizer_class.from_pretrained(snake_case , **snake_case ) # Simple input lowercase = 'This is a simple input' lowercase = ['This is a simple input 1', 'This is a simple input 2'] lowercase = ('This is a simple input', 'This is a pair') lowercase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding='max_length' ) # Simple input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding='max_length' ) # Simple input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding='max_length' , ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding='max_length' ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding='max_length' ) # Pair input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding='max_length' , ) def SCREAMING_SNAKE_CASE__ ( self ): pass @require_ftfy @require_spacy @require_tokenizers class A_ ( __lowerCamelCase ): '''simple docstring''' pass
84
1
import logging import sys from dataclasses import dataclass, field from typing import Any, Dict, List, Optional, Union import librosa import torch from datasets import DatasetDict, load_dataset from packaging import version from torch import nn from transformers import ( HfArgumentParser, Trainer, TrainingArguments, WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForPreTraining, is_apex_available, trainer_utils, ) from transformers.models.wavaveca.modeling_wavaveca import _compute_mask_indices if is_apex_available(): from apex import amp if version.parse(version.parse(torch.__version__).base_version) >= version.parse('''1.6'''): UpperCAmelCase = True from torch.cuda.amp import autocast UpperCAmelCase = logging.getLogger(__name__) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : str = field( metadata={"""help""": """Path to pretrained model or model identifier from huggingface.co/models"""} ) _UpperCamelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """Where do you want to store the pretrained models downloaded from huggingface.co"""} , ) _UpperCamelCase : Optional[bool] = field( default=__lowerCamelCase , metadata={"""help""": """Whether to freeze the feature extractor layers of the model."""} ) _UpperCamelCase : Optional[bool] = field( default=__lowerCamelCase , metadata={"""help""": """Whether to log verbose messages or not."""} , ) _UpperCamelCase : Optional[float] = field( default=2.0 , metadata={"""help""": """Maximum temperature for gumbel softmax."""} ) _UpperCamelCase : Optional[float] = field( default=0.5 , metadata={"""help""": """Minimum temperature for gumbel softmax."""} ) _UpperCamelCase : Optional[float] = field( default=0.999995 , metadata={"""help""": """Decay of gumbel temperature during training."""} ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) lowercase = logging.WARNING if model_args.verbose_logging: lowercase = logging.DEBUG elif trainer_utils.is_main_process(training_args.local_rank ): lowercase = logging.INFO logger.setLevel(__SCREAMING_SNAKE_CASE ) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : str = field( default=__lowerCamelCase , metadata={"""help""": """The name of the dataset to use (via the datasets library)."""} ) _UpperCamelCase : Optional[str] = field( default=__lowerCamelCase , metadata={"""help""": """The configuration name of the dataset to use (via the datasets library)."""} ) _UpperCamelCase : Optional[str] = field( default="""train""" , metadata={ """help""": """The name of the training data set split to use (via the datasets library). Defaults to 'train'""" } , ) _UpperCamelCase : Optional[str] = field( default="""validation""" , metadata={ """help""": ( """The name of the validation data set split to use (via the datasets library). Defaults to 'validation'""" ) } , ) _UpperCamelCase : Optional[str] = field( default="""file""" , metadata={"""help""": """Column in the dataset that contains speech file path. Defaults to 'file'"""} , ) _UpperCamelCase : bool = field( default=__lowerCamelCase , metadata={"""help""": """Overwrite the cached preprocessed datasets or not."""} ) _UpperCamelCase : Optional[int] = field( default=1 , metadata={ """help""": """The percentage of the train set used as validation set in case there's no validation split""" } , ) _UpperCamelCase : Optional[int] = field( default=__lowerCamelCase , metadata={"""help""": """The number of processes to use for the preprocessing."""} , ) _UpperCamelCase : Optional[float] = field( default=20.0 , metadata={"""help""": """Filter audio files that are longer than `max_duration_in_seconds` seconds"""} ) @dataclass class A_ : '''simple docstring''' _UpperCamelCase : WavaVecaForPreTraining _UpperCamelCase : WavaVecaFeatureExtractor _UpperCamelCase : Union[bool, str] = "longest" _UpperCamelCase : Optional[int] = None _UpperCamelCase : Optional[int] = None def __call__( self , snake_case ): # reformat list to dict and set to pytorch format lowercase = self.feature_extractor.pad( snake_case , max_length=self.max_length , padding=self.padding , pad_to_multiple_of=self.pad_to_multiple_of , return_tensors='pt' , ) lowercase = self.model._get_feat_extract_output_lengths(batch['input_values'].shape[-1] ) lowercase = batch['input_values'].shape[0] # make sure that no loss is computed on padded inputs if batch["attention_mask"] is not None: # compute real output lengths according to convolution formula lowercase = self.model._get_feat_extract_output_lengths(batch['attention_mask'].sum(-1 ) ).to( torch.long ) lowercase = torch.zeros( (batch_size, mask_indices_seq_length) , dtype=torch.long , device=batch['input_values'].device ) # these two operations makes sure that all values # before the output lengths indices are attended to lowercase = 1 lowercase = attention_mask.flip([-1] ).cumsum(-1 ).flip([-1] ).bool() # sample randomly masked indices lowercase = _compute_mask_indices( (batch_size, mask_indices_seq_length) , self.model.config.mask_time_prob , self.model.config.mask_time_length , attention_mask=snake_case , min_masks=2 , ) return batch class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , *snake_case , snake_case=1 , snake_case=0 , snake_case=1.0 , **snake_case ): super().__init__(*snake_case , **snake_case ) lowercase = 0 lowercase = max_gumbel_temp lowercase = min_gumbel_temp lowercase = gumbel_temp_decay def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): model.train() lowercase = self._prepare_inputs(snake_case ) if self.use_amp: with autocast(): lowercase = self.compute_loss(snake_case , snake_case ) else: lowercase = self.compute_loss(snake_case , snake_case ) if self.args.n_gpu > 1 or self.deepspeed: if model.module.config.ctc_loss_reduction == "mean": lowercase = loss.mean() elif model.module.config.ctc_loss_reduction == "sum": lowercase = loss.sum() / (inputs['mask_time_indices']).sum() else: raise ValueError(F'''{model.config.ctc_loss_reduction} is not valid. Choose one of [\'mean\', \'sum\']''' ) if self.args.gradient_accumulation_steps > 1: lowercase = loss / self.args.gradient_accumulation_steps if self.use_amp: self.scaler.scale(snake_case ).backward() elif self.use_apex: with amp.scale_loss(snake_case , self.optimizer ) as scaled_loss: scaled_loss.backward() elif self.deepspeed: self.deepspeed.backward(snake_case ) else: loss.backward() self.num_update_step += 1 # make sure gumbel softmax temperature is decayed if self.args.n_gpu > 1 or self.deepspeed: model.module.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) else: model.set_gumbel_temperature( max(self.max_gumbel_temp * self.gumbel_temp_decay**self.num_update_step , self.min_gumbel_temp ) ) return loss.detach() def UpperCAmelCase_ ( ): # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. lowercase = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase , lowercase , lowercase = parser.parse_args_into_dataclasses() configure_logger(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Downloading and loading a dataset from the hub. lowercase = load_dataset(data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) if "validation" not in datasets.keys(): # make sure only "validation" and "train" keys remain" lowercase = DatasetDict() lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[:{data_args.validation_split_percentage}%]''' , cache_dir=model_args.cache_dir , ) lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}[{data_args.validation_split_percentage}%:]''' , cache_dir=model_args.cache_dir , ) else: # make sure only "validation" and "train" keys remain" lowercase = DatasetDict() lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split='validation' , cache_dir=model_args.cache_dir , ) lowercase = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=F'''{data_args.train_split_name}''' , cache_dir=model_args.cache_dir , ) # only normalized-inputs-training is supported lowercase = WavaVecaFeatureExtractor.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , do_normalize=__SCREAMING_SNAKE_CASE ) def prepare_dataset(__SCREAMING_SNAKE_CASE ): # check that all files have the correct sampling rate lowercase , lowercase = librosa.load(batch[data_args.speech_file_column] , sr=feature_extractor.sampling_rate ) return batch # load audio files into numpy arrays lowercase = datasets.map( __SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , remove_columns=datasets['train'].column_names ) # filter audio files that are too long lowercase = vectorized_datasets.filter( lambda __SCREAMING_SNAKE_CASE : len(data['speech'] ) < int(data_args.max_duration_in_seconds * feature_extractor.sampling_rate ) ) def normalize(__SCREAMING_SNAKE_CASE ): return feature_extractor(batch['speech'] , sampling_rate=feature_extractor.sampling_rate ) # normalize and transform to `BatchFeatures` lowercase = vectorized_datasets.map( __SCREAMING_SNAKE_CASE , batched=__SCREAMING_SNAKE_CASE , num_proc=data_args.preprocessing_num_workers , load_from_cache_file=not data_args.overwrite_cache , remove_columns=vectorized_datasets['train'].column_names , ) # pretraining is only supported for "newer" stable layer norm architecture # apply_spec_augment has to be True, mask_feature_prob has to be 0.0 lowercase = WavaVecaConfig.from_pretrained( model_args.model_name_or_path , cache_dir=model_args.cache_dir , gradient_checkpointing=training_args.gradient_checkpointing , ) if not config.do_stable_layer_norm or config.feat_extract_norm != "layer": raise ValueError( 'PreTraining is only supported for ``config.do_stable_layer_norm=True`` and' ' ``config.feat_extract_norm=\'layer\'' ) lowercase = WavaVecaForPreTraining(__SCREAMING_SNAKE_CASE ) lowercase = DataCollatorForWavaVecaPretraining(model=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE ) lowercase = WavaVecaPreTrainer( model=__SCREAMING_SNAKE_CASE , data_collator=__SCREAMING_SNAKE_CASE , args=__SCREAMING_SNAKE_CASE , train_dataset=vectorized_datasets['train'] , eval_dataset=vectorized_datasets['validation'] , tokenizer=__SCREAMING_SNAKE_CASE , max_gumbel_temp=model_args.max_gumbel_temperature , min_gumbel_temp=model_args.min_gumbel_temperature , gumbel_temp_decay=model_args.gumbel_temperature_decay , ) trainer.train() if __name__ == "__main__": main()
84
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. UpperCAmelCase = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): config.addinivalue_line( 'markers' , 'is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested' ) config.addinivalue_line( 'markers' , 'is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested' ) config.addinivalue_line('markers' , 'is_pipeline_test: mark test to run only when pipelines are tested' ) config.addinivalue_line('markers' , 'is_staging_test: mark test to run only in the staging environment' ) config.addinivalue_line('markers' , 'accelerate_tests: mark test that require accelerate' ) config.addinivalue_line('markers' , 'tool_tests: mark the tool tests that are run on their specific schedule' ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_terminal_summary_main lowercase = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(__SCREAMING_SNAKE_CASE , id=__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: lowercase = 0 # Doctest custom flag to ignore output. UpperCAmelCase = doctest.register_optionflag('''IGNORE_RESULT''') UpperCAmelCase = doctest.OutputChecker class A_ ( __lowerCamelCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , snake_case , snake_case , snake_case ) UpperCAmelCase = CustomOutputChecker UpperCAmelCase = HfDoctestModule UpperCAmelCase = HfDocTestParser
84
1
import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = {'''vocab_file''': '''sentencepiece.model'''} UpperCAmelCase = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } UpperCAmelCase = { '''google/rembert''': 256, } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[Any] = VOCAB_FILES_NAMES _UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self , snake_case , snake_case=False , snake_case=True , snake_case=True , snake_case="[CLS]" , snake_case="[SEP]" , snake_case="[UNK]" , snake_case="[SEP]" , snake_case="[PAD]" , snake_case="[CLS]" , snake_case="[MASK]" , **snake_case , ): super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , **snake_case , ) lowercase = do_lower_case lowercase = remove_space lowercase = keep_accents lowercase = vocab_file lowercase = spm.SentencePieceProcessor() self.sp_model.Load(snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ): lowercase = self.__dict__.copy() lowercase = None return state def __setstate__( self , snake_case ): lowercase = d lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=False ): lowercase = self.sp_model.EncodeAsPieces(snake_case ) return pieces def SCREAMING_SNAKE_CASE__ ( self , snake_case ): return self.sp_model.PieceToId(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): return self.sp_model.IdToPiece(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = self.sp_model.decode_pieces(snake_case ) return out_string def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None , snake_case = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( 'You should not supply a second sequence if the provided sequence of ' 'ids is already formatted with special tokens for the model.' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1] return [1] + ([0] * len(snake_case )) + [1] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): if not os.path.isdir(snake_case ): logger.error('Vocabulary path ({}) should be a directory'.format(snake_case ) ) return lowercase = os.path.join( snake_case , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ): copyfile(self.vocab_file , snake_case ) return (out_vocab_file,)
84
import torch from torch import nn class A_ ( nn.Module ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , snake_case , snake_case=1 , snake_case=False ): super().__init__() lowercase = n_token lowercase = d_embed lowercase = d_proj lowercase = cutoffs + [n_token] lowercase = [0] + self.cutoffs lowercase = div_val lowercase = self.cutoffs[0] lowercase = len(self.cutoffs ) - 1 lowercase = self.shortlist_size + self.n_clusters if self.n_clusters > 0: lowercase = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) lowercase = nn.Parameter(torch.zeros(self.n_clusters ) ) lowercase = nn.ModuleList() lowercase = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(snake_case , snake_case ) ) ) else: self.out_projs.append(snake_case ) self.out_layers.append(nn.Linear(snake_case , snake_case ) ) else: for i in range(len(self.cutoffs ) ): lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(snake_case , snake_case ) ) ) self.out_layers.append(nn.Linear(snake_case , r_idx - l_idx ) ) lowercase = keep_order def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): if proj is None: lowercase = nn.functional.linear(snake_case , snake_case , bias=snake_case ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: lowercase = nn.functional.linear(snake_case , proj.t().contiguous() ) lowercase = nn.functional.linear(snake_case , snake_case , bias=snake_case ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=None , snake_case=False ): if labels is not None: # Shift so that tokens < n predict n lowercase = hidden[..., :-1, :].contiguous() lowercase = labels[..., 1:].contiguous() lowercase = hidden.view(-1 , hidden.size(-1 ) ) lowercase = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError('Input and labels should have the same size in the batch dimension.' ) else: lowercase = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: lowercase = self._compute_logit(snake_case , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: lowercase = labels != -100 lowercase = torch.zeros_like(snake_case , dtype=hidden.dtype , device=hidden.device ) lowercase = ( -nn.functional.log_softmax(snake_case , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: lowercase = nn.functional.log_softmax(snake_case , dim=-1 ) else: # construct weights and biases lowercase , lowercase = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = self.out_layers[0].weight[l_idx:r_idx] lowercase = self.out_layers[0].bias[l_idx:r_idx] else: lowercase = self.out_layers[i].weight lowercase = self.out_layers[i].bias if i == 0: lowercase = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowercase = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(snake_case ) biases.append(snake_case ) lowercase , lowercase , lowercase = weights[0], biases[0], self.out_projs[0] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) if labels is None: lowercase = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: lowercase = torch.zeros_like(snake_case , dtype=hidden.dtype , device=hidden.device ) lowercase = 0 lowercase = [0] + self.cutoffs for i in range(len(snake_case ) - 1 ): lowercase , lowercase = cutoff_values[i], cutoff_values[i + 1] if labels is not None: lowercase = (labels >= l_idx) & (labels < r_idx) lowercase = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue lowercase = labels.index_select(0 , snake_case ) - l_idx lowercase = head_logprob.index_select(0 , snake_case ) lowercase = hidden.index_select(0 , snake_case ) else: lowercase = hidden if i == 0: if labels is not None: lowercase = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: lowercase = head_logprob[:, : self.cutoffs[0]] else: lowercase , lowercase , lowercase = weights[i], biases[i], self.out_projs[i] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: lowercase = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: lowercase = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i lowercase = logprob_i if labels is not None: if (hasattr(self , 'keep_order' ) and self.keep_order) or keep_order: out.index_copy_(0 , snake_case , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if self.n_clusters == 0: lowercase = self._compute_logit(snake_case , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(snake_case , dim=-1 ) else: # construct weights and biases lowercase , lowercase = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = self.out_layers[0].weight[l_idx:r_idx] lowercase = self.out_layers[0].bias[l_idx:r_idx] else: lowercase = self.out_layers[i].weight lowercase = self.out_layers[i].bias if i == 0: lowercase = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowercase = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(snake_case ) biases.append(snake_case ) lowercase , lowercase , lowercase = weights[0], biases[0], self.out_projs[0] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = hidden.new_empty((head_logit.size(0 ), self.n_token) ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = [0] + self.cutoffs for i in range(len(snake_case ) - 1 ): lowercase , lowercase = cutoff_values[i], cutoff_values[i + 1] if i == 0: lowercase = head_logprob[:, : self.cutoffs[0]] else: lowercase , lowercase , lowercase = weights[i], biases[i], self.out_projs[i] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = head_logprob[:, -i] + tail_logprob_i lowercase = logprob_i return out
84
1
from __future__ import annotations import unittest from transformers import DistilBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.distilbert.modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertModel, ) class A_ : '''simple docstring''' def __init__( self , snake_case , ): lowercase = parent lowercase = 13 lowercase = 7 lowercase = True lowercase = True lowercase = False lowercase = True lowercase = 99 lowercase = 32 lowercase = 2 lowercase = 4 lowercase = 37 lowercase = 'gelu' lowercase = 0.1 lowercase = 0.1 lowercase = 512 lowercase = 16 lowercase = 2 lowercase = 0.02 lowercase = 3 lowercase = 4 lowercase = None def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length] ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFDistilBertModel(config=snake_case ) lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} lowercase = model(snake_case ) lowercase = [input_ids, input_mask] lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFDistilBertForMaskedLM(config=snake_case ) lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} lowercase = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFDistilBertForQuestionAnswering(config=snake_case ) lowercase = { 'input_ids': input_ids, 'attention_mask': input_mask, } lowercase = model(snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = TFDistilBertForSequenceClassification(snake_case ) lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} lowercase = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_choices lowercase = TFDistilBertForMultipleChoice(snake_case ) lowercase = tf.tile(tf.expand_dims(snake_case , 1 ) , (1, self.num_choices, 1) ) lowercase = tf.tile(tf.expand_dims(snake_case , 1 ) , (1, self.num_choices, 1) ) lowercase = { 'input_ids': multiple_choice_inputs_ids, 'attention_mask': multiple_choice_input_mask, } lowercase = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = TFDistilBertForTokenClassification(snake_case ) lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} lowercase = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ((lowercase) , (lowercase) , (lowercase) , (lowercase) , (lowercase) , (lowercase)) = config_and_inputs lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class A_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : List[Any] = ( ( TFDistilBertModel, TFDistilBertForMaskedLM, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertForMultipleChoice, ) if is_tf_available() else None ) _UpperCamelCase : int = ( { """feature-extraction""": TFDistilBertModel, """fill-mask""": TFDistilBertForMaskedLM, """question-answering""": TFDistilBertForQuestionAnswering, """text-classification""": TFDistilBertForSequenceClassification, """token-classification""": TFDistilBertForTokenClassification, """zero-shot""": TFDistilBertForSequenceClassification, } if is_tf_available() else {} ) _UpperCamelCase : List[str] = False _UpperCamelCase : Dict = False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = TFDistilBertModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , dim=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in list(TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1] ): lowercase = TFDistilBertModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_tf class A_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = TFDistilBertModel.from_pretrained('distilbert-base-uncased' ) lowercase = tf.constant([[0, 1, 2, 3, 4, 5]] ) lowercase = model(snake_case )[0] lowercase = [1, 6, 768] self.assertEqual(output.shape , snake_case ) lowercase = tf.constant( [ [ [0.19_261_885, -0.13_732_955, 0.4_119_799], [0.22_150_156, -0.07_422_661, 0.39_037_204], [0.22_756_018, -0.0_896_414, 0.3_701_467], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case , atol=1E-4 )
84
from __future__ import annotations class A_ : '''simple docstring''' def __init__( self , snake_case ): lowercase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(snake_case ) != 0: lowercase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(snake_case ) != cols: raise error for value in row: if not isinstance(snake_case , (int, float) ): raise error lowercase = rows else: lowercase = [] def SCREAMING_SNAKE_CASE__ ( self ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows ) @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows[0] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return (self.num_rows, self.num_columns) @property def SCREAMING_SNAKE_CASE__ ( self ): return self.order[0] == self.order[1] def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def SCREAMING_SNAKE_CASE__ ( self ): return bool(self.determinant() ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(snake_case ).determinant() def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): if (row + column) % 2 == 0: return self.get_minor(snake_case , snake_case ) return -1 * self.get_minor(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [self.get_minor(snake_case , snake_case ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self ): return str(self.rows ) def __str__( self ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(snake_case ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in row: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(snake_case ) else: lowercase = self.rows[0:position] + [row] + self.rows[position:] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in column: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: lowercase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: lowercase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self , snake_case ): if not isinstance(snake_case , snake_case ): return NotImplemented return self.rows == other.rows def __ne__( self , snake_case ): return not self == other def __neg__( self ): return self * -1 def __add__( self , snake_case ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self , snake_case ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self , snake_case ): if isinstance(snake_case , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(snake_case , snake_case ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(snake_case , snake_case ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self , snake_case ): if not isinstance(snake_case , snake_case ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) lowercase = self for _ in range(other - 1 ): result *= self return result @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , snake_case ): return sum(row[i] * column[i] for i in range(len(snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
84
1
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType UpperCAmelCase = None UpperCAmelCase = '''<''' if sys.byteorder == '''little''' else '''>''' # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image UpperCAmelCase = [ np.dtype('''|b1'''), np.dtype('''|u1'''), np.dtype('''<u2'''), np.dtype('''>u2'''), np.dtype('''<i2'''), np.dtype('''>i2'''), np.dtype('''<u4'''), np.dtype('''>u4'''), np.dtype('''<i4'''), np.dtype('''>i4'''), np.dtype('''<f4'''), np.dtype('''>f4'''), np.dtype('''<f8'''), np.dtype('''>f8'''), ] @dataclass class A_ : '''simple docstring''' _UpperCamelCase : bool = True _UpperCamelCase : Optional[str] = None # Automatically constructed _UpperCamelCase : ClassVar[str] = "PIL.Image.Image" _UpperCamelCase : ClassVar[Any] = pa.struct({"""bytes""": pa.binary(), """path""": pa.string()} ) _UpperCamelCase : str = field(default="""Image""" , init=__lowerCamelCase , repr=__lowerCamelCase ) def __call__( self ): return self.pa_type def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('To support encoding images, please install \'Pillow\'.' ) if isinstance(snake_case , snake_case ): lowercase = np.array(snake_case ) if isinstance(snake_case , snake_case ): return {"path": value, "bytes": None} elif isinstance(snake_case , snake_case ): return {"path": None, "bytes": value} elif isinstance(snake_case , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(snake_case ) elif isinstance(snake_case , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(snake_case ) elif value.get('path' ) is not None and os.path.isfile(value['path'] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get('path' )} elif value.get('bytes' ) is not None or value.get('path' ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get('bytes' ), "path": value.get('path' )} else: raise ValueError( F'''An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.''' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=None ): if not self.decode: raise RuntimeError('Decoding is disabled for this feature. Please use Image(decode=True) instead.' ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('To support decoding images, please install \'Pillow\'.' ) if token_per_repo_id is None: lowercase = {} lowercase , lowercase = value['path'], value['bytes'] if bytes_ is None: if path is None: raise ValueError(F'''An image should have one of \'path\' or \'bytes\' but both are None in {value}.''' ) else: if is_local_path(snake_case ): lowercase = PIL.Image.open(snake_case ) else: lowercase = path.split('::' )[-1] try: lowercase = string_to_dict(snake_case , config.HUB_DATASETS_URL )['repo_id'] lowercase = token_per_repo_id.get(snake_case ) except ValueError: lowercase = None with xopen(snake_case , 'rb' , use_auth_token=snake_case ) as f: lowercase = BytesIO(f.read() ) lowercase = PIL.Image.open(bytes_ ) else: lowercase = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def SCREAMING_SNAKE_CASE__ ( self ): from .features import Value return ( self if self.decode else { "bytes": Value('binary' ), "path": Value('string' ), } ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if pa.types.is_string(storage.type ): lowercase = pa.array([None] * len(snake_case ) , type=pa.binary() ) lowercase = pa.StructArray.from_arrays([bytes_array, storage] , ['bytes', 'path'] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): lowercase = pa.array([None] * len(snake_case ) , type=pa.string() ) lowercase = pa.StructArray.from_arrays([storage, path_array] , ['bytes', 'path'] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index('bytes' ) >= 0: lowercase = storage.field('bytes' ) else: lowercase = pa.array([None] * len(snake_case ) , type=pa.binary() ) if storage.type.get_field_index('path' ) >= 0: lowercase = storage.field('path' ) else: lowercase = pa.array([None] * len(snake_case ) , type=pa.string() ) lowercase = pa.StructArray.from_arrays([bytes_array, path_array] , ['bytes', 'path'] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): lowercase = pa.array( [encode_np_array(np.array(snake_case ) )['bytes'] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) lowercase = pa.array([None] * len(snake_case ) , type=pa.string() ) lowercase = pa.StructArray.from_arrays( [bytes_array, path_array] , ['bytes', 'path'] , mask=bytes_array.is_null() ) return array_cast(snake_case , self.pa_type ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): @no_op_if_value_is_null def path_to_bytes(snake_case ): with xopen(snake_case , 'rb' ) as f: lowercase = f.read() return bytes_ lowercase = pa.array( [ (path_to_bytes(x['path'] ) if x['bytes'] is None else x['bytes']) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) lowercase = pa.array( [os.path.basename(snake_case ) if path is not None else None for path in storage.field('path' ).to_pylist()] , type=pa.string() , ) lowercase = pa.StructArray.from_arrays([bytes_array, path_array] , ['bytes', 'path'] , mask=bytes_array.is_null() ) return array_cast(snake_case , self.pa_type ) def UpperCAmelCase_ ( ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('To support encoding images, please install \'Pillow\'.' ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() lowercase = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = BytesIO() if image.format in list_image_compression_formats(): lowercase = image.format else: lowercase = 'PNG' if image.mode in ['1', 'L', 'LA', 'RGB', 'RGBA'] else 'TIFF' image.save(__SCREAMING_SNAKE_CASE , format=__SCREAMING_SNAKE_CASE ) return buffer.getvalue() def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if hasattr(__SCREAMING_SNAKE_CASE , 'filename' ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(__SCREAMING_SNAKE_CASE )} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('To support encoding images, please install \'Pillow\'.' ) lowercase = array.dtype lowercase = dtype.byteorder if dtype.byteorder != '=' else _NATIVE_BYTEORDER lowercase = dtype.kind lowercase = dtype.itemsize lowercase = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: lowercase = np.dtype('|u1' ) if dtype_kind not in ["u", "i"]: raise TypeError( F'''Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.''' ) if dtype is not dest_dtype: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: lowercase = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: lowercase = dtype_byteorder + dtype_kind + str(__SCREAMING_SNAKE_CASE ) lowercase = np.dtype(__SCREAMING_SNAKE_CASE ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(F'''Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'''' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( F'''Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}''' ) lowercase = PIL.Image.fromarray(array.astype(__SCREAMING_SNAKE_CASE ) ) return {"path": None, "bytes": image_to_bytes(__SCREAMING_SNAKE_CASE )} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError('To support encoding images, please install \'Pillow\'.' ) if objs: lowercase , lowercase = first_non_null_value(__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(__SCREAMING_SNAKE_CASE , np.ndarray ): lowercase = no_op_if_value_is_null(__SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(__SCREAMING_SNAKE_CASE ) for obj in objs] elif isinstance(__SCREAMING_SNAKE_CASE , PIL.Image.Image ): lowercase = no_op_if_value_is_null(__SCREAMING_SNAKE_CASE ) return [obj_to_image_dict_func(__SCREAMING_SNAKE_CASE ) for obj in objs] else: return objs else: return objs
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase = ''' Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior") >>> pipe_prior.to("cuda") >>> prompt = "red cat, 4k photo" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder") >>> pipe.to("cuda") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save("cat.png") ``` ''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=8 ): lowercase = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , ): super().__init__() self.register_modules( unet=snake_case , scheduler=snake_case , movq=snake_case , ) lowercase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): if latents is None: lowercase = randn_tensor(snake_case , generator=snake_case , device=snake_case , dtype=snake_case ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase = latents.to(snake_case ) lowercase = latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) lowercase = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=snake_case ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase , lowercase = cpu_offload_with_hook(snake_case , snake_case , prev_module_hook=snake_case ) # We'll offload the last model manually. lowercase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE__ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(snake_case , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(snake_case ) def __call__( self , snake_case , snake_case , snake_case = 512 , snake_case = 512 , snake_case = 100 , snake_case = 4.0 , snake_case = 1 , snake_case = None , snake_case = None , snake_case = "pil" , snake_case = True , ): lowercase = self._execution_device lowercase = guidance_scale > 1.0 if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) lowercase = image_embeds.shape[0] * num_images_per_prompt if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) if do_classifier_free_guidance: lowercase = image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = negative_image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=snake_case ) self.scheduler.set_timesteps(snake_case , device=snake_case ) lowercase = self.scheduler.timesteps lowercase = self.unet.config.in_channels lowercase , lowercase = downscale_height_and_width(snake_case , snake_case , self.movq_scale_factor ) # create initial latent lowercase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , snake_case , snake_case , snake_case , self.scheduler , ) for i, t in enumerate(self.progress_bar(snake_case ) ): # expand the latents if we are doing classifier free guidance lowercase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase = {'image_embeds': image_embeds} lowercase = self.unet( sample=snake_case , timestep=snake_case , encoder_hidden_states=snake_case , added_cond_kwargs=snake_case , return_dict=snake_case , )[0] if do_classifier_free_guidance: lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) lowercase , lowercase = noise_pred.chunk(2 ) lowercase , lowercase = variance_pred.chunk(2 ) lowercase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase = self.scheduler.step( snake_case , snake_case , snake_case , generator=snake_case , )[0] # post-processing lowercase = self.movq.decode(snake_case , force_not_quantize=snake_case )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase = image * 0.5 + 0.5 lowercase = image.clamp(0 , 1 ) lowercase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase = self.numpy_to_pil(snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=snake_case )
84
1
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=30 , snake_case=2 , snake_case=3 , snake_case=True , snake_case=True , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=10 , snake_case=0.02 , snake_case=3 , snake_case=0.6 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = image_size lowercase = patch_size lowercase = num_channels lowercase = is_training lowercase = use_labels lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = type_sequence_label_size lowercase = initializer_range lowercase = mask_ratio lowercase = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowercase = (image_size // patch_size) ** 2 lowercase = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self ): return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): lowercase = ViTMAEModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): lowercase = ViTMAEForPreTraining(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case ) lowercase = (self.image_size // self.patch_size) ** 2 lowercase = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowercase = 1 lowercase = ViTMAEForPreTraining(snake_case ) model.to(snake_case ) model.eval() lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase = model(snake_case ) lowercase = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() lowercase , lowercase , lowercase = config_and_inputs lowercase = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () _UpperCamelCase : Union[str, Any] = {"""feature-extraction""": ViTMAEModel} if is_torch_available() else {} _UpperCamelCase : Dict = False _UpperCamelCase : int = False _UpperCamelCase : Any = False _UpperCamelCase : str = False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ViTMAEModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case , nn.Linear ) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(snake_case ) lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase = [*signature.parameters.keys()] lowercase = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): # make masks reproducible np.random.seed(2 ) lowercase = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) lowercase = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowercase = torch.from_numpy(snake_case ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowercase = pt_noise super().check_pt_tf_models(snake_case , snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase = model_class(snake_case ) model.to(snake_case ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowercase = model(**self._prepare_for_class(snake_case , snake_case ) ) lowercase = outputs[0].cpu().numpy() lowercase = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case ) lowercase = model_class.from_pretrained(snake_case ) model.to(snake_case ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowercase = model(**self._prepare_for_class(snake_case , snake_case ) ) # Make sure we don't have nans lowercase = after_outputs[0].cpu().numpy() lowercase = 0 lowercase = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(snake_case , 1E-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = ViTMAEModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def UpperCAmelCase_ ( ): lowercase = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class A_ ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE__ ( self ): return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE__ ( self ): # make random mask reproducible across the PT and TF model np.random.seed(2 ) lowercase = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(snake_case ) lowercase = self.default_image_processor lowercase = prepare_img() lowercase = image_processor(images=snake_case , return_tensors='pt' ).to(snake_case ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowercase = ViTMAEConfig() lowercase = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowercase = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): lowercase = model(**snake_case , noise=torch.from_numpy(snake_case ).to(device=snake_case ) ) # verify the logits lowercase = torch.Size((1, 196, 768) ) self.assertEqual(outputs.logits.shape , snake_case ) lowercase = torch.tensor( [[-0.0_548, -1.7_023, -0.9_325], [0.3_721, -0.5_670, -0.2_233], [0.8_235, -1.3_878, -0.3_524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(snake_case ) , atol=1E-4 ) )
84
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): if digit_amount > 0: return round(number - int(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) return number - int(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
84
1
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger UpperCAmelCase = get_logger(__name__) class A_ ( enum.Enum ): '''simple docstring''' _UpperCamelCase : Tuple = """all_checks""" _UpperCamelCase : Optional[Any] = """basic_checks""" _UpperCamelCase : str = """no_checks""" class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None ): if expected_checksums is None: logger.info('Unable to verify checksums.' ) return if len(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) ) if len(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) > 0: raise UnexpectedDownloadedFile(str(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) ) lowercase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] lowercase = ' for ' + verification_name if verification_name is not None else '' if len(__SCREAMING_SNAKE_CASE ) > 0: raise NonMatchingChecksumError( F'''Checksums didn\'t match{for_verification_name}:\n''' F'''{bad_urls}\n''' 'Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error' ) logger.info('All the checksums matched successfully' + for_verification_name ) class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' class A_ ( __lowerCamelCase ): '''simple docstring''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): if expected_splits is None: logger.info('Unable to verify splits sizes.' ) return if len(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) > 0: raise ExpectedMoreSplits(str(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) ) if len(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) > 0: raise UnexpectedSplits(str(set(__SCREAMING_SNAKE_CASE ) - set(__SCREAMING_SNAKE_CASE ) ) ) lowercase = [ {'expected': expected_splits[name], 'recorded': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(__SCREAMING_SNAKE_CASE ) > 0: raise NonMatchingSplitsSizesError(str(__SCREAMING_SNAKE_CASE ) ) logger.info('All the splits matched successfully.' ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = True ): if record_checksum: lowercase = shaaaa() with open(__SCREAMING_SNAKE_CASE , 'rb' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'' ): m.update(__SCREAMING_SNAKE_CASE ) lowercase = m.hexdigest() else: lowercase = None return {"num_bytes": os.path.getsize(__SCREAMING_SNAKE_CASE ), "checksum": checksum} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
84
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) return n == n[::-1] def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 100_0000 ): lowercase = 0 for i in range(1 , __SCREAMING_SNAKE_CASE ): if is_palindrome(__SCREAMING_SNAKE_CASE ) and is_palindrome(bin(__SCREAMING_SNAKE_CASE ).split('b' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
84
1
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = self.vocab_size - 1 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) lowercase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , head_mask=snake_case ) lowercase = model(snake_case , token_type_ids=snake_case ) lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTLMHeadModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTDoubleHeadsModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = self.num_labels lowercase = OpenAIGPTForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask, } return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _UpperCamelCase : Tuple = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _UpperCamelCase : str = ( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case=False ): lowercase = super()._prepare_for_class(snake_case , snake_case , return_labels=snake_case ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=snake_case , ) lowercase = inputs_dict['labels'] lowercase = inputs_dict['labels'] lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=snake_case , ) lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case ) return inputs_dict def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , n_embd=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = OpenAIGPTModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt' ) model.to(snake_case ) lowercase = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=snake_case ) # the president is lowercase = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 4_0477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the lowercase = model.generate(snake_case , do_sample=snake_case ) self.assertListEqual(output_ids[0].tolist() , snake_case )
84
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''microsoft/conditional-detr-resnet-50''': ( '''https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json''' ), } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[Any] = """conditional_detr""" _UpperCamelCase : Any = ["""past_key_values"""] _UpperCamelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , snake_case=True , snake_case=None , snake_case=3 , snake_case=300 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=0.0 , snake_case=0.0 , snake_case=True , snake_case="relu" , snake_case=256 , snake_case=0.1 , snake_case=0.0 , snake_case=0.0 , snake_case=0.02 , snake_case=1.0 , snake_case=False , snake_case="sine" , snake_case="resnet50" , snake_case=True , snake_case=False , snake_case=2 , snake_case=5 , snake_case=2 , snake_case=1 , snake_case=1 , snake_case=2 , snake_case=5 , snake_case=2 , snake_case=0.25 , **snake_case , ): if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) lowercase = CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(snake_case , snake_case ): lowercase = backbone_config.get('model_type' ) lowercase = CONFIG_MAPPING[backbone_model_type] lowercase = config_class.from_dict(snake_case ) lowercase = use_timm_backbone lowercase = backbone_config lowercase = num_channels lowercase = num_queries lowercase = d_model lowercase = encoder_ffn_dim lowercase = encoder_layers lowercase = encoder_attention_heads lowercase = decoder_ffn_dim lowercase = decoder_layers lowercase = decoder_attention_heads lowercase = dropout lowercase = attention_dropout lowercase = activation_dropout lowercase = activation_function lowercase = init_std lowercase = init_xavier_std lowercase = encoder_layerdrop lowercase = decoder_layerdrop lowercase = encoder_layers lowercase = auxiliary_loss lowercase = position_embedding_type lowercase = backbone lowercase = use_pretrained_backbone lowercase = dilation # Hungarian matcher lowercase = class_cost lowercase = bbox_cost lowercase = giou_cost # Loss coefficients lowercase = mask_loss_coefficient lowercase = dice_loss_coefficient lowercase = cls_loss_coefficient lowercase = bbox_loss_coefficient lowercase = giou_loss_coefficient lowercase = focal_alpha super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self ): return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE__ ( self ): return self.d_model def SCREAMING_SNAKE_CASE__ ( self ): lowercase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowercase = self.backbone_config.to_dict() lowercase = self.__class__.model_type return output class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[str] = version.parse("""1.11""" ) @property def SCREAMING_SNAKE_CASE__ ( self ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return 1E-5 @property def SCREAMING_SNAKE_CASE__ ( self ): return 12
84
1
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if not head: return True # split the list to two parts lowercase , lowercase = head.next, head while fast and fast.next: lowercase = fast.next.next lowercase = slow.next lowercase = slow.next lowercase = None # Don't forget here! But forget still works! # reverse the second part lowercase = None while second: lowercase = second.next lowercase = node lowercase = second lowercase = nxt # compare two parts # second part has the same or one less node while node: if node.val != head.val: return False lowercase = node.next lowercase = head.next return True def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if not head or not head.next: return True # 1. Get the midpoint (slow) lowercase = lowercase = lowercase = head while fast and fast.next: lowercase , lowercase = fast.next.next, slow.next # 2. Push the second half into the stack lowercase = [slow.val] while slow.next: lowercase = slow.next stack.append(slow.val ) # 3. Comparison while stack: if stack.pop() != cur.val: return False lowercase = cur.next return True def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if not head or not head.next: return True lowercase = {} lowercase = 0 while head: if head.val in d: d[head.val].append(__SCREAMING_SNAKE_CASE ) else: lowercase = [pos] lowercase = head.next pos += 1 lowercase = pos - 1 lowercase = 0 for v in d.values(): if len(__SCREAMING_SNAKE_CASE ) % 2 != 0: middle += 1 else: lowercase = 0 for i in range(0 , len(__SCREAMING_SNAKE_CASE ) ): if v[i] + v[len(__SCREAMING_SNAKE_CASE ) - 1 - step] != checksum: return False step += 1 if middle > 1: return False return True
84
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available UpperCAmelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''MLukeTokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
1
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = 0 while num > 0: digit_sum += num % 10 num //= 10 return digit_sum def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 100 ): lowercase = 1 lowercase = 2 for i in range(2 , max_n + 1 ): lowercase = pre_numerator lowercase = 2 * i // 3 if i % 3 == 0 else 1 lowercase = cur_numerator lowercase = e_cont * pre_numerator + temp return sum_digits(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(F"""{solution() = }""")
84
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = [0] * len(__SCREAMING_SNAKE_CASE ) lowercase = [] lowercase = [] lowercase = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__SCREAMING_SNAKE_CASE ) ): if indegree[i] == 0: queue.append(__SCREAMING_SNAKE_CASE ) while queue: lowercase = queue.pop(0 ) cnt += 1 topo.append(__SCREAMING_SNAKE_CASE ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__SCREAMING_SNAKE_CASE ) if cnt != len(__SCREAMING_SNAKE_CASE ): print('Cycle exists' ) else: print(__SCREAMING_SNAKE_CASE ) # Adjacency List of Graph UpperCAmelCase = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
84
1
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class A_ : '''simple docstring''' @property def SCREAMING_SNAKE_CASE__ ( self ): return self.get_dummy_input() @property def SCREAMING_SNAKE_CASE__ ( self ): if self.block_type == "down": return (4, 32, 16, 16) elif self.block_type == "mid": return (4, 32, 32, 32) elif self.block_type == "up": return (4, 32, 64, 64) raise ValueError(F'''\'{self.block_type}\' is not a supported block_type. Set it to \'up\', \'mid\', or \'down\'.''' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=True , snake_case=False , snake_case=False , snake_case=False , ): lowercase = 4 lowercase = 32 lowercase = (32, 32) lowercase = torch.manual_seed(0 ) lowercase = torch.device(snake_case ) lowercase = (batch_size, num_channels) + sizes lowercase = randn_tensor(snake_case , generator=snake_case , device=snake_case ) lowercase = {'hidden_states': hidden_states} if include_temb: lowercase = 128 lowercase = randn_tensor((batch_size, temb_channels) , generator=snake_case , device=snake_case ) if include_res_hidden_states_tuple: lowercase = torch.manual_seed(1 ) lowercase = (randn_tensor(snake_case , generator=snake_case , device=snake_case ),) if include_encoder_hidden_states: lowercase = floats_tensor((batch_size, 32, 32) ).to(snake_case ) if include_skip_sample: lowercase = randn_tensor(((batch_size, 3) + sizes) , generator=snake_case , device=snake_case ) return dummy_input def SCREAMING_SNAKE_CASE__ ( self ): lowercase = { 'in_channels': 32, 'out_channels': 32, 'temb_channels': 128, } if self.block_type == "up": lowercase = 32 if self.block_type == "mid": init_dict.pop('out_channels' ) lowercase = self.dummy_input return init_dict, inputs_dict def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase , lowercase = self.prepare_init_args_and_inputs_for_common() lowercase = self.block_class(**snake_case ) unet_block.to(snake_case ) unet_block.eval() with torch.no_grad(): lowercase = unet_block(**snake_case ) if isinstance(snake_case , snake_case ): lowercase = output[0] self.assertEqual(output.shape , self.output_shape ) lowercase = output[0, -1, -3:, -3:] lowercase = torch.tensor(snake_case ).to(snake_case ) assert torch_all_close(output_slice.flatten() , snake_case , atol=5E-3 ) @unittest.skipIf(torch_device == 'mps' , 'Training is not supported in mps' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.prepare_init_args_and_inputs_for_common() lowercase = self.block_class(**snake_case ) model.to(snake_case ) model.train() lowercase = model(**snake_case ) if isinstance(snake_case , snake_case ): lowercase = output[0] lowercase = torch.device(snake_case ) lowercase = randn_tensor(output.shape , device=snake_case ) lowercase = torch.nn.functional.mse_loss(snake_case , snake_case ) loss.backward()
84
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available UpperCAmelCase = { '''configuration_gpt_neo''': ['''GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GPTNeoConfig''', '''GPTNeoOnnxConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ '''GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GPTNeoForCausalLM''', '''GPTNeoForQuestionAnswering''', '''GPTNeoForSequenceClassification''', '''GPTNeoForTokenClassification''', '''GPTNeoModel''', '''GPTNeoPreTrainedModel''', '''load_tf_weights_in_gpt_neo''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ '''FlaxGPTNeoForCausalLM''', '''FlaxGPTNeoModel''', '''FlaxGPTNeoPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_gpt_neo import GPT_NEO_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoConfig, GPTNeoOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neo import ( GPT_NEO_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoForCausalLM, GPTNeoForQuestionAnswering, GPTNeoForSequenceClassification, GPTNeoForTokenClassification, GPTNeoModel, GPTNeoPreTrainedModel, load_tf_weights_in_gpt_neo, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_gpt_neo import FlaxGPTNeoForCausalLM, FlaxGPTNeoModel, FlaxGPTNeoPreTrainedModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
1
import argparse import os import sys from unittest.mock import patch import pytorch_lightning as pl import timeout_decorator import torch from distillation import SummarizationDistiller, distill_main from finetune import SummarizationModule, main from transformers import MarianMTModel from transformers.file_utils import cached_path from transformers.testing_utils import TestCasePlus, require_torch_gpu, slow from utils import load_json UpperCAmelCase = '''sshleifer/mar_enro_6_3_student''' class A_ ( __lowerCamelCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() lowercase = cached_path( 'https://cdn-datasets.huggingface.co/translation/wmt_en_ro-tr40k-va0.5k-te0.5k.tar.gz' , extract_compressed_file=snake_case , ) lowercase = F'''{data_cached}/wmt_en_ro-tr40k-va0.5k-te0.5k''' @slow @require_torch_gpu def SCREAMING_SNAKE_CASE__ ( self ): MarianMTModel.from_pretrained(snake_case ) @slow @require_torch_gpu def SCREAMING_SNAKE_CASE__ ( self ): lowercase = { '$MAX_LEN': 64, '$BS': 64, '$GAS': 1, '$ENRO_DIR': self.data_dir, 'facebook/mbart-large-cc25': MARIAN_MODEL, # "val_check_interval=0.25": "val_check_interval=1.0", '--learning_rate=3e-5': '--learning_rate 3e-4', '--num_train_epochs 6': '--num_train_epochs 1', } # Clean up bash script lowercase = (self.test_file_dir / 'train_mbart_cc25_enro.sh').open().read().split('finetune.py' )[1].strip() lowercase = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) for k, v in env_vars_to_replace.items(): lowercase = bash_script.replace(snake_case , str(snake_case ) ) lowercase = self.get_auto_remove_tmp_dir() # bash_script = bash_script.replace("--fp16 ", "") lowercase = F''' --output_dir {output_dir} --tokenizer_name Helsinki-NLP/opus-mt-en-ro --sortish_sampler --do_predict --gpus 1 --freeze_encoder --n_train 40000 --n_val 500 --n_test 500 --fp16_opt_level O1 --num_sanity_val_steps 0 --eval_beams 2 '''.split() # XXX: args.gpus > 1 : handle multi_gpu in the future lowercase = ['finetune.py'] + bash_script.split() + args with patch.object(snake_case , 'argv' , snake_case ): lowercase = argparse.ArgumentParser() lowercase = pl.Trainer.add_argparse_args(snake_case ) lowercase = SummarizationModule.add_model_specific_args(snake_case , os.getcwd() ) lowercase = parser.parse_args() lowercase = main(snake_case ) # Check metrics lowercase = load_json(model.metrics_save_path ) lowercase = metrics['val'][0] lowercase = metrics['val'][-1] self.assertEqual(len(metrics['val'] ) , (args.max_epochs / args.val_check_interval) ) assert isinstance(last_step_stats[F'''val_avg_{model.val_metric}'''] , snake_case ) self.assertGreater(last_step_stats['val_avg_gen_time'] , 0.01 ) # model hanging on generate. Maybe bad config was saved. (XXX: old comment/assert?) self.assertLessEqual(last_step_stats['val_avg_gen_time'] , 1.0 ) # test learning requirements: # 1. BLEU improves over the course of training by more than 2 pts self.assertGreater(last_step_stats['val_avg_bleu'] - first_step_stats['val_avg_bleu'] , 2 ) # 2. BLEU finishes above 17 self.assertGreater(last_step_stats['val_avg_bleu'] , 17 ) # 3. test BLEU and val BLEU within ~1.1 pt. self.assertLess(abs(metrics['val'][-1]['val_avg_bleu'] - metrics['test'][-1]['test_avg_bleu'] ) , 1.1 ) # check lightning ckpt can be loaded and has a reasonable statedict lowercase = os.listdir(snake_case ) lowercase = [x for x in contents if x.endswith('.ckpt' )][0] lowercase = os.path.join(args.output_dir , snake_case ) lowercase = torch.load(snake_case , map_location='cpu' ) lowercase = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowercase = {os.path.basename(snake_case ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1 class A_ ( __lowerCamelCase ): '''simple docstring''' @timeout_decorator.timeout(600 ) @slow @require_torch_gpu def SCREAMING_SNAKE_CASE__ ( self ): lowercase = F'''{self.test_file_dir_str}/test_data/wmt_en_ro''' lowercase = { '--fp16_opt_level=O1': '', '$MAX_LEN': 128, '$BS': 16, '$GAS': 1, '$ENRO_DIR': data_dir, '$m': 'sshleifer/student_marian_en_ro_6_1', 'val_check_interval=0.25': 'val_check_interval=1.0', } # Clean up bash script lowercase = ( (self.test_file_dir / 'distil_marian_no_teacher.sh').open().read().split('distillation.py' )[1].strip() ) lowercase = bash_script.replace('\\\n' , '' ).strip().replace('"$@"' , '' ) lowercase = bash_script.replace('--fp16 ' , ' ' ) for k, v in env_vars_to_replace.items(): lowercase = bash_script.replace(snake_case , str(snake_case ) ) lowercase = self.get_auto_remove_tmp_dir() lowercase = bash_script.replace('--fp16' , '' ) lowercase = 6 lowercase = ( ['distillation.py'] + bash_script.split() + [ F'''--output_dir={output_dir}''', '--gpus=1', '--learning_rate=1e-3', F'''--num_train_epochs={epochs}''', '--warmup_steps=10', '--val_check_interval=1.0', '--do_predict', ] ) with patch.object(snake_case , 'argv' , snake_case ): lowercase = argparse.ArgumentParser() lowercase = pl.Trainer.add_argparse_args(snake_case ) lowercase = SummarizationDistiller.add_model_specific_args(snake_case , os.getcwd() ) lowercase = parser.parse_args() # assert args.gpus == gpus THIS BREAKS for multi_gpu lowercase = distill_main(snake_case ) # Check metrics lowercase = load_json(model.metrics_save_path ) lowercase = metrics['val'][0] lowercase = metrics['val'][-1] assert len(metrics['val'] ) >= (args.max_epochs / args.val_check_interval) # +1 accounts for val_sanity_check assert last_step_stats["val_avg_gen_time"] >= 0.01 assert first_step_stats["val_avg_bleu"] < last_step_stats["val_avg_bleu"] # model learned nothing assert 1.0 >= last_step_stats["val_avg_gen_time"] # model hanging on generate. Maybe bad config was saved. assert isinstance(last_step_stats[F'''val_avg_{model.val_metric}'''] , snake_case ) # check lightning ckpt can be loaded and has a reasonable statedict lowercase = os.listdir(snake_case ) lowercase = [x for x in contents if x.endswith('.ckpt' )][0] lowercase = os.path.join(args.output_dir , snake_case ) lowercase = torch.load(snake_case , map_location='cpu' ) lowercase = 'model.model.decoder.layers.0.encoder_attn_layer_norm.weight' assert expected_key in ckpt["state_dict"] assert ckpt["state_dict"]["model.model.decoder.layers.0.encoder_attn_layer_norm.weight"].dtype == torch.floataa # TODO: turn on args.do_predict when PL bug fixed. if args.do_predict: lowercase = {os.path.basename(snake_case ) for p in contents} assert "test_generations.txt" in contents assert "test_results.txt" in contents # assert len(metrics["val"]) == desired_n_evals assert len(metrics['test'] ) == 1
84
from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : jnp.ndarray _UpperCamelCase : jnp.ndarray class A_ ( nn.Module ): '''simple docstring''' _UpperCamelCase : int _UpperCamelCase : Tuple[int] = (16, 32, 96, 256) _UpperCamelCase : jnp.dtype = jnp.floataa def SCREAMING_SNAKE_CASE__ ( self ): lowercase = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) lowercase = [] for i in range(len(self.block_out_channels ) - 1 ): lowercase = self.block_out_channels[i] lowercase = self.block_out_channels[i + 1] lowercase = nn.Conv( snake_case , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(snake_case ) lowercase = nn.Conv( snake_case , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(snake_case ) lowercase = blocks lowercase = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , snake_case ): lowercase = self.conv_in(snake_case ) lowercase = nn.silu(snake_case ) for block in self.blocks: lowercase = block(snake_case ) lowercase = nn.silu(snake_case ) lowercase = self.conv_out(snake_case ) return embedding @flax_register_to_config class A_ ( nn.Module , __lowerCamelCase , __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : int = 32 _UpperCamelCase : int = 4 _UpperCamelCase : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) _UpperCamelCase : Union[bool, Tuple[bool]] = False _UpperCamelCase : Tuple[int] = (320, 640, 1280, 1280) _UpperCamelCase : int = 2 _UpperCamelCase : Union[int, Tuple[int]] = 8 _UpperCamelCase : Optional[Union[int, Tuple[int]]] = None _UpperCamelCase : int = 1280 _UpperCamelCase : float = 0.0 _UpperCamelCase : bool = False _UpperCamelCase : jnp.dtype = jnp.floataa _UpperCamelCase : bool = True _UpperCamelCase : int = 0 _UpperCamelCase : str = "rgb" _UpperCamelCase : Tuple[int] = (16, 32, 96, 256) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): # init input tensors lowercase = (1, self.in_channels, self.sample_size, self.sample_size) lowercase = jnp.zeros(snake_case , dtype=jnp.floataa ) lowercase = jnp.ones((1,) , dtype=jnp.intaa ) lowercase = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) lowercase = (1, 3, self.sample_size * 8, self.sample_size * 8) lowercase = jnp.zeros(snake_case , dtype=jnp.floataa ) lowercase , lowercase = jax.random.split(snake_case ) lowercase = {'params': params_rng, 'dropout': dropout_rng} return self.init(snake_case , snake_case , snake_case , snake_case , snake_case )["params"] def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.block_out_channels lowercase = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. lowercase = self.num_attention_heads or self.attention_head_dim # input lowercase = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time lowercase = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) lowercase = FlaxTimestepEmbedding(snake_case , dtype=self.dtype ) lowercase = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) lowercase = self.only_cross_attention if isinstance(snake_case , snake_case ): lowercase = (only_cross_attention,) * len(self.down_block_types ) if isinstance(snake_case , snake_case ): lowercase = (num_attention_heads,) * len(self.down_block_types ) # down lowercase = [] lowercase = [] lowercase = block_out_channels[0] lowercase = nn.Conv( snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case ) for i, down_block_type in enumerate(self.down_block_types ): lowercase = output_channel lowercase = block_out_channels[i] lowercase = i == len(snake_case ) - 1 if down_block_type == "CrossAttnDownBlock2D": lowercase = FlaxCrossAttnDownBlockaD( in_channels=snake_case , out_channels=snake_case , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: lowercase = FlaxDownBlockaD( in_channels=snake_case , out_channels=snake_case , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(snake_case ) for _ in range(self.layers_per_block ): lowercase = nn.Conv( snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case ) if not is_final_block: lowercase = nn.Conv( snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case ) lowercase = down_blocks lowercase = controlnet_down_blocks # mid lowercase = block_out_channels[-1] lowercase = FlaxUNetMidBlockaDCrossAttn( in_channels=snake_case , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) lowercase = nn.Conv( snake_case , kernel_size=(1, 1) , padding='VALID' , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , snake_case , snake_case , snake_case , snake_case , snake_case = 1.0 , snake_case = True , snake_case = False , ): lowercase = self.controlnet_conditioning_channel_order if channel_order == "bgr": lowercase = jnp.flip(snake_case , axis=1 ) # 1. time if not isinstance(snake_case , jnp.ndarray ): lowercase = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(snake_case , jnp.ndarray ) and len(timesteps.shape ) == 0: lowercase = timesteps.astype(dtype=jnp.floataa ) lowercase = jnp.expand_dims(snake_case , 0 ) lowercase = self.time_proj(snake_case ) lowercase = self.time_embedding(snake_case ) # 2. pre-process lowercase = jnp.transpose(snake_case , (0, 2, 3, 1) ) lowercase = self.conv_in(snake_case ) lowercase = jnp.transpose(snake_case , (0, 2, 3, 1) ) lowercase = self.controlnet_cond_embedding(snake_case ) sample += controlnet_cond # 3. down lowercase = (sample,) for down_block in self.down_blocks: if isinstance(snake_case , snake_case ): lowercase , lowercase = down_block(snake_case , snake_case , snake_case , deterministic=not train ) else: lowercase , lowercase = down_block(snake_case , snake_case , deterministic=not train ) down_block_res_samples += res_samples # 4. mid lowercase = self.mid_block(snake_case , snake_case , snake_case , deterministic=not train ) # 5. contronet blocks lowercase = () for down_block_res_sample, controlnet_block in zip(snake_case , self.controlnet_down_blocks ): lowercase = controlnet_block(snake_case ) controlnet_down_block_res_samples += (down_block_res_sample,) lowercase = controlnet_down_block_res_samples lowercase = self.controlnet_mid_block(snake_case ) # 6. scaling lowercase = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=snake_case , mid_block_res_sample=snake_case )
84
1
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # Checks if the entire collection has been sorted if len(__SCREAMING_SNAKE_CASE ) <= 1 or n <= 1: return insert_next(__SCREAMING_SNAKE_CASE , n - 1 ) rec_insertion_sort(__SCREAMING_SNAKE_CASE , n - 1 ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # Checks order between adjacent elements if index >= len(__SCREAMING_SNAKE_CASE ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order lowercase , lowercase = ( collection[index], collection[index - 1], ) insert_next(__SCREAMING_SNAKE_CASE , index + 1 ) if __name__ == "__main__": UpperCAmelCase = input('''Enter integers separated by spaces: ''') UpperCAmelCase = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
84
import math import os from copy import deepcopy import datasets import evaluate import torch import transformers from datasets import load_dataset from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer from accelerate import Accelerator from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import is_tpu_available, set_seed UpperCAmelCase = '''true''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=82 , __SCREAMING_SNAKE_CASE=16 ): set_seed(42 ) lowercase = RegressionModel() lowercase = deepcopy(__SCREAMING_SNAKE_CASE ) lowercase = RegressionDataset(length=__SCREAMING_SNAKE_CASE ) lowercase = DataLoader(__SCREAMING_SNAKE_CASE , batch_size=__SCREAMING_SNAKE_CASE ) model.to(accelerator.device ) lowercase , lowercase = accelerator.prepare(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return model, ddp_model, dataloader def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=False ): lowercase = AutoTokenizer.from_pretrained('hf-internal-testing/mrpc-bert-base-cased' ) lowercase = load_dataset('glue' , 'mrpc' , split='validation' ) def tokenize_function(__SCREAMING_SNAKE_CASE ): lowercase = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE ) return outputs with accelerator.main_process_first(): lowercase = dataset.map( __SCREAMING_SNAKE_CASE , batched=__SCREAMING_SNAKE_CASE , remove_columns=['idx', 'sentence1', 'sentence2'] , ) lowercase = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__SCREAMING_SNAKE_CASE ): if use_longest: return tokenizer.pad(__SCREAMING_SNAKE_CASE , padding='longest' , return_tensors='pt' ) return tokenizer.pad(__SCREAMING_SNAKE_CASE , padding='max_length' , max_length=128 , return_tensors='pt' ) return DataLoader(__SCREAMING_SNAKE_CASE , shuffle=__SCREAMING_SNAKE_CASE , collate_fn=__SCREAMING_SNAKE_CASE , batch_size=16 ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = Accelerator(dispatch_batches=__SCREAMING_SNAKE_CASE , split_batches=__SCREAMING_SNAKE_CASE ) lowercase = get_dataloader(__SCREAMING_SNAKE_CASE , not dispatch_batches ) lowercase = AutoModelForSequenceClassification.from_pretrained( 'hf-internal-testing/mrpc-bert-base-cased' , return_dict=__SCREAMING_SNAKE_CASE ) lowercase , lowercase = accelerator.prepare(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return {"ddp": [ddp_model, ddp_dataloader, "cuda:0"], "no": [model, dataloader, accelerator.device]}, accelerator def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = [] for batch in dataloader: lowercase , lowercase = batch.values() with torch.no_grad(): lowercase = model(__SCREAMING_SNAKE_CASE ) lowercase , lowercase = accelerator.gather_for_metrics((logit, target) ) logits_and_targets.append((logit, target) ) lowercase , lowercase = [], [] for logit, targ in logits_and_targets: logits.append(__SCREAMING_SNAKE_CASE ) targs.append(__SCREAMING_SNAKE_CASE ) lowercase , lowercase = torch.cat(__SCREAMING_SNAKE_CASE ), torch.cat(__SCREAMING_SNAKE_CASE ) return logits, targs def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=82 , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=False , __SCREAMING_SNAKE_CASE=16 ): lowercase , lowercase , lowercase = get_basic_setup(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase , lowercase = generate_predictions(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) assert ( len(__SCREAMING_SNAKE_CASE ) == num_samples ), F'''Unexpected number of inputs:\n Expected: {num_samples}\n Actual: {len(__SCREAMING_SNAKE_CASE )}''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = False , __SCREAMING_SNAKE_CASE = False ): lowercase = evaluate.load('glue' , 'mrpc' ) lowercase , lowercase = get_mrpc_setup(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # First do baseline lowercase , lowercase , lowercase = setup['no'] model.to(__SCREAMING_SNAKE_CASE ) model.eval() for batch in dataloader: batch.to(__SCREAMING_SNAKE_CASE ) with torch.inference_mode(): lowercase = model(**__SCREAMING_SNAKE_CASE ) lowercase = outputs.logits.argmax(dim=-1 ) metric.add_batch(predictions=__SCREAMING_SNAKE_CASE , references=batch['labels'] ) lowercase = metric.compute() # Then do distributed lowercase , lowercase , lowercase = setup['ddp'] model.eval() for batch in dataloader: with torch.inference_mode(): lowercase = model(**__SCREAMING_SNAKE_CASE ) lowercase = outputs.logits.argmax(dim=-1 ) lowercase = batch['labels'] lowercase , lowercase = accelerator.gather_for_metrics((preds, references) ) metric.add_batch(predictions=__SCREAMING_SNAKE_CASE , references=__SCREAMING_SNAKE_CASE ) lowercase = metric.compute() for key in "accuracy f1".split(): assert math.isclose( baseline[key] , distributed[key] ), F'''Baseline and Distributed are not the same for key {key}:\n\tBaseline: {baseline[key]}\n\tDistributed: {distributed[key]}\n''' def UpperCAmelCase_ ( ): lowercase = Accelerator(split_batches=__SCREAMING_SNAKE_CASE , dispatch_batches=__SCREAMING_SNAKE_CASE ) if accelerator.is_local_main_process: datasets.utils.logging.set_verbosity_warning() transformers.utils.logging.set_verbosity_warning() else: datasets.utils.logging.set_verbosity_error() transformers.utils.logging.set_verbosity_error() # These are a bit slower so they should only be ran on the GPU or TPU if torch.cuda.is_available() or is_tpu_available(): if accelerator.is_local_main_process: print('**Testing gather_for_metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`''' ) test_mrpc(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test torch metrics**' ) for split_batches in [True, False]: for dispatch_batches in [True, False]: lowercase = Accelerator(split_batches=__SCREAMING_SNAKE_CASE , dispatch_batches=__SCREAMING_SNAKE_CASE ) if accelerator.is_local_main_process: print(F'''With: `split_batches={split_batches}`, `dispatch_batches={dispatch_batches}`, length=99''' ) test_torch_metrics(__SCREAMING_SNAKE_CASE , 99 ) accelerator.state._reset_state() if accelerator.is_local_main_process: print('**Test last batch is not dropped when perfectly divisible**' ) lowercase = Accelerator() test_torch_metrics(__SCREAMING_SNAKE_CASE , 512 ) accelerator.state._reset_state() def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
84
1
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar UpperCAmelCase = TypeVar('''T''') class A_ ( Generic[T] ): '''simple docstring''' def __init__( self , snake_case ): lowercase = data lowercase = None def __str__( self ): return F'''{self.data}''' class A_ ( Generic[T] ): '''simple docstring''' def __init__( self ): lowercase = None def __iter__( self ): lowercase = self.top while node: yield node.data lowercase = node.next def __str__( self ): return "->".join([str(snake_case ) for item in self] ) def __len__( self ): return len(tuple(iter(self ) ) ) def SCREAMING_SNAKE_CASE__ ( self ): return self.top is None def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = Node(snake_case ) if not self.is_empty(): lowercase = self.top lowercase = node def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top , snake_case ) lowercase = self.top lowercase = self.top.next return pop_node.data def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def SCREAMING_SNAKE_CASE__ ( self ): lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
84
import warnings from typing import List import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import is_flax_available, is_tf_available, is_torch_available class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[str] = ["""image_processor""", """tokenizer"""] _UpperCamelCase : Any = """OwlViTImageProcessor""" _UpperCamelCase : Dict = ("""CLIPTokenizer""", """CLIPTokenizerFast""") def __init__( self , snake_case=None , snake_case=None , **snake_case ): lowercase = None if "feature_extractor" in kwargs: warnings.warn( 'The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`' ' instead.' , snake_case , ) lowercase = kwargs.pop('feature_extractor' ) lowercase = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError('You need to specify an `image_processor`.' ) if tokenizer is None: raise ValueError('You need to specify a `tokenizer`.' ) super().__init__(snake_case , snake_case ) def __call__( self , snake_case=None , snake_case=None , snake_case=None , snake_case="max_length" , snake_case="np" , **snake_case ): if text is None and query_images is None and images is None: raise ValueError( 'You have to specify at least one text or query image or image. All three cannot be none.' ) if text is not None: if isinstance(snake_case , snake_case ) or (isinstance(snake_case , snake_case ) and not isinstance(text[0] , snake_case )): lowercase = [self.tokenizer(snake_case , padding=snake_case , return_tensors=snake_case , **snake_case )] elif isinstance(snake_case , snake_case ) and isinstance(text[0] , snake_case ): lowercase = [] # Maximum number of queries across batch lowercase = max([len(snake_case ) for t in text] ) # Pad all batch samples to max number of text queries for t in text: if len(snake_case ) != max_num_queries: lowercase = t + [' '] * (max_num_queries - len(snake_case )) lowercase = self.tokenizer(snake_case , padding=snake_case , return_tensors=snake_case , **snake_case ) encodings.append(snake_case ) else: raise TypeError('Input text should be a string, a list of strings or a nested list of strings' ) if return_tensors == "np": lowercase = np.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = np.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "jax" and is_flax_available(): import jax.numpy as jnp lowercase = jnp.concatenate([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = jnp.concatenate([encoding['attention_mask'] for encoding in encodings] , axis=0 ) elif return_tensors == "pt" and is_torch_available(): import torch lowercase = torch.cat([encoding['input_ids'] for encoding in encodings] , dim=0 ) lowercase = torch.cat([encoding['attention_mask'] for encoding in encodings] , dim=0 ) elif return_tensors == "tf" and is_tf_available(): import tensorflow as tf lowercase = tf.stack([encoding['input_ids'] for encoding in encodings] , axis=0 ) lowercase = tf.stack([encoding['attention_mask'] for encoding in encodings] , axis=0 ) else: raise ValueError('Target return tensor type could not be returned' ) lowercase = BatchEncoding() lowercase = input_ids lowercase = attention_mask if query_images is not None: lowercase = BatchEncoding() lowercase = self.image_processor( snake_case , return_tensors=snake_case , **snake_case ).pixel_values lowercase = query_pixel_values if images is not None: lowercase = self.image_processor(snake_case , return_tensors=snake_case , **snake_case ) if text is not None and images is not None: lowercase = image_features.pixel_values return encoding elif query_images is not None and images is not None: lowercase = image_features.pixel_values return encoding elif text is not None or query_images is not None: return encoding else: return BatchEncoding(data=dict(**snake_case ) , tensor_type=snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process_object_detection(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.image_processor.post_process_image_guided_detection(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.batch_decode(*snake_case , **snake_case ) def SCREAMING_SNAKE_CASE__ ( self , *snake_case , **snake_case ): return self.tokenizer.decode(*snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self ): warnings.warn( '`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead.' , snake_case , ) return self.image_processor_class @property def SCREAMING_SNAKE_CASE__ ( self ): warnings.warn( '`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead.' , snake_case , ) return self.image_processor
84
1
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) return n == n[::-1] def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 100_0000 ): lowercase = 0 for i in range(1 , __SCREAMING_SNAKE_CASE ): if is_palindrome(__SCREAMING_SNAKE_CASE ) and is_palindrome(bin(__SCREAMING_SNAKE_CASE ).split('b' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
84
from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_config_file''': '''tokenizer_config.json''', } UpperCAmelCase = { '''vocab_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json''' }, '''merges_file''': { '''facebook/blenderbot_small-90M''': '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt''' }, '''tokenizer_config_file''': { '''facebook/blenderbot_small-90M''': ( '''https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json''' ) }, } UpperCAmelCase = { '''facebook/blenderbot_small-90M''': 512, } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : Dict = VOCAB_FILES_NAMES _UpperCamelCase : Any = PRETRAINED_VOCAB_FILES_MAP _UpperCamelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _UpperCamelCase : str = BlenderbotSmallTokenizer def __init__( self , snake_case=None , snake_case=None , snake_case="<|endoftext|>" , snake_case="<|endoftext|>" , snake_case="<|endoftext|>" , snake_case=False , snake_case=True , **snake_case , ): super().__init__( ByteLevelBPETokenizer( vocab=snake_case , merges=snake_case , add_prefix_space=snake_case , trim_offsets=snake_case , ) , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , **snake_case , ) lowercase = add_prefix_space def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=None ): lowercase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = [self.sep_token_id] lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
84
1
from __future__ import annotations import unittest import numpy as np from transformers import LayoutLMConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers.models.layoutlm.modeling_tf_layoutlm import ( TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST, TFLayoutLMForMaskedLM, TFLayoutLMForQuestionAnswering, TFLayoutLMForSequenceClassification, TFLayoutLMForTokenClassification, TFLayoutLMModel, ) class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=2 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , snake_case=1000 , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = range_bbox def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # convert bbox to numpy since TF does not support item assignment lowercase = ids_tensor([self.batch_size, self.seq_length, 4] , self.range_bbox ).numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: lowercase = bbox[i, j, 3] lowercase = bbox[i, j, 1] lowercase = t if bbox[i, j, 2] < bbox[i, j, 0]: lowercase = bbox[i, j, 2] lowercase = bbox[i, j, 0] lowercase = t lowercase = tf.convert_to_tensor(snake_case ) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length] ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = LayoutLMConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , ) return config, input_ids, bbox, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFLayoutLMModel(config=snake_case ) lowercase = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) lowercase = model(snake_case , snake_case , token_type_ids=snake_case ) lowercase = model(snake_case , snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFLayoutLMForMaskedLM(config=snake_case ) lowercase = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = TFLayoutLMForSequenceClassification(config=snake_case ) lowercase = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = TFLayoutLMForTokenClassification(config=snake_case ) lowercase = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = TFLayoutLMForQuestionAnswering(config=snake_case ) lowercase = model(snake_case , snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = { 'input_ids': input_ids, 'bbox': bbox, 'token_type_ids': token_type_ids, 'attention_mask': input_mask, } return config, inputs_dict @require_tf class A_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = ( ( TFLayoutLMModel, TFLayoutLMForMaskedLM, TFLayoutLMForTokenClassification, TFLayoutLMForSequenceClassification, TFLayoutLMForQuestionAnswering, ) if is_tf_available() else () ) _UpperCamelCase : Tuple = ( { """feature-extraction""": TFLayoutLMModel, """fill-mask""": TFLayoutLMForMaskedLM, """text-classification""": TFLayoutLMForSequenceClassification, """token-classification""": TFLayoutLMForTokenClassification, """zero-shot""": TFLayoutLMForSequenceClassification, } if is_tf_available() else {} ) _UpperCamelCase : List[str] = False _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : Optional[int] = 10 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = TFLayoutLMModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in TF_LAYOUTLM_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = TFLayoutLMModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip('Onnx compliancy broke with TF 2.10' ) def SCREAMING_SNAKE_CASE__ ( self ): pass def UpperCAmelCase_ ( ): # Here we prepare a batch of 2 sequences to test a LayoutLM forward pass on: # fmt: off lowercase = tf.convert_to_tensor([[101,1019,1014,1016,1037,1_2849,4747,1004,1_4246,2278,5439,4524,5002,2930,2193,2930,4341,3208,1005,1055,2171,2848,1_1300,3531,102],[101,4070,4034,7020,1024,3058,1015,1013,2861,1013,6070,1_9274,2772,6205,2_7814,1_6147,1_6147,4343,2047,1_0283,1_0969,1_4389,1012,2338,102]] ) # noqa: E231 lowercase = tf.convert_to_tensor([[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1],] ) # noqa: E231 lowercase = tf.convert_to_tensor([[[0,0,0,0],[423,237,440,251],[427,272,441,287],[419,115,437,129],[961,885,992,912],[256,38,330,58],[256,38,330,58],[336,42,353,57],[360,39,401,56],[360,39,401,56],[411,39,471,59],[479,41,528,59],[533,39,630,60],[67,113,134,131],[141,115,209,132],[68,149,133,166],[141,149,187,164],[195,148,287,165],[195,148,287,165],[195,148,287,165],[295,148,349,165],[441,149,492,166],[497,149,546,164],[64,201,125,218],[1000,1000,1000,1000]],[[0,0,0,0],[662,150,754,166],[665,199,742,211],[519,213,554,228],[519,213,554,228],[134,433,187,454],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[130,467,204,480],[314,469,376,482],[504,684,582,706],[941,825,973,900],[941,825,973,900],[941,825,973,900],[941,825,973,900],[610,749,652,765],[130,659,168,672],[176,657,237,672],[238,657,312,672],[443,653,628,672],[443,653,628,672],[716,301,825,317],[1000,1000,1000,1000]]] ) # noqa: E231 lowercase = tf.convert_to_tensor([[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0]] ) # noqa: E231 # these are sequence labels (i.e. at the token level) lowercase = tf.convert_to_tensor([[-100,10,10,10,9,1,-100,7,7,-100,7,7,4,2,5,2,8,8,-100,-100,5,0,3,2,-100],[-100,12,12,12,-100,12,10,-100,-100,-100,-100,10,12,9,-100,-100,-100,10,10,10,9,12,-100,10,-100]] ) # noqa: E231 # fmt: on return input_ids, attention_mask, bbox, token_type_ids, labels @require_tf class A_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = TFLayoutLMModel.from_pretrained('microsoft/layoutlm-base-uncased' ) lowercase , lowercase , lowercase , lowercase , lowercase = prepare_layoutlm_batch_inputs() # forward pass lowercase = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the sequence output on [0, :3, :3] lowercase = tf.convert_to_tensor( [[0.1_785, -0.1_947, -0.0_425], [-0.3_254, -0.2_807, 0.2_553], [-0.5_391, -0.3_322, 0.3_364]] , ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , snake_case , atol=1E-3 ) ) # test the pooled output on [1, :3] lowercase = tf.convert_to_tensor([-0.6_580, -0.0_214, 0.8_552] ) self.assertTrue(np.allclose(outputs.pooler_output[1, :3] , snake_case , atol=1E-3 ) ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized sequence classification head lowercase = TFLayoutLMForSequenceClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=2 ) lowercase , lowercase , lowercase , lowercase , lowercase = prepare_layoutlm_batch_inputs() # forward pass lowercase = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=tf.convert_to_tensor([1, 1] ) , ) # test whether we get a loss as a scalar lowercase = outputs.loss lowercase = (2,) self.assertEqual(loss.shape , snake_case ) # test the shape of the logits lowercase = outputs.logits lowercase = (2, 2) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized token classification head lowercase = TFLayoutLMForTokenClassification.from_pretrained('microsoft/layoutlm-base-uncased' , num_labels=13 ) lowercase , lowercase , lowercase , lowercase , lowercase = prepare_layoutlm_batch_inputs() # forward pass lowercase = model( input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) # test the shape of the logits lowercase = outputs.logits lowercase = tf.convert_to_tensor((2, 25, 13) ) self.assertEqual(logits.shape , snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): # initialize model with randomly initialized token classification head lowercase = TFLayoutLMForQuestionAnswering.from_pretrained('microsoft/layoutlm-base-uncased' ) lowercase , lowercase , lowercase , lowercase , lowercase = prepare_layoutlm_batch_inputs() # forward pass lowercase = model(input_ids=snake_case , bbox=snake_case , attention_mask=snake_case , token_type_ids=snake_case ) # test the shape of the logits lowercase = tf.convert_to_tensor((2, 25) ) self.assertEqual(outputs.start_logits.shape , snake_case ) self.assertEqual(outputs.end_logits.shape , snake_case )
84
import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST, OpenAIGPTConfig, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification, OpenAIGPTLMHeadModel, OpenAIGPTModel, ) class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope lowercase = self.vocab_size - 1 def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = OpenAIGPTConfig( vocab_size=self.vocab_size , n_embd=self.hidden_size , n_layer=self.num_hidden_layers , n_head=self.num_attention_heads , n_positions=self.max_position_embeddings , pad_token_id=self.pad_token_id , ) lowercase = ids_tensor([self.num_hidden_layers, self.num_attention_heads] , 2 ) return ( config, input_ids, head_mask, token_type_ids, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , head_mask=snake_case ) lowercase = model(snake_case , token_type_ids=snake_case ) lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTLMHeadModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = OpenAIGPTDoubleHeadsModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , *snake_case ): lowercase = self.num_labels lowercase = OpenAIGPTForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = model(snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = { 'input_ids': input_ids, 'token_type_ids': token_type_ids, 'head_mask': head_mask, } return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Optional[Any] = ( (OpenAIGPTModel, OpenAIGPTLMHeadModel, OpenAIGPTDoubleHeadsModel, OpenAIGPTForSequenceClassification) if is_torch_available() else () ) _UpperCamelCase : Tuple = ( (OpenAIGPTLMHeadModel,) if is_torch_available() else () ) # TODO (PVP): Add Double HeadsModel when generate() function is changed accordingly _UpperCamelCase : str = ( { """feature-extraction""": OpenAIGPTModel, """text-classification""": OpenAIGPTForSequenceClassification, """text-generation""": OpenAIGPTLMHeadModel, """zero-shot""": OpenAIGPTForSequenceClassification, } if is_torch_available() else {} ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case ): if pipeline_test_casse_name == "ZeroShotClassificationPipelineTests": # Get `tokenizer does not have a padding token` error for both fast/slow tokenizers. # `OpenAIGPTConfig` was never used in pipeline tests, either because of a missing checkpoint or because a # tiny config could not be created. return True return False def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case=False ): lowercase = super()._prepare_for_class(snake_case , snake_case , return_labels=snake_case ) if return_labels: if model_class.__name__ == "OpenAIGPTDoubleHeadsModel": lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices, self.model_tester.seq_length) , dtype=torch.long , device=snake_case , ) lowercase = inputs_dict['labels'] lowercase = inputs_dict['labels'] lowercase = torch.zeros( (self.model_tester.batch_size, self.model_tester.num_choices) , dtype=torch.long , device=snake_case , ) lowercase = torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=snake_case ) return inputs_dict def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , n_embd=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_double_lm_head_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_openai_gpt_for_sequence_classification(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in OPENAI_GPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = OpenAIGPTModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTLMHeadModel.from_pretrained('openai-gpt' ) model.to(snake_case ) lowercase = torch.tensor([[481, 4735, 544]] , dtype=torch.long , device=snake_case ) # the president is lowercase = [ 481, 4735, 544, 246, 963, 870, 762, 239, 244, 4_0477, 244, 249, 719, 881, 487, 544, 240, 244, 603, 481, ] # the president is a very good man. " \n " i\'m sure he is, " said the lowercase = model.generate(snake_case , do_sample=snake_case ) self.assertListEqual(output_ids[0].tolist() , snake_case )
84
1
from __future__ import annotations UpperCAmelCase = 8.988e9 # units = N * m^s * C^-2 def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = abs(chargea * chargea ) if (force, chargea, chargea, distance).count(0 ) != 1: raise ValueError('One and only one argument must be 0' ) if distance < 0: raise ValueError('Distance cannot be negative' ) if force == 0: lowercase = COULOMBS_CONSTANT * charge_product / (distance**2) return {"force": force} elif chargea == 0: lowercase = abs(__SCREAMING_SNAKE_CASE ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge1": chargea} elif chargea == 0: lowercase = abs(__SCREAMING_SNAKE_CASE ) * (distance**2) / (COULOMBS_CONSTANT * chargea) return {"charge2": chargea} elif distance == 0: lowercase = (COULOMBS_CONSTANT * charge_product / abs(__SCREAMING_SNAKE_CASE )) ** 0.5 return {"distance": distance} raise ValueError('Exactly one argument must be 0' ) if __name__ == "__main__": import doctest doctest.testmod()
84
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCAmelCase = {'''configuration_vit_msn''': ['''VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMSNConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = [ '''VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMSNModel''', '''ViTMSNForImageClassification''', '''ViTMSNPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_msn import VIT_MSN_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMSNConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_msn import ( VIT_MSN_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMSNForImageClassification, ViTMSNModel, ViTMSNPreTrainedModel, ) else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
1
import unittest from transformers import DebertaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DebertaForMaskedLM, DebertaForQuestionAnswering, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaModel, ) from transformers.models.deberta.modeling_deberta import DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=True , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=False , snake_case=True , snake_case="None" , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = relative_attention lowercase = position_biased_input lowercase = pos_att_type lowercase = scope def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_input_mask: lowercase = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self ): return DebertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , pos_att_type=self.pos_att_type , ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.get_config() lowercase = 300 return config def SCREAMING_SNAKE_CASE__ ( self , snake_case ): self.parent.assertListEqual(list(result.loss.size() ) , [] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = DebertaModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case )[0] lowercase = model(snake_case , token_type_ids=snake_case )[0] lowercase = model(snake_case )[0] self.parent.assertListEqual(list(sequence_output.size() ) , [self.batch_size, self.seq_length, self.hidden_size] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = DebertaForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = DebertaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertListEqual(list(result.logits.size() ) , [self.batch_size, self.num_labels] ) self.check_loss_output(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = self.num_labels lowercase = DebertaForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = DebertaForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : List[Any] = ( ( DebertaModel, DebertaForMaskedLM, DebertaForSequenceClassification, DebertaForTokenClassification, DebertaForQuestionAnswering, ) if is_torch_available() else () ) _UpperCamelCase : int = ( { """feature-extraction""": DebertaModel, """fill-mask""": DebertaForMaskedLM, """question-answering""": DebertaForQuestionAnswering, """text-classification""": DebertaForSequenceClassification, """token-classification""": DebertaForTokenClassification, """zero-shot""": DebertaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase : Union[str, Any] = True _UpperCamelCase : Any = False _UpperCamelCase : Optional[Any] = False _UpperCamelCase : Union[str, Any] = False _UpperCamelCase : Optional[Any] = False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = DebertaModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_sequence_classification(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_masked_lm(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_question_answering(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_deberta_for_token_classification(*snake_case ) @slow def SCREAMING_SNAKE_CASE__ ( self ): for model_name in DEBERTA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase = DebertaModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @require_torch @require_sentencepiece @require_tokenizers class A_ ( unittest.TestCase ): '''simple docstring''' @unittest.skip(reason='Model not available yet' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = DebertaModel.from_pretrained('microsoft/deberta-base' ) lowercase = torch.tensor([[0, 3_1414, 232, 328, 740, 1140, 1_2695, 69, 4_6078, 1588, 2]] ) lowercase = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): lowercase = model(snake_case , attention_mask=snake_case )[0] # compare the actual values for a slice. lowercase = torch.tensor( [[[-0.5_986, -0.8_055, -0.8_462], [1.4_484, -0.9_348, -0.8_059], [0.3_123, 0.0_032, -1.4_131]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , snake_case , atol=1E-4 ) , F'''{output[:, 1:4, 1:4]}''' )
84
import math def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = [True] * n lowercase = False lowercase = False lowercase = True for i in range(3 , int(n**0.5 + 1 ) , 2 ): lowercase = i * 2 while index < n: lowercase = False lowercase = index + i lowercase = [2] for i in range(3 , __SCREAMING_SNAKE_CASE , 2 ): if is_prime[i]: primes.append(__SCREAMING_SNAKE_CASE ) return primes def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 9999_6666_3333 ): lowercase = math.floor(math.sqrt(__SCREAMING_SNAKE_CASE ) ) + 100 lowercase = prime_sieve(__SCREAMING_SNAKE_CASE ) lowercase = 0 lowercase = 0 lowercase = primes[prime_index] while (last_prime**2) <= limit: lowercase = primes[prime_index + 1] lowercase = last_prime**2 lowercase = next_prime**2 # Get numbers divisible by lps(current) lowercase = lower_bound + last_prime while upper_bound > current <= limit: matches_sum += current current += last_prime # Reset the upper_bound while (upper_bound - next_prime) > limit: upper_bound -= next_prime # Add the numbers divisible by ups(current) lowercase = upper_bound - next_prime while current > lower_bound: matches_sum += current current -= next_prime # Remove the numbers divisible by both ups and lps lowercase = 0 while upper_bound > current <= limit: if current <= lower_bound: # Increment the current number current += last_prime * next_prime continue if current > limit: break # Remove twice since it was added by both ups and lps matches_sum -= current * 2 # Increment the current number current += last_prime * next_prime # Setup for next pair lowercase = next_prime prime_index += 1 return matches_sum if __name__ == "__main__": print(solution())
84
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..utils import cached_file # docstyle-ignore UpperCAmelCase = ''' Human: <<task>> Assistant: ''' UpperCAmelCase = '''huggingface-tools/default-prompts''' UpperCAmelCase = {'''chat''': '''chat_prompt_template.txt''', '''run''': '''run_prompt_template.txt'''} def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE="run" ): if prompt_or_repo_id is None: lowercase = DEFAULT_PROMPTS_REPO # prompt is considered a repo ID when it does not contain any kind of space if re.search('\\s' , __SCREAMING_SNAKE_CASE ) is not None: return prompt_or_repo_id lowercase = cached_file( __SCREAMING_SNAKE_CASE , PROMPT_FILES[mode] , repo_type='dataset' , user_agent={'agent': agent_name} ) with open(__SCREAMING_SNAKE_CASE , 'r' , encoding='utf-8' ) as f: return f.read()
84
import collections import os import re from pathlib import Path UpperCAmelCase = '''src/transformers''' # Matches is_xxx_available() UpperCAmelCase = re.compile(R'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} UpperCAmelCase = re.compile(R'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] UpperCAmelCase = re.compile(R'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available UpperCAmelCase = re.compile(R'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") UpperCAmelCase = re.compile(R'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] UpperCAmelCase = re.compile(R'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", UpperCAmelCase = re.compile(R'''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], UpperCAmelCase = re.compile(R'''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo UpperCAmelCase = re.compile(R'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: UpperCAmelCase = re.compile(R'''^\s*try:''') # Catches a line with else: UpperCAmelCase = re.compile(R'''^\s*else:''') def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): if _re_test_backend.search(__SCREAMING_SNAKE_CASE ) is None: return None lowercase = [b[0] for b in _re_backend.findall(__SCREAMING_SNAKE_CASE )] backends.sort() return "_and_".join(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): with open(__SCREAMING_SNAKE_CASE , 'r' , encoding='utf-8' , newline='\n' ) as f: lowercase = f.readlines() lowercase = 0 while line_index < len(__SCREAMING_SNAKE_CASE ) and not lines[line_index].startswith('_import_structure = {' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(__SCREAMING_SNAKE_CASE ): return None # First grab the objects without a specific backend in _import_structure lowercase = [] while not lines[line_index].startswith('if TYPE_CHECKING' ) and find_backend(lines[line_index] ) is None: lowercase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(__SCREAMING_SNAKE_CASE ): lowercase = _re_one_line_import_struct.search(__SCREAMING_SNAKE_CASE ).groups()[0] lowercase = re.findall(r'\[([^\]]+)\]' , __SCREAMING_SNAKE_CASE ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(', ' )] ) line_index += 1 continue lowercase = _re_import_struct_key_value.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: lowercase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(', ' ) if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) line_index += 1 lowercase = {'none': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('if TYPE_CHECKING' ): # If the line is an if not is_backend_available, we grab all objects associated. lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 4 ): lowercase = lines[line_index] if _re_import_struct_add_one.search(__SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_import_struct_add_one.search(__SCREAMING_SNAKE_CASE ).groups()[0] ) elif _re_import_struct_add_many.search(__SCREAMING_SNAKE_CASE ) is not None: lowercase = _re_import_struct_add_many.search(__SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) lowercase = [obj[1:-1] for obj in imports if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif _re_between_brackets.search(__SCREAMING_SNAKE_CASE ) is not None: lowercase = _re_between_brackets.search(__SCREAMING_SNAKE_CASE ).groups()[0].split(', ' ) lowercase = [obj[1:-1] for obj in imports if len(__SCREAMING_SNAKE_CASE ) > 0] objects.extend(__SCREAMING_SNAKE_CASE ) elif _re_quote_object.search(__SCREAMING_SNAKE_CASE ) is not None: objects.append(_re_quote_object.search(__SCREAMING_SNAKE_CASE ).groups()[0] ) elif line.startswith(' ' * 8 + '"' ): objects.append(line[9:-3] ) elif line.startswith(' ' * 12 + '"' ): objects.append(line[13:-3] ) line_index += 1 lowercase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend lowercase = [] while ( line_index < len(__SCREAMING_SNAKE_CASE ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('else' ) ): lowercase = lines[line_index] lowercase = _re_import.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 8 ): objects.append(line[8:-2] ) line_index += 1 lowercase = {'none': objects} # Let's continue with backend-specific objects while line_index < len(__SCREAMING_SNAKE_CASE ): # If the line is an if is_backend_available, we grab all objects associated. lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(' ' * 8 ): lowercase = lines[line_index] lowercase = _re_import.search(__SCREAMING_SNAKE_CASE ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(', ' ) ) elif line.startswith(' ' * 12 ): objects.append(line[12:-2] ) line_index += 1 lowercase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): def find_duplicates(__SCREAMING_SNAKE_CASE ): return [k for k, v in collections.Counter(__SCREAMING_SNAKE_CASE ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] lowercase = [] for key in import_dict_objects.keys(): lowercase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F'''Duplicate _import_structure definitions for: {duplicate_imports}''' ) lowercase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F'''Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}''' ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): lowercase = 'base imports' if key == 'none' else F'''{key} backend''' errors.append(F'''Differences for {name}:''' ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F''' {a} in TYPE_HINT but not in _import_structure.''' ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F''' {a} in _import_structure but not in TYPE_HINT.''' ) return errors def UpperCAmelCase_ ( ): lowercase = [] for root, _, files in os.walk(__SCREAMING_SNAKE_CASE ): if "__init__.py" in files: lowercase = os.path.join(__SCREAMING_SNAKE_CASE , '__init__.py' ) lowercase = parse_init(__SCREAMING_SNAKE_CASE ) if objects is not None: lowercase = analyze_results(*__SCREAMING_SNAKE_CASE ) if len(__SCREAMING_SNAKE_CASE ) > 0: lowercase = F'''Problem in {fname}, both halves do not define the same objects.\n{errors[0]}''' failures.append('\n'.join(__SCREAMING_SNAKE_CASE ) ) if len(__SCREAMING_SNAKE_CASE ) > 0: raise ValueError('\n\n'.join(__SCREAMING_SNAKE_CASE ) ) def UpperCAmelCase_ ( ): lowercase = [] for path, directories, files in os.walk(__SCREAMING_SNAKE_CASE ): for folder in directories: # Ignore private modules if folder.startswith('_' ): directories.remove(__SCREAMING_SNAKE_CASE ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(__SCREAMING_SNAKE_CASE ) / folder).glob('*.py' ) ) ) == 0: continue lowercase = str((Path(__SCREAMING_SNAKE_CASE ) / folder).relative_to(__SCREAMING_SNAKE_CASE ) ) lowercase = short_path.replace(os.path.sep , '.' ) submodules.append(__SCREAMING_SNAKE_CASE ) for fname in files: if fname == "__init__.py": continue lowercase = str((Path(__SCREAMING_SNAKE_CASE ) / fname).relative_to(__SCREAMING_SNAKE_CASE ) ) lowercase = short_path.replace('.py' , '' ).replace(os.path.sep , '.' ) if len(submodule.split('.' ) ) == 1: submodules.append(__SCREAMING_SNAKE_CASE ) return submodules UpperCAmelCase = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', '''models.esm.openfold_utils''', ] def UpperCAmelCase_ ( ): # This is to make sure the transformers module imported is the one in the repo. from transformers.utils import direct_transformers_import lowercase = direct_transformers_import(__SCREAMING_SNAKE_CASE ) lowercase = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(__SCREAMING_SNAKE_CASE , '__init__.py' ) , 'r' ) as f: lowercase = f.read() import_structure_keys.update(set(re.findall(r'import_structure\[\"([^\"]*)\"\]' , __SCREAMING_SNAKE_CASE ) ) ) lowercase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(__SCREAMING_SNAKE_CASE ) > 0: lowercase = '\n'.join(F'''- {module}''' for module in module_not_registered ) raise ValueError( 'The following submodules are not properly registed in the main init of Transformers:\n' F'''{list_of_modules}\n''' 'Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.' ) if __name__ == "__main__": check_all_inits() check_submodules()
84
1
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): for i in range(len(__SCREAMING_SNAKE_CASE ) - 1 , 0 , -1 ): lowercase = False for j in range(__SCREAMING_SNAKE_CASE , 0 , -1 ): if unsorted[j] < unsorted[j - 1]: lowercase , lowercase = unsorted[j - 1], unsorted[j] lowercase = True for j in range(__SCREAMING_SNAKE_CASE ): if unsorted[j] > unsorted[j + 1]: lowercase , lowercase = unsorted[j + 1], unsorted[j] lowercase = True if not swapped: break return unsorted if __name__ == "__main__": import doctest doctest.testmod() UpperCAmelCase = input('''Enter numbers separated by a comma:\n''').strip() UpperCAmelCase = [int(item) for item in user_input.split(''',''')] print(F"""{cocktail_shaker_sort(unsorted) = }""")
84
from __future__ import annotations from collections.abc import Iterator from typing import Generic, TypeVar UpperCAmelCase = TypeVar('''T''') class A_ ( Generic[T] ): '''simple docstring''' def __init__( self , snake_case ): lowercase = data lowercase = None def __str__( self ): return F'''{self.data}''' class A_ ( Generic[T] ): '''simple docstring''' def __init__( self ): lowercase = None def __iter__( self ): lowercase = self.top while node: yield node.data lowercase = node.next def __str__( self ): return "->".join([str(snake_case ) for item in self] ) def __len__( self ): return len(tuple(iter(self ) ) ) def SCREAMING_SNAKE_CASE__ ( self ): return self.top is None def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = Node(snake_case ) if not self.is_empty(): lowercase = self.top lowercase = node def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('pop from empty stack' ) assert isinstance(self.top , snake_case ) lowercase = self.top lowercase = self.top.next return pop_node.data def SCREAMING_SNAKE_CASE__ ( self ): if self.is_empty(): raise IndexError('peek from empty stack' ) assert self.top is not None return self.top.data def SCREAMING_SNAKE_CASE__ ( self ): lowercase = None if __name__ == "__main__": from doctest import testmod testmod()
84
1
from sklearn.metrics import fa_score import datasets UpperCAmelCase = ''' The F1 score is the harmonic mean of the precision and recall. It can be computed with the equation: F1 = 2 * (precision * recall) / (precision + recall) ''' UpperCAmelCase = ''' Args: predictions (`list` of `int`): Predicted labels. references (`list` of `int`): Ground truth labels. labels (`list` of `int`): The set of labels to include when `average` is not set to `\'binary\'`, and the order of the labels if `average` is `None`. Labels present in the data can be excluded, for example to calculate a multiclass average ignoring a majority negative class. Labels not present in the data will result in 0 components in a macro average. For multilabel targets, labels are column indices. By default, all labels in `predictions` and `references` are used in sorted order. Defaults to None. pos_label (`int`): The class to be considered the positive class, in the case where `average` is set to `binary`. Defaults to 1. average (`string`): This parameter is required for multiclass/multilabel targets. If set to `None`, the scores for each class are returned. Otherwise, this determines the type of averaging performed on the data. Defaults to `\'binary\'`. - \'binary\': Only report results for the class specified by `pos_label`. This is applicable only if the classes found in `predictions` and `references` are binary. - \'micro\': Calculate metrics globally by counting the total true positives, false negatives and false positives. - \'macro\': Calculate metrics for each label, and find their unweighted mean. This does not take label imbalance into account. - \'weighted\': Calculate metrics for each label, and find their average weighted by support (the number of true instances for each label). This alters `\'macro\'` to account for label imbalance. This option can result in an F-score that is not between precision and recall. - \'samples\': Calculate metrics for each instance, and find their average (only meaningful for multilabel classification). sample_weight (`list` of `float`): Sample weights Defaults to None. Returns: f1 (`float` or `array` of `float`): F1 score or list of f1 scores, depending on the value passed to `average`. Minimum possible value is 0. Maximum possible value is 1. Higher f1 scores are better. Examples: Example 1-A simple binary example >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0]) >>> print(results) {\'f1\': 0.5} Example 2-The same simple binary example as in Example 1, but with `pos_label` set to `0`. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], pos_label=0) >>> print(round(results[\'f1\'], 2)) 0.67 Example 3-The same simple binary example as in Example 1, but with `sample_weight` included. >>> f1_metric = datasets.load_metric("f1") >>> results = f1_metric.compute(references=[0, 1, 0, 1, 0], predictions=[0, 0, 1, 1, 0], sample_weight=[0.9, 0.5, 3.9, 1.2, 0.3]) >>> print(round(results[\'f1\'], 2)) 0.35 Example 4-A multiclass example, with different values for the `average` input. >>> predictions = [0, 2, 1, 0, 0, 1] >>> references = [0, 1, 2, 0, 1, 2] >>> results = f1_metric.compute(predictions=predictions, references=references, average="macro") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average="micro") >>> print(round(results[\'f1\'], 2)) 0.33 >>> results = f1_metric.compute(predictions=predictions, references=references, average="weighted") >>> print(round(results[\'f1\'], 2)) 0.27 >>> results = f1_metric.compute(predictions=predictions, references=references, average=None) >>> print(results) {\'f1\': array([0.8, 0. , 0. ])} ''' UpperCAmelCase = ''' @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class A_ ( datasets.Metric ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Sequence(datasets.Value('int32' ) ), 'references': datasets.Sequence(datasets.Value('int32' ) ), } if self.config_name == 'multilabel' else { 'predictions': datasets.Value('int32' ), 'references': datasets.Value('int32' ), } ) , reference_urls=['https://scikit-learn.org/stable/modules/generated/sklearn.metrics.f1_score.html'] , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case=None , snake_case=1 , snake_case="binary" , snake_case=None ): lowercase = fa_score( snake_case , snake_case , labels=snake_case , pos_label=snake_case , average=snake_case , sample_weight=snake_case ) return {"f1": float(snake_case ) if score.size == 1 else score}
84
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case=13 , snake_case=7 , snake_case=True , snake_case=True , snake_case=False , snake_case=True , snake_case=99 , snake_case=32 , snake_case=5 , snake_case=4 , snake_case=37 , snake_case="gelu" , snake_case=0.1 , snake_case=0.1 , snake_case=512 , snake_case=16 , snake_case=2 , snake_case=0.02 , snake_case=3 , snake_case=4 , snake_case=None , ): lowercase = parent lowercase = batch_size lowercase = seq_length lowercase = is_training lowercase = use_input_mask lowercase = use_token_type_ids lowercase = use_labels lowercase = vocab_size lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = max_position_embeddings lowercase = type_vocab_size lowercase = type_sequence_label_size lowercase = initializer_range lowercase = num_labels lowercase = num_choices lowercase = scope def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase = None if self.use_input_mask: lowercase = random_attention_mask([self.batch_size, self.seq_length] ) lowercase = None if self.use_token_type_ids: lowercase = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) lowercase = None lowercase = None lowercase = None if self.use_labels: lowercase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) lowercase = ids_tensor([self.batch_size] , self.num_choices ) lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE__ ( self ): return LlamaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): lowercase = LlamaModel(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case ) lowercase = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = True lowercase = LlamaModel(snake_case ) model.to(snake_case ) model.eval() lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , ) lowercase = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = LlamaForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = True lowercase = True lowercase = LlamaForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , use_cache=snake_case , ) lowercase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids lowercase = ids_tensor((self.batch_size, 3) , config.vocab_size ) lowercase = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and lowercase = torch.cat([input_ids, next_tokens] , dim=-1 ) lowercase = torch.cat([input_mask, next_mask] , dim=-1 ) lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , output_hidden_states=snake_case , )['hidden_states'][0] lowercase = model( snake_case , attention_mask=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )['hidden_states'][0] # select random slice lowercase = ids_tensor((1,) , output_from_past.shape[-1] ).item() lowercase = output_from_no_past[:, -3:, random_slice_idx].detach() lowercase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1E-3 ) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.prepare_config_and_inputs() ( ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ( lowercase ) , ) = config_and_inputs lowercase = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class A_ ( __lowerCamelCase , __lowerCamelCase , __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Tuple = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () _UpperCamelCase : List[Any] = (LlamaForCausalLM,) if is_torch_available() else () _UpperCamelCase : int = ( { """feature-extraction""": LlamaModel, """text-classification""": LlamaForSequenceClassification, """text-generation""": LlamaForCausalLM, """zero-shot""": LlamaForSequenceClassification, } if is_torch_available() else {} ) _UpperCamelCase : int = False _UpperCamelCase : int = False def SCREAMING_SNAKE_CASE__ ( self ): lowercase = LlamaModelTester(self ) lowercase = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def SCREAMING_SNAKE_CASE__ ( self ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: lowercase = type self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = 'single_label_classification' lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = 3 lowercase = 'multi_label_classification' lowercase = input_dict['input_ids'] lowercase = input_ids.ne(1 ).to(snake_case ) lowercase = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) lowercase = LlamaForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() lowercase = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('LLaMA buffers include complex numbers, which breaks this test' ) def SCREAMING_SNAKE_CASE__ ( self ): pass @parameterized.expand([('linear',), ('dynamic',)] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase , lowercase = self.model_tester.prepare_config_and_inputs_for_common() lowercase = ids_tensor([1, 10] , config.vocab_size ) lowercase = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase = LlamaModel(snake_case ) original_model.to(snake_case ) original_model.eval() lowercase = original_model(snake_case ).last_hidden_state lowercase = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights lowercase = {'type': scaling_type, 'factor': 10.0} lowercase = LlamaModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() lowercase = scaled_model(snake_case ).last_hidden_state lowercase = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1E-5 ) ) @require_torch class A_ ( unittest.TestCase ): '''simple docstring''' @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-7b-hf' , device_map='auto' ) lowercase = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-6.6_550, -4.1_227, -4.9_859, -3.2_406, 0.8_262, -3.0_033, 1.2_964, -3.3_699]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-12.8_281, -7.4_453, -0.4_639, -8.0_625, -7.2_500, -8.0_000, -6.4_883, -7.7_695, -7.8_438, -7.0_312, -6.2_188, -7.1_328, -1.8_496, 1.9_961, -8.6_250, -6.7_227, -12.8_281, -6.9_492, -7.0_742, -7.7_852, -7.5_820, -7.9_062, -6.9_375, -7.9_805, -8.3_438, -8.1_562, -8.0_469, -7.6_250, -7.7_422, -7.3_398,] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-2.0_622, -1.2_794, -1.1_638, -0.9_788, -1.4_603, -1.0_238, -1.7_893, -1.4_411]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-8.1_406, -8.0_547, 2.7_461, -1.2_344, -0.1_448, -1.8_262, -1.0_020, -1.8_154, -1.6_895, -1.8_516, -2.3_574, -0.9_277, 3.7_598, 6.5_742, -1.2_998, -0.1_177, -8.1_406, -2.9_688, -2.9_199, -3.1_699, -3.5_254, -2.3_555, -2.7_988, -3.4_141, -2.8_262, -4.5_195, -3.3_379, -3.3_164, -2.7_832, -3.0_273] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Logits are not exactly the same, once we fix the instabalities somehow, will update!' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-13b-chat-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) # Expected mean on dim = -1 lowercase = torch.tensor([[-0.8_562, -1.8_520, -0.7_551, -0.4_162, -1.5_161, -1.2_038, -2.4_823, -2.3_254]] ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # slicing logits[0, 0, 0:30] # fmt: off lowercase = torch.tensor([-2.2_227, 4.8_828, 0.9_023, -0.4_578, -0.7_871, -0.1_033, -0.6_221, -0.5_786, -0.7_803, -1.0_674, -1.2_920, -0.1_570, 0.8_008, 2.0_723, -0.9_497, 0.2_771, -2.2_227, -0.7_612, -1.4_346, -1.2_061, -1.6_426, -0.3_000, -0.7_139, -1.1_934, -1.8_691, -1.6_973, -1.5_947, -1.2_705, -0.3_523, -0.5_513] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) @unittest.skip( 'Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [1, 306, 4658, 278, 6593, 310, 2834, 338] lowercase = LlamaForCausalLM.from_pretrained('meta-llama/Llama-2-70b-hf' , device_map='auto' ) lowercase = model(torch.tensor(snake_case ) ) lowercase = torch.tensor( [[-4.2_327, -3.3_360, -4.6_665, -4.7_631, -1.8_180, -3.4_170, -1.4_211, -3.1_810]] , dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) , snake_case , atol=1E-2 , rtol=1E-2 ) # fmt: off lowercase = torch.tensor([-9.4_922, -3.9_551, 1.7_998, -5.6_758, -5.1_055, -5.8_984, -4.8_320, -6.8_086, -6.5_391, -5.6_172, -5.5_820, -5.5_352, 1.7_881, 3.6_289, -6.5_117, -3.4_785, -9.5_000, -6.0_352, -6.8_125, -6.0_195, -6.6_836, -5.4_727, -6.2_812, -6.0_391, -7.3_398, -7.4_297, -7.4_844, -6.5_820, -5.8_789, -5.5_312] ) # fmt: on torch.testing.assert_close(out[0, 0, :30] , snake_case , atol=1E-5 , rtol=1E-5 ) @unittest.skip('Model is curently gated' ) @slow def SCREAMING_SNAKE_CASE__ ( self ): lowercase = 'Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi' lowercase = 'Simply put, the theory of relativity states that ' lowercase = LlamaTokenizer.from_pretrained('meta-llama/Llama-2-13b-chat-hf' ) lowercase = tokenizer.encode(snake_case , return_tensors='pt' ) lowercase = LlamaForCausalLM.from_pretrained( 'meta-llama/Llama-2-13b-chat-hf' , device_map='sequential' , use_safetensors=snake_case ) # greedy generation outputs lowercase = model.generate(snake_case , max_new_tokens=64 , top_p=snake_case , temperature=1 , do_sample=snake_case ) lowercase = tokenizer.decode(generated_ids[0] , skip_special_tokens=snake_case ) self.assertEqual(snake_case , snake_case )
84
1
import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## UpperCAmelCase = 16 UpperCAmelCase = 32 def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE = 16 ): lowercase = AutoTokenizer.from_pretrained('bert-base-cased' ) lowercase = load_dataset('glue' , 'mrpc' ) def tokenize_function(__SCREAMING_SNAKE_CASE ): # max_length=None => use the model max length (it's actually the default) lowercase = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=__SCREAMING_SNAKE_CASE , max_length=__SCREAMING_SNAKE_CASE ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase = datasets.map( __SCREAMING_SNAKE_CASE , batched=__SCREAMING_SNAKE_CASE , remove_columns=['idx', 'sentence1', 'sentence2'] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(__SCREAMING_SNAKE_CASE ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase = 128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase = 16 elif accelerator.mixed_precision != "no": lowercase = 8 else: lowercase = None return tokenizer.pad( __SCREAMING_SNAKE_CASE , padding='longest' , max_length=__SCREAMING_SNAKE_CASE , pad_to_multiple_of=__SCREAMING_SNAKE_CASE , return_tensors='pt' , ) # Instantiate dataloaders. lowercase = DataLoader( tokenized_datasets['train'] , shuffle=__SCREAMING_SNAKE_CASE , collate_fn=__SCREAMING_SNAKE_CASE , batch_size=__SCREAMING_SNAKE_CASE ) lowercase = DataLoader( tokenized_datasets['validation'] , shuffle=__SCREAMING_SNAKE_CASE , collate_fn=__SCREAMING_SNAKE_CASE , batch_size=__SCREAMING_SNAKE_CASE ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''', None) == "1": from accelerate.test_utils.training import mocked_dataloaders UpperCAmelCase = mocked_dataloaders # noqa: F811 def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS' , __SCREAMING_SNAKE_CASE ) == "1": lowercase = 2 # Initialize accelerator lowercase = Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase = config['lr'] lowercase = int(config['num_epochs'] ) lowercase = int(config['seed'] ) lowercase = int(config['batch_size'] ) lowercase = evaluate.load('glue' , 'mrpc' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__SCREAMING_SNAKE_CASE ) def inner_training_loop(__SCREAMING_SNAKE_CASE ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__SCREAMING_SNAKE_CASE ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase = AutoModelForSequenceClassification.from_pretrained('bert-base-cased' , return_dict=__SCREAMING_SNAKE_CASE ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase = model.to(accelerator.device ) # Instantiate optimizer lowercase = AdamW(params=model.parameters() , lr=__SCREAMING_SNAKE_CASE ) lowercase , lowercase = get_dataloaders(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Instantiate scheduler lowercase = get_linear_schedule_with_warmup( optimizer=__SCREAMING_SNAKE_CASE , num_warmup_steps=100 , num_training_steps=(len(__SCREAMING_SNAKE_CASE ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase , lowercase , lowercase , lowercase , lowercase = accelerator.prepare( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # Now we train the model for epoch in range(__SCREAMING_SNAKE_CASE ): model.train() for step, batch in enumerate(__SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowercase = model(**__SCREAMING_SNAKE_CASE ) lowercase = outputs.loss accelerator.backward(__SCREAMING_SNAKE_CASE ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__SCREAMING_SNAKE_CASE ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase = model(**__SCREAMING_SNAKE_CASE ) lowercase = outputs.logits.argmax(dim=-1 ) lowercase , lowercase = accelerator.gather_for_metrics((predictions, batch['labels']) ) metric.add_batch( predictions=__SCREAMING_SNAKE_CASE , references=__SCREAMING_SNAKE_CASE , ) lowercase = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , __SCREAMING_SNAKE_CASE ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def UpperCAmelCase_ ( ): lowercase = argparse.ArgumentParser(description='Simple example of training script.' ) parser.add_argument( '--mixed_precision' , type=__SCREAMING_SNAKE_CASE , default=__SCREAMING_SNAKE_CASE , choices=['no', 'fp16', 'bf16', 'fp8'] , help='Whether to use mixed precision. Choose' 'between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.' 'and an Nvidia Ampere GPU.' , ) parser.add_argument('--cpu' , action='store_true' , help='If passed, will train on the CPU.' ) lowercase = parser.parse_args() lowercase = {'lr': 2e-5, 'num_epochs': 3, 'seed': 42, 'batch_size': 16} training_function(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if __name__ == "__main__": main()
84
# Lint as: python3 import os import re import urllib.parse from pathlib import Path from typing import Callable, List, Optional, Union from zipfile import ZipFile from ..utils.file_utils import cached_path, hf_github_url from ..utils.logging import get_logger from ..utils.version import Version UpperCAmelCase = get_logger(__name__) class A_ : '''simple docstring''' _UpperCamelCase : Dict = """dummy_data""" _UpperCamelCase : Optional[int] = """datasets""" _UpperCamelCase : Tuple = False def __init__( self , snake_case , snake_case , snake_case , snake_case = None , snake_case = False , snake_case = True , snake_case = None , ): lowercase = 0 lowercase = dataset_name lowercase = cache_dir lowercase = use_local_dummy_data lowercase = config # download_callbacks take a single url as input lowercase = download_callbacks or [] # if False, it doesn't load existing files and it returns the paths of the dummy files relative # to the dummy_data zip file root lowercase = load_existing_dummy_data # TODO(PVP, QL) might need to make this more general lowercase = str(snake_case ) # to be downloaded lowercase = None lowercase = None @property def SCREAMING_SNAKE_CASE__ ( self ): if self._dummy_file is None: lowercase = self.download_dummy_data() return self._dummy_file @property def SCREAMING_SNAKE_CASE__ ( self ): if self.config is not None: # structure is dummy / config_name / version_name return os.path.join('dummy' , self.config.name , self.version_name ) # structure is dummy / version_name return os.path.join('dummy' , self.version_name ) @property def SCREAMING_SNAKE_CASE__ ( self ): return os.path.join(self.dummy_data_folder , 'dummy_data.zip' ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = ( self.local_path_to_dummy_data if self.use_local_dummy_data is True else self.github_path_to_dummy_data ) lowercase = cached_path( snake_case , cache_dir=self.cache_dir , extract_compressed_file=snake_case , force_extract=snake_case ) return os.path.join(snake_case , self.dummy_file_name ) @property def SCREAMING_SNAKE_CASE__ ( self ): return os.path.join(self.datasets_scripts_dir , self.dataset_name , self.dummy_zip_file ) @property def SCREAMING_SNAKE_CASE__ ( self ): if self._bucket_url is None: lowercase = hf_github_url(self.dataset_name , self.dummy_zip_file.replace(os.sep , '/' ) ) return self._bucket_url @property def SCREAMING_SNAKE_CASE__ ( self ): # return full path if its a dir if os.path.isdir(self.dummy_file ): return self.dummy_file # else cut off path to file -> example `xsum`. return "/".join(self.dummy_file.replace(os.sep , '/' ).split('/' )[:-1] ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case ): if self.load_existing_dummy_data: # dummy data is downloaded and tested lowercase = self.dummy_file else: # dummy data cannot be downloaded and only the path to dummy file is returned lowercase = self.dummy_file_name # special case when data_url is a dict if isinstance(snake_case , snake_case ): return self.create_dummy_data_dict(snake_case , snake_case ) elif isinstance(snake_case , (list, tuple) ): return self.create_dummy_data_list(snake_case , snake_case ) else: return self.create_dummy_data_single(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case ): return self.download_and_extract(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): return self.download_and_extract(snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , *snake_case , **snake_case ): return path def SCREAMING_SNAKE_CASE__ ( self ): return {} def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = {} for key, single_urls in data_url.items(): for download_callback in self.download_callbacks: if isinstance(snake_case , snake_case ): for single_url in single_urls: download_callback(snake_case ) else: lowercase = single_urls download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus if isinstance(snake_case , snake_case ): lowercase = [os.path.join(snake_case , urllib.parse.quote_plus(Path(snake_case ).name ) ) for x in single_urls] else: lowercase = single_urls lowercase = os.path.join(snake_case , urllib.parse.quote_plus(Path(snake_case ).name ) ) lowercase = value # make sure that values are unique if all(isinstance(snake_case , snake_case ) for i in dummy_data_dict.values() ) and len(set(dummy_data_dict.values() ) ) < len( dummy_data_dict.values() ): # append key to value to make its name unique lowercase = {key: value + key for key, value in dummy_data_dict.items()} return dummy_data_dict def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = [] # trick: if there are many shards named like `data.txt-000001-of-00300`, only use the first one lowercase = all(bool(re.findall('[0-9]{3,}-of-[0-9]{3,}' , snake_case ) ) for url in data_url ) lowercase = all( url.startswith('https://ftp.ncbi.nlm.nih.gov/pubmed/baseline/pubmed' ) for url in data_url ) if data_url and (is_tf_records or is_pubmed_records): lowercase = [data_url[0]] * len(snake_case ) for single_url in data_url: for download_callback in self.download_callbacks: download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowercase = os.path.join(snake_case , urllib.parse.quote_plus(single_url.split('/' )[-1] ) ) dummy_data_list.append(snake_case ) return dummy_data_list def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): for download_callback in self.download_callbacks: download_callback(snake_case ) # we force the name of each key to be the last file / folder name of the url path # if the url has arguments, we need to encode them with urllib.parse.quote_plus lowercase = os.path.join(snake_case , urllib.parse.quote_plus(data_url.split('/' )[-1] ) ) if os.path.exists(snake_case ) or not self.load_existing_dummy_data: return value else: # Backward compatibility, maybe deprecate at one point. # For many datasets with single url calls to dl_manager.download_and_extract, # the dummy_data.zip file is actually the zipped downloaded file # while now we expected the dummy_data.zip file to be a directory containing # the downloaded file. return path_to_dummy_data def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self ): pass def SCREAMING_SNAKE_CASE__ ( self , snake_case ): def _iter_archive_members(snake_case ): # this preserves the order of the members inside the ZIP archive lowercase = Path(self.dummy_file ).parent lowercase = path.relative_to(snake_case ) with ZipFile(self.local_path_to_dummy_data ) as zip_file: lowercase = zip_file.namelist() for member in members: if member.startswith(relative_path.as_posix() ): yield dummy_parent_path.joinpath(snake_case ) lowercase = Path(snake_case ) lowercase = _iter_archive_members(snake_case ) if self.use_local_dummy_data else path.rglob('*' ) for file_path in file_paths: if file_path.is_file() and not file_path.name.startswith(('.', '__') ): yield file_path.relative_to(snake_case ).as_posix(), file_path.open('rb' ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if not isinstance(snake_case , snake_case ): lowercase = [paths] for path in paths: if os.path.isfile(snake_case ): if os.path.basename(snake_case ).startswith(('.', '__') ): return yield path else: for dirpath, dirnames, filenames in os.walk(snake_case ): if os.path.basename(snake_case ).startswith(('.', '__') ): continue dirnames.sort() for filename in sorted(snake_case ): if filename.startswith(('.', '__') ): continue yield os.path.join(snake_case , snake_case )
84
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) UpperCAmelCase = {'''configuration_encoder_decoder''': ['''EncoderDecoderConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''EncoderDecoderModel'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''TFEncoderDecoderModel'''] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''FlaxEncoderDecoderModel'''] if TYPE_CHECKING: from .configuration_encoder_decoder import EncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encoder_decoder import EncoderDecoderModel try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_encoder_decoder import TFEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_encoder_decoder import FlaxEncoderDecoderModel else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class A_ ( __lowerCamelCase , unittest.TestCase ): '''simple docstring''' _UpperCamelCase : Tuple = OpenAIGPTTokenizer _UpperCamelCase : List[Any] = OpenAIGPTTokenizerFast _UpperCamelCase : int = True _UpperCamelCase : List[Any] = False def SCREAMING_SNAKE_CASE__ ( self ): super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt lowercase = [ 'l', 'o', 'w', 'e', 'r', 's', 't', 'i', 'd', 'n', 'w</w>', 'r</w>', 't</w>', 'lo', 'low', 'er</w>', 'low</w>', 'lowest</w>', 'newer</w>', 'wider</w>', '<unk>', ] lowercase = dict(zip(snake_case , range(len(snake_case ) ) ) ) lowercase = ['#version: 0.2', 'l o', 'lo w', 'e r</w>', ''] lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['vocab_file'] ) lowercase = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES['merges_file'] ) with open(self.vocab_file , 'w' ) as fp: fp.write(json.dumps(snake_case ) ) with open(self.merges_file , 'w' ) as fp: fp.write('\n'.join(snake_case ) ) def SCREAMING_SNAKE_CASE__ ( self , snake_case ): return "lower newer", "lower newer" def SCREAMING_SNAKE_CASE__ ( self ): lowercase = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) lowercase = 'lower' lowercase = ['low', 'er</w>'] lowercase = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , snake_case ) lowercase = tokens + ['<unk>'] lowercase = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=15 ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): lowercase = self.rust_tokenizer_class.from_pretrained(snake_case , **snake_case ) # Simple input lowercase = 'This is a simple input' lowercase = ['This is a simple input 1', 'This is a simple input 2'] lowercase = ('This is a simple input', 'This is a pair') lowercase = [ ('This is a simple input 1', 'This is a simple input 2'), ('This is a simple pair 1', 'This is a simple pair 2'), ] # Simple input tests self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding='max_length' ) # Simple input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding='max_length' ) # Simple input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding='max_length' , ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding='max_length' ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding='max_length' ) # Pair input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding='max_length' , ) def SCREAMING_SNAKE_CASE__ ( self ): pass @require_ftfy @require_spacy @require_tokenizers class A_ ( __lowerCamelCase ): '''simple docstring''' pass
84
1
from __future__ import annotations class A_ : '''simple docstring''' def __init__( self , snake_case ): lowercase = order # a_{0} ... a_{k} lowercase = [1.0] + [0.0] * order # b_{0} ... b_{k} lowercase = [1.0] + [0.0] * order # x[n-1] ... x[n-k] lowercase = [0.0] * self.order # y[n-1] ... y[n-k] lowercase = [0.0] * self.order def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): if len(snake_case ) < self.order: lowercase = [1.0, *a_coeffs] if len(snake_case ) != self.order + 1: lowercase = ( F'''Expected a_coeffs to have {self.order + 1} elements ''' F'''for {self.order}-order filter, got {len(snake_case )}''' ) raise ValueError(snake_case ) if len(snake_case ) != self.order + 1: lowercase = ( F'''Expected b_coeffs to have {self.order + 1} elements ''' F'''for {self.order}-order filter, got {len(snake_case )}''' ) raise ValueError(snake_case ) lowercase = a_coeffs lowercase = b_coeffs def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) lowercase = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] lowercase = self.input_history[:-1] lowercase = self.output_history[:-1] lowercase = sample lowercase = result return result
84
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. UpperCAmelCase = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): config.addinivalue_line( 'markers' , 'is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested' ) config.addinivalue_line( 'markers' , 'is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested' ) config.addinivalue_line('markers' , 'is_pipeline_test: mark test to run only when pipelines are tested' ) config.addinivalue_line('markers' , 'is_staging_test: mark test to run only in the staging environment' ) config.addinivalue_line('markers' , 'accelerate_tests: mark test that require accelerate' ) config.addinivalue_line('markers' , 'tool_tests: mark the tool tests that are run on their specific schedule' ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_terminal_summary_main lowercase = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(__SCREAMING_SNAKE_CASE , id=__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: lowercase = 0 # Doctest custom flag to ignore output. UpperCAmelCase = doctest.register_optionflag('''IGNORE_RESULT''') UpperCAmelCase = doctest.OutputChecker class A_ ( __lowerCamelCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , snake_case , snake_case , snake_case ) UpperCAmelCase = CustomOutputChecker UpperCAmelCase = HfDoctestModule UpperCAmelCase = HfDocTestParser
84
1
from __future__ import annotations import time from math import sqrt # 1 for manhattan, 0 for euclidean UpperCAmelCase = 0 UpperCAmelCase = [ [0, 0, 0, 0, 0, 0, 0], [0, 1, 0, 0, 0, 0, 0], # 0 are free path whereas 1's are obstacles [0, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0], [1, 0, 1, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 1, 0, 0], ] UpperCAmelCase = [[-1, 0], [0, -1], [1, 0], [0, 1]] # up, left, down, right UpperCAmelCase = tuple[int, int] class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case , ): lowercase = pos_x lowercase = pos_y lowercase = (pos_y, pos_x) lowercase = goal_x lowercase = goal_y lowercase = g_cost lowercase = parent lowercase = self.calculate_heuristic() lowercase = self.g_cost + self.h_cost def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.pos_x - self.goal_x lowercase = self.pos_y - self.goal_y if HEURISTIC == 1: return abs(snake_case ) + abs(snake_case ) else: return sqrt(dy**2 + dx**2 ) def __lt__( self , snake_case ): return self.f_cost < other.f_cost class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case ): lowercase = Node(start[1] , start[0] , goal[1] , goal[0] , 0 , snake_case ) lowercase = Node(goal[1] , goal[0] , goal[1] , goal[0] , 9_9999 , snake_case ) lowercase = [self.start] lowercase = [] lowercase = False def SCREAMING_SNAKE_CASE__ ( self ): while self.open_nodes: # Open Nodes are sorted using __lt__ self.open_nodes.sort() lowercase = self.open_nodes.pop(0 ) if current_node.pos == self.target.pos: return self.retrace_path(snake_case ) self.closed_nodes.append(snake_case ) lowercase = self.get_successors(snake_case ) for child_node in successors: if child_node in self.closed_nodes: continue if child_node not in self.open_nodes: self.open_nodes.append(snake_case ) else: # retrieve the best current path lowercase = self.open_nodes.pop(self.open_nodes.index(snake_case ) ) if child_node.g_cost < better_node.g_cost: self.open_nodes.append(snake_case ) else: self.open_nodes.append(snake_case ) return [self.start.pos] def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = [] for action in delta: lowercase = parent.pos_x + action[1] lowercase = parent.pos_y + action[0] if not (0 <= pos_x <= len(grid[0] ) - 1 and 0 <= pos_y <= len(snake_case ) - 1): continue if grid[pos_y][pos_x] != 0: continue successors.append( Node( snake_case , snake_case , self.target.pos_y , self.target.pos_x , parent.g_cost + 1 , snake_case , ) ) return successors def SCREAMING_SNAKE_CASE__ ( self , snake_case ): lowercase = node lowercase = [] while current_node is not None: path.append((current_node.pos_y, current_node.pos_x) ) lowercase = current_node.parent path.reverse() return path class A_ : '''simple docstring''' def __init__( self , snake_case , snake_case ): lowercase = AStar(snake_case , snake_case ) lowercase = AStar(snake_case , snake_case ) lowercase = False def SCREAMING_SNAKE_CASE__ ( self ): while self.fwd_astar.open_nodes or self.bwd_astar.open_nodes: self.fwd_astar.open_nodes.sort() self.bwd_astar.open_nodes.sort() lowercase = self.fwd_astar.open_nodes.pop(0 ) lowercase = self.bwd_astar.open_nodes.pop(0 ) if current_bwd_node.pos == current_fwd_node.pos: return self.retrace_bidirectional_path( snake_case , snake_case ) self.fwd_astar.closed_nodes.append(snake_case ) self.bwd_astar.closed_nodes.append(snake_case ) lowercase = current_bwd_node lowercase = current_fwd_node lowercase = { self.fwd_astar: self.fwd_astar.get_successors(snake_case ), self.bwd_astar: self.bwd_astar.get_successors(snake_case ), } for astar in [self.fwd_astar, self.bwd_astar]: for child_node in successors[astar]: if child_node in astar.closed_nodes: continue if child_node not in astar.open_nodes: astar.open_nodes.append(snake_case ) else: # retrieve the best current path lowercase = astar.open_nodes.pop( astar.open_nodes.index(snake_case ) ) if child_node.g_cost < better_node.g_cost: astar.open_nodes.append(snake_case ) else: astar.open_nodes.append(snake_case ) return [self.fwd_astar.start.pos] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = self.fwd_astar.retrace_path(snake_case ) lowercase = self.bwd_astar.retrace_path(snake_case ) bwd_path.pop() bwd_path.reverse() lowercase = fwd_path + bwd_path return path if __name__ == "__main__": # all coordinates are given in format [y,x] UpperCAmelCase = (0, 0) UpperCAmelCase = (len(grid) - 1, len(grid[0]) - 1) for elem in grid: print(elem) UpperCAmelCase = time.time() UpperCAmelCase = AStar(init, goal) UpperCAmelCase = a_star.search() UpperCAmelCase = time.time() - start_time print(F"""AStar execution time = {end_time:f} seconds""") UpperCAmelCase = time.time() UpperCAmelCase = BidirectionalAStar(init, goal) UpperCAmelCase = time.time() - bd_start_time print(F"""BidirectionalAStar execution time = {bd_end_time:f} seconds""")
84
import torch from torch import nn class A_ ( nn.Module ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , snake_case , snake_case=1 , snake_case=False ): super().__init__() lowercase = n_token lowercase = d_embed lowercase = d_proj lowercase = cutoffs + [n_token] lowercase = [0] + self.cutoffs lowercase = div_val lowercase = self.cutoffs[0] lowercase = len(self.cutoffs ) - 1 lowercase = self.shortlist_size + self.n_clusters if self.n_clusters > 0: lowercase = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) lowercase = nn.Parameter(torch.zeros(self.n_clusters ) ) lowercase = nn.ModuleList() lowercase = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(snake_case , snake_case ) ) ) else: self.out_projs.append(snake_case ) self.out_layers.append(nn.Linear(snake_case , snake_case ) ) else: for i in range(len(self.cutoffs ) ): lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(snake_case , snake_case ) ) ) self.out_layers.append(nn.Linear(snake_case , r_idx - l_idx ) ) lowercase = keep_order def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case ): if proj is None: lowercase = nn.functional.linear(snake_case , snake_case , bias=snake_case ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: lowercase = nn.functional.linear(snake_case , proj.t().contiguous() ) lowercase = nn.functional.linear(snake_case , snake_case , bias=snake_case ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case=None , snake_case=False ): if labels is not None: # Shift so that tokens < n predict n lowercase = hidden[..., :-1, :].contiguous() lowercase = labels[..., 1:].contiguous() lowercase = hidden.view(-1 , hidden.size(-1 ) ) lowercase = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError('Input and labels should have the same size in the batch dimension.' ) else: lowercase = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: lowercase = self._compute_logit(snake_case , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: lowercase = labels != -100 lowercase = torch.zeros_like(snake_case , dtype=hidden.dtype , device=hidden.device ) lowercase = ( -nn.functional.log_softmax(snake_case , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: lowercase = nn.functional.log_softmax(snake_case , dim=-1 ) else: # construct weights and biases lowercase , lowercase = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = self.out_layers[0].weight[l_idx:r_idx] lowercase = self.out_layers[0].bias[l_idx:r_idx] else: lowercase = self.out_layers[i].weight lowercase = self.out_layers[i].bias if i == 0: lowercase = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowercase = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(snake_case ) biases.append(snake_case ) lowercase , lowercase , lowercase = weights[0], biases[0], self.out_projs[0] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) if labels is None: lowercase = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: lowercase = torch.zeros_like(snake_case , dtype=hidden.dtype , device=hidden.device ) lowercase = 0 lowercase = [0] + self.cutoffs for i in range(len(snake_case ) - 1 ): lowercase , lowercase = cutoff_values[i], cutoff_values[i + 1] if labels is not None: lowercase = (labels >= l_idx) & (labels < r_idx) lowercase = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue lowercase = labels.index_select(0 , snake_case ) - l_idx lowercase = head_logprob.index_select(0 , snake_case ) lowercase = hidden.index_select(0 , snake_case ) else: lowercase = hidden if i == 0: if labels is not None: lowercase = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: lowercase = head_logprob[:, : self.cutoffs[0]] else: lowercase , lowercase , lowercase = weights[i], biases[i], self.out_projs[i] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: lowercase = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: lowercase = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i lowercase = logprob_i if labels is not None: if (hasattr(self , 'keep_order' ) and self.keep_order) or keep_order: out.index_copy_(0 , snake_case , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def SCREAMING_SNAKE_CASE__ ( self , snake_case ): if self.n_clusters == 0: lowercase = self._compute_logit(snake_case , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(snake_case , dim=-1 ) else: # construct weights and biases lowercase , lowercase = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: lowercase , lowercase = self.cutoff_ends[i], self.cutoff_ends[i + 1] lowercase = self.out_layers[0].weight[l_idx:r_idx] lowercase = self.out_layers[0].bias[l_idx:r_idx] else: lowercase = self.out_layers[i].weight lowercase = self.out_layers[i].bias if i == 0: lowercase = torch.cat([weight_i, self.cluster_weight] , dim=0 ) lowercase = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(snake_case ) biases.append(snake_case ) lowercase , lowercase , lowercase = weights[0], biases[0], self.out_projs[0] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = hidden.new_empty((head_logit.size(0 ), self.n_token) ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = [0] + self.cutoffs for i in range(len(snake_case ) - 1 ): lowercase , lowercase = cutoff_values[i], cutoff_values[i + 1] if i == 0: lowercase = head_logprob[:, : self.cutoffs[0]] else: lowercase , lowercase , lowercase = weights[i], biases[i], self.out_projs[i] lowercase = self._compute_logit(snake_case , snake_case , snake_case , snake_case ) lowercase = nn.functional.log_softmax(snake_case , dim=1 ) lowercase = head_logprob[:, -i] + tail_logprob_i lowercase = logprob_i return out
84
1
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase = ''' Examples: ```py >>> import torch >>> import numpy as np >>> from diffusers import KandinskyV22PriorPipeline, KandinskyV22ControlnetPipeline >>> from transformers import pipeline >>> from diffusers.utils import load_image >>> def make_hint(image, depth_estimator): ... image = depth_estimator(image)["depth"] ... image = np.array(image) ... image = image[:, :, None] ... image = np.concatenate([image, image, image], axis=2) ... detected_map = torch.from_numpy(image).float() / 255.0 ... hint = detected_map.permute(2, 0, 1) ... return hint >>> depth_estimator = pipeline("depth-estimation") >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-prior", torch_dtype=torch.float16 ... ) >>> pipe_prior = pipe_prior.to("cuda") >>> pipe = KandinskyV22ControlnetPipeline.from_pretrained( ... "kandinsky-community/kandinsky-2-2-controlnet-depth", torch_dtype=torch.float16 ... ) >>> pipe = pipe.to("cuda") >>> img = load_image( ... "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" ... "/kandinsky/cat.png" ... ).resize((768, 768)) >>> hint = make_hint(img, depth_estimator).unsqueeze(0).half().to("cuda") >>> prompt = "A robot, 4k photo" >>> negative_prior_prompt = "lowres, text, error, cropped, worst quality, low quality, jpeg artifacts, ugly, duplicate, morbid, mutilated, out of frame, extra fingers, mutated hands, poorly drawn hands, poorly drawn face, mutation, deformed, blurry, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, username, watermark, signature" >>> generator = torch.Generator(device="cuda").manual_seed(43) >>> image_emb, zero_image_emb = pipe_prior( ... prompt=prompt, negative_prompt=negative_prior_prompt, generator=generator ... ).to_tuple() >>> images = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... hint=hint, ... num_inference_steps=50, ... generator=generator, ... height=768, ... width=768, ... ).images >>> images[0].save("robot_cat.png") ``` ''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=8 ): lowercase = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , ): super().__init__() self.register_modules( unet=snake_case , scheduler=snake_case , movq=snake_case , ) lowercase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): if latents is None: lowercase = randn_tensor(snake_case , generator=snake_case , device=snake_case , dtype=snake_case ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase = latents.to(snake_case ) lowercase = latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) lowercase = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=snake_case ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase , lowercase = cpu_offload_with_hook(snake_case , snake_case , prev_module_hook=snake_case ) # We'll offload the last model manually. lowercase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE__ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(snake_case , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(snake_case ) def __call__( self , snake_case , snake_case , snake_case , snake_case = 512 , snake_case = 512 , snake_case = 100 , snake_case = 4.0 , snake_case = 1 , snake_case = None , snake_case = None , snake_case = "pil" , snake_case = True , ): lowercase = self._execution_device lowercase = guidance_scale > 1.0 if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) lowercase = image_embeds.shape[0] * num_images_per_prompt if do_classifier_free_guidance: lowercase = image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = negative_image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = hint.repeat_interleave(snake_case , dim=0 ) lowercase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=snake_case ) lowercase = torch.cat([hint, hint] , dim=0 ).to(dtype=self.unet.dtype , device=snake_case ) self.scheduler.set_timesteps(snake_case , device=snake_case ) lowercase = self.scheduler.timesteps lowercase = self.movq.config.latent_channels lowercase , lowercase = downscale_height_and_width(snake_case , snake_case , self.movq_scale_factor ) # create initial latent lowercase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , snake_case , snake_case , snake_case , self.scheduler , ) for i, t in enumerate(self.progress_bar(snake_case ) ): # expand the latents if we are doing classifier free guidance lowercase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase = {'image_embeds': image_embeds, 'hint': hint} lowercase = self.unet( sample=snake_case , timestep=snake_case , encoder_hidden_states=snake_case , added_cond_kwargs=snake_case , return_dict=snake_case , )[0] if do_classifier_free_guidance: lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) lowercase , lowercase = noise_pred.chunk(2 ) lowercase , lowercase = variance_pred.chunk(2 ) lowercase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase = self.scheduler.step( snake_case , snake_case , snake_case , generator=snake_case , )[0] # post-processing lowercase = self.movq.decode(snake_case , force_not_quantize=snake_case )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase = image * 0.5 + 0.5 lowercase = image.clamp(0 , 1 ) lowercase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase = self.numpy_to_pil(snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=snake_case )
84
from __future__ import annotations class A_ : '''simple docstring''' def __init__( self , snake_case ): lowercase = TypeError( 'Matrices must be formed from a list of zero or more lists containing at ' 'least one and the same number of values, each of which must be of type ' 'int or float.' ) if len(snake_case ) != 0: lowercase = len(rows[0] ) if cols == 0: raise error for row in rows: if len(snake_case ) != cols: raise error for value in row: if not isinstance(snake_case , (int, float) ): raise error lowercase = rows else: lowercase = [] def SCREAMING_SNAKE_CASE__ ( self ): return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows ) @property def SCREAMING_SNAKE_CASE__ ( self ): return len(self.rows[0] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return (self.num_rows, self.num_columns) @property def SCREAMING_SNAKE_CASE__ ( self ): return self.order[0] == self.order[1] def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def SCREAMING_SNAKE_CASE__ ( self ): return bool(self.determinant() ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): lowercase = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(snake_case ).determinant() def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case ): if (row + column) % 2 == 0: return self.get_minor(snake_case , snake_case ) return -1 * self.get_minor(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [self.get_minor(snake_case , snake_case ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(snake_case ) def SCREAMING_SNAKE_CASE__ ( self ): lowercase = self.determinant() if not determinant: raise TypeError('Only matrices with a non-zero determinant have an inverse' ) return self.adjugate() * (1 / determinant) def __repr__( self ): return str(self.rows ) def __str__( self ): if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ '[' + '. '.join([str(snake_case ) for value in row] ) + '.]' for row in self.rows ] ) + "]" ) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError('Row must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in row: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_columns: raise ValueError( 'Row must be equal in length to the other rows in the matrix' ) if position is None: self.rows.append(snake_case ) else: lowercase = self.rows[0:position] + [row] + self.rows[position:] def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case = None ): lowercase = TypeError( 'Column must be a list containing all ints and/or floats' ) if not isinstance(snake_case , snake_case ): raise type_error for value in column: if not isinstance(snake_case , (int, float) ): raise type_error if len(snake_case ) != self.num_rows: raise ValueError( 'Column must be equal in length to the other columns in the matrix' ) if position is None: lowercase = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: lowercase = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__( self , snake_case ): if not isinstance(snake_case , snake_case ): return NotImplemented return self.rows == other.rows def __ne__( self , snake_case ): return not self == other def __neg__( self ): return self * -1 def __add__( self , snake_case ): if self.order != other.order: raise ValueError('Addition requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__( self , snake_case ): if self.order != other.order: raise ValueError('Subtraction requires matrices of the same order' ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__( self , snake_case ): if isinstance(snake_case , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(snake_case , snake_case ): if self.num_columns != other.num_rows: raise ValueError( 'The number of columns in the first matrix must ' 'be equal to the number of rows in the second' ) return Matrix( [ [Matrix.dot_product(snake_case , snake_case ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( 'A Matrix can only be multiplied by an int, float, or another matrix' ) def __pow__( self , snake_case ): if not isinstance(snake_case , snake_case ): raise TypeError('A Matrix can only be raised to the power of an int' ) if not self.is_square: raise ValueError('Only square matrices can be raised to a power' ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( 'Only invertable matrices can be raised to a negative power' ) lowercase = self for _ in range(other - 1 ): result *= self return result @classmethod def SCREAMING_SNAKE_CASE__ ( cls , snake_case , snake_case ): return sum(row[i] * column[i] for i in range(len(snake_case ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
84
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''MIT/ast-finetuned-audioset-10-10-0.4593''': ( '''https://huggingface.co/MIT/ast-finetuned-audioset-10-10-0.4593/resolve/main/config.json''' ), } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[Any] = """audio-spectrogram-transformer""" def __init__( self , snake_case=768 , snake_case=12 , snake_case=12 , snake_case=3072 , snake_case="gelu" , snake_case=0.0 , snake_case=0.0 , snake_case=0.02 , snake_case=1E-12 , snake_case=16 , snake_case=True , snake_case=10 , snake_case=10 , snake_case=1024 , snake_case=128 , **snake_case , ): super().__init__(**snake_case ) lowercase = hidden_size lowercase = num_hidden_layers lowercase = num_attention_heads lowercase = intermediate_size lowercase = hidden_act lowercase = hidden_dropout_prob lowercase = attention_probs_dropout_prob lowercase = initializer_range lowercase = layer_norm_eps lowercase = patch_size lowercase = qkv_bias lowercase = frequency_stride lowercase = time_stride lowercase = max_length lowercase = num_mel_bins
84
from typing import List, Optional, Union import torch from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name UpperCAmelCase = ''' Examples: ```py >>> from diffusers import KandinskyV22Pipeline, KandinskyV22PriorPipeline >>> import torch >>> pipe_prior = KandinskyV22PriorPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-prior") >>> pipe_prior.to("cuda") >>> prompt = "red cat, 4k photo" >>> out = pipe_prior(prompt) >>> image_emb = out.image_embeds >>> zero_image_emb = out.negative_image_embeds >>> pipe = KandinskyV22Pipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder") >>> pipe.to("cuda") >>> image = pipe( ... image_embeds=image_emb, ... negative_image_embeds=zero_image_emb, ... height=768, ... width=768, ... num_inference_steps=50, ... ).images >>> image[0].save("cat.png") ``` ''' def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=8 ): lowercase = height // scale_factor**2 if height % scale_factor**2 != 0: new_height += 1 lowercase = width // scale_factor**2 if width % scale_factor**2 != 0: new_width += 1 return new_height * scale_factor, new_width * scale_factor class A_ ( __lowerCamelCase ): '''simple docstring''' def __init__( self , snake_case , snake_case , snake_case , ): super().__init__() self.register_modules( unet=snake_case , scheduler=snake_case , movq=snake_case , ) lowercase = 2 ** (len(self.movq.config.block_out_channels ) - 1) def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case , snake_case , snake_case , snake_case ): if latents is None: lowercase = randn_tensor(snake_case , generator=snake_case , device=snake_case , dtype=snake_case ) else: if latents.shape != shape: raise ValueError(F'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase = latents.to(snake_case ) lowercase = latents * scheduler.init_noise_sigma return latents def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) lowercase = [ self.unet, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self , snake_case=0 ): if is_accelerate_available() and is_accelerate_version('>=' , '0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowercase = torch.device(F'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' , silence_dtype_warnings=snake_case ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase = None for cpu_offloaded_model in [self.unet, self.movq]: lowercase , lowercase = cpu_offload_with_hook(snake_case , snake_case , prev_module_hook=snake_case ) # We'll offload the last model manually. lowercase = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def SCREAMING_SNAKE_CASE__ ( self ): if not hasattr(self.unet , '_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(snake_case , '_hf_hook' ) and hasattr(module._hf_hook , 'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(snake_case ) def __call__( self , snake_case , snake_case , snake_case = 512 , snake_case = 512 , snake_case = 100 , snake_case = 4.0 , snake_case = 1 , snake_case = None , snake_case = None , snake_case = "pil" , snake_case = True , ): lowercase = self._execution_device lowercase = guidance_scale > 1.0 if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) lowercase = image_embeds.shape[0] * num_images_per_prompt if isinstance(snake_case , snake_case ): lowercase = torch.cat(snake_case , dim=0 ) if do_classifier_free_guidance: lowercase = image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = negative_image_embeds.repeat_interleave(snake_case , dim=0 ) lowercase = torch.cat([negative_image_embeds, image_embeds] , dim=0 ).to(dtype=self.unet.dtype , device=snake_case ) self.scheduler.set_timesteps(snake_case , device=snake_case ) lowercase = self.scheduler.timesteps lowercase = self.unet.config.in_channels lowercase , lowercase = downscale_height_and_width(snake_case , snake_case , self.movq_scale_factor ) # create initial latent lowercase = self.prepare_latents( (batch_size, num_channels_latents, height, width) , image_embeds.dtype , snake_case , snake_case , snake_case , self.scheduler , ) for i, t in enumerate(self.progress_bar(snake_case ) ): # expand the latents if we are doing classifier free guidance lowercase = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase = {'image_embeds': image_embeds} lowercase = self.unet( sample=snake_case , timestep=snake_case , encoder_hidden_states=snake_case , added_cond_kwargs=snake_case , return_dict=snake_case , )[0] if do_classifier_free_guidance: lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) lowercase , lowercase = noise_pred.chunk(2 ) lowercase , lowercase = variance_pred.chunk(2 ) lowercase = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase = torch.cat([noise_pred, variance_pred_text] , dim=1 ) if not ( hasattr(self.scheduler.config , 'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase , lowercase = noise_pred.split(latents.shape[1] , dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase = self.scheduler.step( snake_case , snake_case , snake_case , generator=snake_case , )[0] # post-processing lowercase = self.movq.decode(snake_case , force_not_quantize=snake_case )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(F'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase = image * 0.5 + 0.5 lowercase = image.clamp(0 , 1 ) lowercase = image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() if output_type == "pil": lowercase = self.numpy_to_pil(snake_case ) if not return_dict: return (image,) return ImagePipelineOutput(images=snake_case )
84
1
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase , lowercase = set(__SCREAMING_SNAKE_CASE ), [start] while stack: lowercase = stack.pop() explored.add(__SCREAMING_SNAKE_CASE ) # Differences from BFS: # 1) pop last element instead of first one # 2) add adjacent elements to stack without exploring them for adj in reversed(graph[v] ): if adj not in explored: stack.append(__SCREAMING_SNAKE_CASE ) return explored UpperCAmelCase = { '''A''': ['''B''', '''C''', '''D'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F'''], '''D''': ['''B''', '''D'''], '''E''': ['''B''', '''F'''], '''F''': ['''C''', '''E''', '''G'''], '''G''': ['''F'''], } if __name__ == "__main__": import doctest doctest.testmod() print(depth_first_search(G, '''A'''))
84
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): if digit_amount > 0: return round(number - int(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) return number - int(__SCREAMING_SNAKE_CASE ) if __name__ == "__main__": print(decimal_isolate(1.53, 0)) print(decimal_isolate(35.345, 1)) print(decimal_isolate(35.345, 2)) print(decimal_isolate(35.345, 3)) print(decimal_isolate(-14.789, 3)) print(decimal_isolate(0, 2)) print(decimal_isolate(-14.123, 1)) print(decimal_isolate(-14.123, 2)) print(decimal_isolate(-14.123, 3))
84
1
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE=None ): if subparsers is not None: lowercase = subparsers.add_parser('env' ) else: lowercase = argparse.ArgumentParser('Accelerate env command' ) parser.add_argument( '--config_file' , default=__SCREAMING_SNAKE_CASE , help='The config file to use for the default values in the launching script.' ) if subparsers is not None: parser.set_defaults(func=__SCREAMING_SNAKE_CASE ) return parser def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = torch.__version__ lowercase = torch.cuda.is_available() lowercase = is_xpu_available() lowercase = is_npu_available() lowercase = 'Not found' # Get the default from the config file. if args.config_file is not None or os.path.isfile(__SCREAMING_SNAKE_CASE ): lowercase = load_config_from_file(args.config_file ).to_dict() lowercase = { '`Accelerate` version': version, 'Platform': platform.platform(), 'Python version': platform.python_version(), 'Numpy version': np.__version__, 'PyTorch version (GPU?)': F'''{pt_version} ({pt_cuda_available})''', 'PyTorch XPU available': str(__SCREAMING_SNAKE_CASE ), 'PyTorch NPU available': str(__SCREAMING_SNAKE_CASE ), 'System RAM': F'''{psutil.virtual_memory().total / 1024 ** 3:.2f} GB''', } if pt_cuda_available: lowercase = torch.cuda.get_device_name() print('\nCopy-and-paste the text below in your GitHub issue\n' ) print('\n'.join([F'''- {prop}: {val}''' for prop, val in info.items()] ) ) print('- `Accelerate` default config:' if args.config_file is None else '- `Accelerate` config passed:' ) lowercase = ( '\n'.join([F'''\t- {prop}: {val}''' for prop, val in accelerate_config.items()] ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else F'''\t{accelerate_config}''' ) print(__SCREAMING_SNAKE_CASE ) lowercase = accelerate_config return info def UpperCAmelCase_ ( ): lowercase = env_command_parser() lowercase = parser.parse_args() env_command(__SCREAMING_SNAKE_CASE ) return 0 if __name__ == "__main__": raise SystemExit(main())
84
from __future__ import annotations def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) return n == n[::-1] def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE = 100_0000 ): lowercase = 0 for i in range(1 , __SCREAMING_SNAKE_CASE ): if is_palindrome(__SCREAMING_SNAKE_CASE ) and is_palindrome(bin(__SCREAMING_SNAKE_CASE ).split('b' )[1] ): total += i return total if __name__ == "__main__": print(solution(int(str(input().strip()))))
84
1
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import doctest import sys import warnings from os.path import abspath, dirname, join import _pytest from transformers.testing_utils import HfDoctestModule, HfDocTestParser # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. UpperCAmelCase = abspath(join(dirname(__file__), '''src''')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='''ignore''', category=FutureWarning) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): config.addinivalue_line( 'markers' , 'is_pt_tf_cross_test: mark test to run only when PT and TF interactions are tested' ) config.addinivalue_line( 'markers' , 'is_pt_flax_cross_test: mark test to run only when PT and FLAX interactions are tested' ) config.addinivalue_line('markers' , 'is_pipeline_test: mark test to run only when pipelines are tested' ) config.addinivalue_line('markers' , 'is_staging_test: mark test to run only in the staging environment' ) config.addinivalue_line('markers' , 'accelerate_tests: mark test that require accelerate' ) config.addinivalue_line('markers' , 'tool_tests: mark the tool tests that are run on their specific schedule' ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_addoption_shared pytest_addoption_shared(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): from transformers.testing_utils import pytest_terminal_summary_main lowercase = terminalreporter.config.getoption('--make-reports' ) if make_reports: pytest_terminal_summary_main(__SCREAMING_SNAKE_CASE , id=__SCREAMING_SNAKE_CASE ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): # If no tests are collected, pytest exists with code 5, which makes the CI fail. if exitstatus == 5: lowercase = 0 # Doctest custom flag to ignore output. UpperCAmelCase = doctest.register_optionflag('''IGNORE_RESULT''') UpperCAmelCase = doctest.OutputChecker class A_ ( __lowerCamelCase ): '''simple docstring''' def SCREAMING_SNAKE_CASE__ ( self , snake_case , snake_case , snake_case ): if IGNORE_RESULT & optionflags: return True return OutputChecker.check_output(self , snake_case , snake_case , snake_case ) UpperCAmelCase = CustomOutputChecker UpperCAmelCase = HfDoctestModule UpperCAmelCase = HfDocTestParser
84
import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING UpperCAmelCase = logging.get_logger(__name__) UpperCAmelCase = { '''microsoft/conditional-detr-resnet-50''': ( '''https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json''' ), } class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[Any] = """conditional_detr""" _UpperCamelCase : Any = ["""past_key_values"""] _UpperCamelCase : Optional[Any] = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", } def __init__( self , snake_case=True , snake_case=None , snake_case=3 , snake_case=300 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=6 , snake_case=2048 , snake_case=8 , snake_case=0.0 , snake_case=0.0 , snake_case=True , snake_case="relu" , snake_case=256 , snake_case=0.1 , snake_case=0.0 , snake_case=0.0 , snake_case=0.02 , snake_case=1.0 , snake_case=False , snake_case="sine" , snake_case="resnet50" , snake_case=True , snake_case=False , snake_case=2 , snake_case=5 , snake_case=2 , snake_case=1 , snake_case=1 , snake_case=2 , snake_case=5 , snake_case=2 , snake_case=0.25 , **snake_case , ): if backbone_config is not None and use_timm_backbone: raise ValueError('You can\'t specify both `backbone_config` and `use_timm_backbone`.' ) if not use_timm_backbone: if backbone_config is None: logger.info('`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.' ) lowercase = CONFIG_MAPPING['resnet'](out_features=['stage4'] ) elif isinstance(snake_case , snake_case ): lowercase = backbone_config.get('model_type' ) lowercase = CONFIG_MAPPING[backbone_model_type] lowercase = config_class.from_dict(snake_case ) lowercase = use_timm_backbone lowercase = backbone_config lowercase = num_channels lowercase = num_queries lowercase = d_model lowercase = encoder_ffn_dim lowercase = encoder_layers lowercase = encoder_attention_heads lowercase = decoder_ffn_dim lowercase = decoder_layers lowercase = decoder_attention_heads lowercase = dropout lowercase = attention_dropout lowercase = activation_dropout lowercase = activation_function lowercase = init_std lowercase = init_xavier_std lowercase = encoder_layerdrop lowercase = decoder_layerdrop lowercase = encoder_layers lowercase = auxiliary_loss lowercase = position_embedding_type lowercase = backbone lowercase = use_pretrained_backbone lowercase = dilation # Hungarian matcher lowercase = class_cost lowercase = bbox_cost lowercase = giou_cost # Loss coefficients lowercase = mask_loss_coefficient lowercase = dice_loss_coefficient lowercase = cls_loss_coefficient lowercase = bbox_loss_coefficient lowercase = giou_loss_coefficient lowercase = focal_alpha super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def SCREAMING_SNAKE_CASE__ ( self ): return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE__ ( self ): return self.d_model def SCREAMING_SNAKE_CASE__ ( self ): lowercase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: lowercase = self.backbone_config.to_dict() lowercase = self.__class__.model_type return output class A_ ( __lowerCamelCase ): '''simple docstring''' _UpperCamelCase : List[str] = version.parse("""1.11""" ) @property def SCREAMING_SNAKE_CASE__ ( self ): return OrderedDict( [ ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ('pixel_mask', {0: 'batch'}), ] ) @property def SCREAMING_SNAKE_CASE__ ( self ): return 1E-5 @property def SCREAMING_SNAKE_CASE__ ( self ): return 12
84
1
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return number | (1 << position) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return number & ~(1 << position) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return number ^ (1 << position) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return ((number >> position) & 1) == 1 def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
84
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available UpperCAmelCase = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCAmelCase = ['''MLukeTokenizer'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_mluke import MLukeTokenizer else: import sys UpperCAmelCase = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
84
1
import os import pytest import yaml from datasets.features.features import Features, Value from datasets.info import DatasetInfo, DatasetInfosDict @pytest.mark.parametrize( 'files' , [ ['full:README.md', 'dataset_infos.json'], ['empty:README.md', 'dataset_infos.json'], ['dataset_infos.json'], ['full:README.md'], ] , ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = tmp_path_factory.mktemp('dset_infos_dir' ) if "full:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('---\ndataset_info:\n dataset_size: 42\n---' ) if "empty:README.md" in files: with open(dataset_infos_dir / 'README.md' , 'w' ) as f: f.write('' ) # we want to support dataset_infos.json for backward compatibility if "dataset_infos.json" in files: with open(dataset_infos_dir / 'dataset_infos.json' , 'w' ) as f: f.write('{"default": {"dataset_size": 42}}' ) lowercase = DatasetInfosDict.from_directory(__SCREAMING_SNAKE_CASE ) assert dataset_infos assert dataset_infos["default"].dataset_size == 42 @pytest.mark.parametrize( 'dataset_info' , [ DatasetInfo(), DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ), ] , ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) dataset_info.write_to_directory(__SCREAMING_SNAKE_CASE ) lowercase = DatasetInfo.from_directory(__SCREAMING_SNAKE_CASE ) assert dataset_info == reloaded assert os.path.exists(os.path.join(__SCREAMING_SNAKE_CASE , 'dataset_info.json' ) ) def UpperCAmelCase_ ( ): lowercase = DatasetInfo( description='foo' , citation='bar' , homepage='https://foo.bar' , license='CC0' , features=Features({'a': Value('int32' )} ) , post_processed={} , supervised_keys=() , task_templates=[] , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train', 'num_examples': 42}] , download_checksums={} , download_size=1337 , post_processing_size=442 , dataset_size=1234 , size_in_bytes=1337 + 442 + 1234 , ) lowercase = dataset_info._to_yaml_dict() assert sorted(__SCREAMING_SNAKE_CASE ) == sorted(DatasetInfo._INCLUDED_INFO_IN_YAML ) for key in DatasetInfo._INCLUDED_INFO_IN_YAML: assert key in dataset_info_yaml_dict assert isinstance(dataset_info_yaml_dict[key] , (list, dict, int, str) ) lowercase = yaml.safe_dump(__SCREAMING_SNAKE_CASE ) lowercase = yaml.safe_load(__SCREAMING_SNAKE_CASE ) assert dataset_info_yaml_dict == reloaded def UpperCAmelCase_ ( ): lowercase = DatasetInfo() lowercase = dataset_info._to_yaml_dict() assert dataset_info_yaml_dict == {} @pytest.mark.parametrize( 'dataset_infos_dict' , [ DatasetInfosDict(), DatasetInfosDict({'default': DatasetInfo()} ), DatasetInfosDict({'my_config_name': DatasetInfo()} ), DatasetInfosDict( { 'default': DatasetInfo( description='foo' , features=Features({'a': Value('int32' )} ) , builder_name='builder' , config_name='config' , version='1.0.0' , splits=[{'name': 'train'}] , download_size=42 , ) } ), DatasetInfosDict( { 'v1': DatasetInfo(dataset_size=42 ), 'v2': DatasetInfo(dataset_size=1337 ), } ), ] , ) def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): lowercase = str(__SCREAMING_SNAKE_CASE ) dataset_infos_dict.write_to_directory(__SCREAMING_SNAKE_CASE ) lowercase = DatasetInfosDict.from_directory(__SCREAMING_SNAKE_CASE ) # the config_name of the dataset_infos_dict take over the attribute for config_name, dataset_info in dataset_infos_dict.items(): lowercase = config_name # the yaml representation doesn't include fields like description or citation # so we just test that we can recover what we can from the yaml lowercase = DatasetInfo._from_yaml_dict(dataset_info._to_yaml_dict() ) assert dataset_infos_dict == reloaded if dataset_infos_dict: assert os.path.exists(os.path.join(__SCREAMING_SNAKE_CASE , 'README.md' ) )
84
def UpperCAmelCase_ ( __SCREAMING_SNAKE_CASE ): lowercase = [0] * len(__SCREAMING_SNAKE_CASE ) lowercase = [] lowercase = [] lowercase = 0 for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(__SCREAMING_SNAKE_CASE ) ): if indegree[i] == 0: queue.append(__SCREAMING_SNAKE_CASE ) while queue: lowercase = queue.pop(0 ) cnt += 1 topo.append(__SCREAMING_SNAKE_CASE ) for x in graph[vertex]: indegree[x] -= 1 if indegree[x] == 0: queue.append(__SCREAMING_SNAKE_CASE ) if cnt != len(__SCREAMING_SNAKE_CASE ): print('Cycle exists' ) else: print(__SCREAMING_SNAKE_CASE ) # Adjacency List of Graph UpperCAmelCase = {0: [1, 2], 1: [3], 2: [3], 3: [4, 5], 4: [], 5: []} topological_sort(graph)
84
1