code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
'''simple docstring''' def _A ( A__ , A__ ): """simple docstring""" return number | (1 << position) def _A ( A__ , A__ ): """simple docstring""" return number & ~(1 << position) def _A ( A__ , A__ ): """simple docstring""" return number ^ (1 << position) def _A ( A__ , A__ ): """simple docstring""" return ((number >> position) & 1) == 1 def _A ( A__ , A__ ): """simple docstring""" return int((number & (1 << position)) != 0 ) if __name__ == "__main__": import doctest doctest.testmod()
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
def _A ( A__ ): """simple docstring""" if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise ValueError('''multiplicative_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''multiplicative_persistence() does not accept negative values''' ) __lowercase = 0 __lowercase = str(UpperCAmelCase__ ) while len(UpperCAmelCase__ ) != 1: __lowercase = [int(UpperCAmelCase__ ) for i in num_string] __lowercase = 1 for i in range(0 , len(UpperCAmelCase__ ) ): total *= numbers[i] __lowercase = str(UpperCAmelCase__ ) steps += 1 return steps def _A ( A__ ): """simple docstring""" if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise ValueError('''additive_persistence() only accepts integral values''' ) if num < 0: raise ValueError('''additive_persistence() does not accept negative values''' ) __lowercase = 0 __lowercase = str(UpperCAmelCase__ ) while len(UpperCAmelCase__ ) != 1: __lowercase = [int(UpperCAmelCase__ ) for i in num_string] __lowercase = 0 for i in range(0 , len(UpperCAmelCase__ ) ): total += numbers[i] __lowercase = str(UpperCAmelCase__ ) steps += 1 return steps if __name__ == "__main__": import doctest doctest.testmod()
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer lowerCAmelCase__ = '''bart''' lowerCAmelCase__ = True @st.cache(allow_output_mutation=_snake_case ) def _A ( ): """simple docstring""" if LOAD_DENSE_INDEX: __lowercase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) __lowercase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) __lowercase = qar_model.eval() else: __lowercase , __lowercase = (None, None) if MODEL_TYPE == "bart": __lowercase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) __lowercase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) __lowercase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) __lowercase = sas_model.eval() else: __lowercase , __lowercase = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=_snake_case ) def _A ( ): """simple docstring""" if LOAD_DENSE_INDEX: __lowercase = faiss.StandardGpuResources() __lowercase = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] __lowercase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 128) , ) __lowercase = faiss.IndexFlatIP(128 ) __lowercase = faiss.index_cpu_to_gpu(_snake_case , 1 , _snake_case ) wikiaab_gpu_index_flat.add(_snake_case ) # TODO fix for larger GPU else: __lowercase , __lowercase = (None, None) __lowercase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=_snake_case ) def _A ( ): """simple docstring""" __lowercase = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) __lowercase = elia['''train_eli5'''] __lowercase = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 128) ) __lowercase = faiss.IndexFlatIP(128 ) eli5_train_q_index.add(_snake_case ) return (elia_train, eli5_train_q_index) lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = load_indexes() lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = load_models() lowerCAmelCase__ , lowerCAmelCase__ = load_train_data() def _A ( A__ , A__=10 ): """simple docstring""" __lowercase = embed_questions_for_retrieval([question] , _snake_case , _snake_case ) __lowercase , __lowercase = eli5_train_q_index.search(_snake_case , _snake_case ) __lowercase = [elia_train[int(_snake_case )] for i in I[0]] return nn_examples def _A ( A__ , A__="wiki40b" , A__="dense" , A__=10 ): """simple docstring""" if source == "none": __lowercase , __lowercase = (''' <P> '''.join(['''''' for _ in range(11 )] ).strip(), []) else: if method == "dense": __lowercase , __lowercase = query_qa_dense_index( _snake_case , _snake_case , _snake_case , _snake_case , _snake_case , _snake_case ) else: __lowercase , __lowercase = query_es_index( _snake_case , _snake_case , index_name='''english_wiki40b_snippets_100w''' , n_results=_snake_case , ) __lowercase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] __lowercase = '''question: {} context: {}'''.format(_snake_case , _snake_case ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda A__ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda A__ : None), } ) def _A ( A__ , A__ , A__ , A__=64 , A__=256 , A__=False , A__=2 , A__=0.9_5 , A__=0.8 ): """simple docstring""" with torch.no_grad(): __lowercase = qa_sas_generate( _snake_case , _snake_case , _snake_case , num_answers=1 , num_beams=_snake_case , min_len=_snake_case , max_len=_snake_case , do_sample=_snake_case , temp=_snake_case , top_p=_snake_case , top_k=_snake_case , max_input_length=1024 , device='''cuda:0''' , )[0] return (answer, support_list) st.title('''Long Form Question Answering with ELI5''') # Start sidebar lowerCAmelCase__ = '''<img src=\'https://huggingface.co/front/assets/huggingface_logo.svg\'>''' lowerCAmelCase__ = ''' <html> <head> <style> .img-container { padding-left: 90px; padding-right: 90px; padding-top: 50px; padding-bottom: 50px; background-color: #f0f3f9; } </style> </head> <body> <span class=\"img-container\"> <!-- Inline parent element --> %s </span> </body> </html> ''' % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia lowerCAmelCase__ = ''' This demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html). First, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset, a pre-processed fixed snapshot of Wikipedia. ''' st.sidebar.markdown(description, unsafe_allow_html=True) lowerCAmelCase__ = [ '''Answer the question''', '''View the retrieved document only''', '''View the most similar ELI5 question and answer''', '''Show me everything, please!''', ] lowerCAmelCase__ = st.sidebar.checkbox('''Demo options''') if demo_options: lowerCAmelCase__ = st.sidebar.selectbox( '''''', action_list, index=3, ) lowerCAmelCase__ = action_list.index(action_st) lowerCAmelCase__ = st.sidebar.selectbox( '''''', ['''Show full text of passages''', '''Show passage section titles'''], index=0, ) lowerCAmelCase__ = show_type == '''Show full text of passages''' else: lowerCAmelCase__ = 3 lowerCAmelCase__ = True lowerCAmelCase__ = st.sidebar.checkbox('''Retrieval options''') if retrieval_options: lowerCAmelCase__ = ''' ### Information retriever options The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs. The answer is then generated by sequence to sequence model which takes the question and retrieved document as input. ''' st.sidebar.markdown(retriever_info) lowerCAmelCase__ = st.sidebar.selectbox('''Which Wikipedia format should the model use?''', ['''wiki40b''', '''none''']) lowerCAmelCase__ = st.sidebar.selectbox('''Which Wikipedia indexer should the model use?''', ['''dense''', '''sparse''', '''mixed''']) else: lowerCAmelCase__ = '''wiki40b''' lowerCAmelCase__ = '''dense''' lowerCAmelCase__ = '''beam''' lowerCAmelCase__ = 2 lowerCAmelCase__ = 64 lowerCAmelCase__ = 256 lowerCAmelCase__ = None lowerCAmelCase__ = None lowerCAmelCase__ = st.sidebar.checkbox('''Generation options''') if generate_options: lowerCAmelCase__ = ''' ### Answer generation options The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large) weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with **beam** search, or **sample** from the decoder\'s output probabilities. ''' st.sidebar.markdown(generate_info) lowerCAmelCase__ = st.sidebar.selectbox('''Would you like to use beam search or sample an answer?''', ['''beam''', '''sampled''']) lowerCAmelCase__ = st.sidebar.slider( '''Minimum generation length''', min_value=8, max_value=256, value=64, step=8, format=None, key=None ) lowerCAmelCase__ = st.sidebar.slider( '''Maximum generation length''', min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": lowerCAmelCase__ = st.sidebar.slider('''Beam size''', min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: lowerCAmelCase__ = st.sidebar.slider( '''Nucleus sampling p''', min_value=0.1, max_value=1.0, value=0.95, step=0.01, format=None, key=None ) lowerCAmelCase__ = st.sidebar.slider( '''Temperature''', min_value=0.1, max_value=1.0, value=0.7, step=0.01, format=None, key=None ) lowerCAmelCase__ = None # start main text lowerCAmelCase__ = [ '''<MY QUESTION>''', '''How do people make chocolate?''', '''Why do we get a fever when we are sick?''', '''How can different animals perceive different colors?''', '''What is natural language processing?''', '''What\'s the best way to treat a sunburn?''', '''What exactly are vitamins ?''', '''How does nuclear energy provide electricity?''', '''What\'s the difference between viruses and bacteria?''', '''Why are flutes classified as woodwinds when most of them are made out of metal ?''', '''Why do people like drinking coffee even though it tastes so bad?''', '''What happens when wine ages? How does it make the wine taste better?''', '''If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?''', '''How can we set a date to the beginning or end of an artistic period? Doesn\'t the change happen gradually?''', '''How does New Zealand have so many large bird predators?''', ] lowerCAmelCase__ = st.selectbox( '''What would you like to ask? ---- select <MY QUESTION> to enter a new query''', questions_list, index=1, ) if question_s == "<MY QUESTION>": lowerCAmelCase__ = st.text_input('''Enter your question here:''', '''''') else: lowerCAmelCase__ = question_s if st.button('''Show me!'''): if action in [0, 1, 3]: if index_type == "mixed": lowerCAmelCase__ , lowerCAmelCase__ = make_support(question, source=wiki_source, method='''dense''', n_results=10) lowerCAmelCase__ , lowerCAmelCase__ = make_support(question, source=wiki_source, method='''sparse''', n_results=10) lowerCAmelCase__ = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] lowerCAmelCase__ = support_list[:10] lowerCAmelCase__ = '''<P> ''' + ''' <P> '''.join([res[-1] for res in support_list]) else: lowerCAmelCase__ , lowerCAmelCase__ = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: lowerCAmelCase__ , lowerCAmelCase__ = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == '''sampled'''), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown('''### The model generated answer is:''') st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown('''--- \n ### The model is drawing information from the following Wikipedia passages:''') for i, res in enumerate(support_list): lowerCAmelCase__ = '''https://en.wikipedia.org/wiki/{}'''.format(res[0].replace(''' ''', '''_''')) lowerCAmelCase__ = res[1].strip() if sec_titles == "": lowerCAmelCase__ = '''[{}]({})'''.format(res[0], wiki_url) else: lowerCAmelCase__ = sec_titles.split(''' & ''') lowerCAmelCase__ = ''' & '''.join( ['''[{}]({}#{})'''.format(sec.strip(), wiki_url, sec.strip().replace(''' ''', '''_''')) for sec in sec_list] ) st.markdown( '''{0:02d} - **Article**: {1:<18} <br> _Section_: {2}'''.format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( '''> <span style=\"font-family:arial; font-size:10pt;\">''' + res[-1] + '''</span>''', unsafe_allow_html=True ) if action in [2, 3]: lowerCAmelCase__ = find_nearest_training(question) lowerCAmelCase__ = nn_train_list[0] st.markdown( '''--- \n ### The most similar question in the ELI5 training set was: \n\n {}'''.format(train_exple['''title''']) ) lowerCAmelCase__ = [ '''{}. {}'''.format(i + 1, ''' \n'''.join([line.strip() for line in ans.split('''\n''') if line.strip() != ''''''])) for i, (ans, sc) in enumerate(zip(train_exple['''answers''']['''text'''], train_exple['''answers''']['''score'''])) if i == 0 or sc > 2 ] st.markdown('''##### Its answers were: \n\n {}'''.format('''\n'''.join(answers_st))) lowerCAmelCase__ = ''' --- **Disclaimer** *The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system. Evaluating biases of such a model and ensuring factual generations are still very much open research problems. Therefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.* ''' st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : Dict ,lowercase__ : Any=1_3 ,lowercase__ : Any=7 ,lowercase__ : Tuple=6 ,lowercase__ : int=1_7 ,lowercase__ : List[str]=2_3 ,lowercase__ : Optional[Any]=1_1 ,lowercase__ : List[Any]=True ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = act_dim __lowercase = state_dim __lowercase = hidden_size __lowercase = max_length __lowercase = is_training def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) __lowercase = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) __lowercase = floats_tensor((self.batch_size, self.seq_length, 1) ) __lowercase = floats_tensor((self.batch_size, self.seq_length, 1) ) __lowercase = ids_tensor((self.batch_size, self.seq_length) ,vocab_size=1_0_0_0 ) __lowercase = random_attention_mask((self.batch_size, self.seq_length) ) __lowercase = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def SCREAMING_SNAKE_CASE ( self : Dict ): return DecisionTransformerConfig( batch_size=self.batch_size ,seq_length=self.seq_length ,act_dim=self.act_dim ,state_dim=self.state_dim ,hidden_size=self.hidden_size ,max_length=self.max_length ,) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Any ,lowercase__ : Dict ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,lowercase__ : List[Any] ,lowercase__ : Dict ,): __lowercase = DecisionTransformerModel(config=UpperCamelCase__ ) model.to(UpperCamelCase__ ) model.eval() __lowercase = model(UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ,UpperCamelCase__ ) self.parent.assertEqual(result.state_preds.shape ,states.shape ) self.parent.assertEqual(result.action_preds.shape ,actions.shape ) self.parent.assertEqual(result.return_preds.shape ,returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.prepare_config_and_inputs() ( __lowercase ) = config_and_inputs __lowercase = { '''states''': states, '''actions''': actions, '''rewards''': rewards, '''returns_to_go''': returns_to_go, '''timesteps''': timesteps, '''attention_mask''': attention_mask, } return config, inputs_dict @require_torch class lowercase_ (UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = (DecisionTransformerModel,) if is_torch_available() else () SCREAMING_SNAKE_CASE : Tuple = () SCREAMING_SNAKE_CASE : Union[str, Any] = {'feature-extraction': DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids SCREAMING_SNAKE_CASE : str = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = DecisionTransformerModelTester(self ) __lowercase = ConfigTester(self ,config_class=UpperCamelCase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : int ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCamelCase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = DecisionTransformerModel.from_pretrained(UpperCamelCase__ ) self.assertIsNotNone(UpperCamelCase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(UpperCamelCase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = [ '''states''', '''actions''', '''rewards''', '''returns_to_go''', '''timesteps''', '''attention_mask''', ] self.assertListEqual(arg_names[: len(UpperCamelCase__ )] ,UpperCamelCase__ ) @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" @slow def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = 2 # number of steps of autoregressive prediction we will perform __lowercase = 1_0 # defined by the RL environment, may be normalized __lowercase = DecisionTransformerModel.from_pretrained('''edbeeching/decision-transformer-gym-hopper-expert''' ) __lowercase = model.to(UpperCamelCase__ ) __lowercase = model.config torch.manual_seed(0 ) __lowercase = torch.randn(1 ,1 ,config.state_dim ).to(device=UpperCamelCase__ ,dtype=torch.floataa ) # env.reset() __lowercase = torch.tensor( [[0.2_4_2_7_9_3, -0.2_8_6_9_3_0_7_4, 0.8_7_4_2_6_1_3], [0.6_7_8_1_5_2_7_4, -0.0_8_1_0_1_0_8_5, -0.1_2_9_5_2_1_4_7]] ,device=UpperCamelCase__ ) __lowercase = torch.tensor(UpperCamelCase__ ,device=UpperCamelCase__ ,dtype=torch.floataa ).reshape(1 ,1 ,1 ) __lowercase = state __lowercase = torch.zeros(1 ,0 ,config.act_dim ,device=UpperCamelCase__ ,dtype=torch.floataa ) __lowercase = torch.zeros(1 ,0 ,device=UpperCamelCase__ ,dtype=torch.floataa ) __lowercase = torch.tensor(0 ,device=UpperCamelCase__ ,dtype=torch.long ).reshape(1 ,1 ) for step in range(UpperCamelCase__ ): __lowercase = torch.cat([actions, torch.zeros(1 ,1 ,config.act_dim ,device=UpperCamelCase__ )] ,dim=1 ) __lowercase = torch.cat([rewards, torch.zeros(1 ,1 ,device=UpperCamelCase__ )] ,dim=1 ) __lowercase = torch.ones(1 ,states.shape[1] ).to(dtype=torch.long ,device=states.device ) with torch.no_grad(): __lowercase = model( states=UpperCamelCase__ ,actions=UpperCamelCase__ ,rewards=UpperCamelCase__ ,returns_to_go=UpperCamelCase__ ,timesteps=UpperCamelCase__ ,attention_mask=UpperCamelCase__ ,return_dict=UpperCamelCase__ ,) self.assertEqual(action_pred.shape ,actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] ,expected_outputs[step] ,atol=1e-4 ) ) __lowercase = ( # env.step(action) torch.randn(1 ,1 ,config.state_dim ).to(device=UpperCamelCase__ ,dtype=torch.floataa ), 1.0, False, {}, ) __lowercase = action_pred[0, -1] __lowercase = torch.cat([states, state] ,dim=1 ) __lowercase = returns_to_go[0, -1] - reward __lowercase = torch.cat([returns_to_go, pred_return.reshape(1 ,1 ,1 )] ,dim=1 ) __lowercase = torch.cat( [timesteps, torch.ones((1, 1) ,device=UpperCamelCase__ ,dtype=torch.long ) * (step + 1)] ,dim=1 )
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import torch from diffusers import CMStochasticIterativeScheduler from .test_schedulers import SchedulerCommonTest class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = (CMStochasticIterativeScheduler,) SCREAMING_SNAKE_CASE : int = 1_0 def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,**lowercase__ : str ): __lowercase = { '''num_train_timesteps''': 2_0_1, '''sigma_min''': 0.0_0_2, '''sigma_max''': 8_0.0, } config.update(**_SCREAMING_SNAKE_CASE ) return config def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = 1_0 __lowercase = self.get_scheduler_config() __lowercase = self.scheduler_classes[0](**_SCREAMING_SNAKE_CASE ) scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) __lowercase = scheduler.timesteps[0] __lowercase = scheduler.timesteps[1] __lowercase = self.dummy_sample __lowercase = 0.1 * sample __lowercase = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).prev_sample __lowercase = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ).prev_sample self.assertEqual(output_a.shape ,sample.shape ) self.assertEqual(output_a.shape ,output_a.shape ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for timesteps in [1_0, 5_0, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self : List[str] ): for clip_denoised in [True, False]: self.check_over_configs(clip_denoised=_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**_SCREAMING_SNAKE_CASE ) __lowercase = 1 scheduler.set_timesteps(_SCREAMING_SNAKE_CASE ) __lowercase = scheduler.timesteps __lowercase = torch.manual_seed(0 ) __lowercase = self.dummy_model() __lowercase = self.dummy_sample_deter * scheduler.init_noise_sigma for i, t in enumerate(_SCREAMING_SNAKE_CASE ): # 1. scale model input __lowercase = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # 2. predict noise residual __lowercase = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # 3. predict previous sample x_t-1 __lowercase = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,generator=_SCREAMING_SNAKE_CASE ).prev_sample __lowercase = pred_prev_sample __lowercase = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) __lowercase = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 1_9_2.7_6_1_4 ) < 1e-2 assert abs(result_mean.item() - 0.2_5_1_0 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**_SCREAMING_SNAKE_CASE ) __lowercase = [1_0_6, 0] scheduler.set_timesteps(timesteps=_SCREAMING_SNAKE_CASE ) __lowercase = scheduler.timesteps __lowercase = torch.manual_seed(0 ) __lowercase = self.dummy_model() __lowercase = self.dummy_sample_deter * scheduler.init_noise_sigma for t in timesteps: # 1. scale model input __lowercase = scheduler.scale_model_input(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # 2. predict noise residual __lowercase = model(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ) # 3. predict previous sample x_t-1 __lowercase = scheduler.step(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ,generator=_SCREAMING_SNAKE_CASE ).prev_sample __lowercase = pred_prev_sample __lowercase = torch.sum(torch.abs(_SCREAMING_SNAKE_CASE ) ) __lowercase = torch.mean(torch.abs(_SCREAMING_SNAKE_CASE ) ) assert abs(result_sum.item() - 3_4_7.6_3_5_7 ) < 1e-2 assert abs(result_mean.item() - 0.4_5_2_7 ) < 1e-3 def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**_SCREAMING_SNAKE_CASE ) __lowercase = [3_9, 3_0, 1_2, 1_5, 0] with self.assertRaises(_SCREAMING_SNAKE_CASE ,msg='''`timesteps` must be in descending order.''' ): scheduler.set_timesteps(timesteps=_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**_SCREAMING_SNAKE_CASE ) __lowercase = [3_9, 3_0, 1_2, 1, 0] __lowercase = len(_SCREAMING_SNAKE_CASE ) with self.assertRaises(_SCREAMING_SNAKE_CASE ,msg='''Can only pass one of `num_inference_steps` or `timesteps`.''' ): scheduler.set_timesteps(num_inference_steps=_SCREAMING_SNAKE_CASE ,timesteps=_SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.scheduler_classes[0] __lowercase = self.get_scheduler_config() __lowercase = scheduler_class(**_SCREAMING_SNAKE_CASE ) __lowercase = [scheduler.config.num_train_timesteps] with self.assertRaises( _SCREAMING_SNAKE_CASE ,msg='''`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}''' ,): scheduler.set_timesteps(timesteps=_SCREAMING_SNAKE_CASE )
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' def _A ( A__ = 2000000 ): """simple docstring""" __lowercase = [0 for i in range(n + 1 )] __lowercase = 1 __lowercase = 1 for i in range(2 , int(n**0.5 ) + 1 ): if primality_list[i] == 0: for j in range(i * i , n + 1 , A__ ): __lowercase = 1 __lowercase = 0 for i in range(A__ ): if primality_list[i] == 0: sum_of_primes += i return sum_of_primes if __name__ == "__main__": print(f'{solution() = }')
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' from __future__ import annotations from collections import namedtuple def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = namedtuple('''result''' , '''name value''' ) if (voltage, current, power).count(0 ) != 1: raise ValueError('''Only one argument must be 0''' ) elif power < 0: raise ValueError( '''Power cannot be negative in any electrical/electronics system''' ) elif voltage == 0: return result('''voltage''' , power / current ) elif current == 0: return result('''current''' , power / voltage ) elif power == 0: return result('''power''' , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError('''Exactly one argument must be 0''' ) if __name__ == "__main__": import doctest doctest.testmod()
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : List[str] ,lowercase__ : Optional[Any] ): __lowercase = num_of_nodes __lowercase = [] __lowercase = {} def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : int ,lowercase__ : Optional[Any] ): self.m_edges.append([u_node, v_node, weight] ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[Any] ): if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ): if self.m_component[u_node] != u_node: for k in self.m_component: __lowercase = self.find_component(A_ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): if component_size[u_node] <= component_size[v_node]: __lowercase = v_node component_size[v_node] += component_size[u_node] self.set_component(A_ ) elif component_size[u_node] >= component_size[v_node]: __lowercase = self.find_component(A_ ) component_size[u_node] += component_size[v_node] self.set_component(A_ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = [] __lowercase = 0 __lowercase = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) __lowercase = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: __lowercase = edge __lowercase = self.m_component[u] __lowercase = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): __lowercase = [u, v, w] for edge in minimum_weight_edge: if isinstance(A_ ,A_ ): __lowercase = edge __lowercase = self.m_component[u] __lowercase = self.m_component[v] if u_component != v_component: mst_weight += w self.union(A_ ,A_ ,A_ ) print(F"Added edge [{u} - {v}]\nAdded weight: {w}\n" ) num_of_components -= 1 __lowercase = [-1] * self.m_num_of_nodes print(F"The total weight of the minimal spanning tree is: {mst_weight}" ) def _A ( ): """simple docstring""" pass if __name__ == "__main__": import doctest doctest.testmod()
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from math import factorial def _A ( A__ , A__ ): """simple docstring""" if n < k or k < 0: raise ValueError('''Please enter positive integers for n and k where n >= k''' ) return factorial(A__ ) // (factorial(A__ ) * factorial(n - k )) if __name__ == "__main__": print( '''The number of five-card hands possible from a standard''', f'fifty-two card deck is: {combinations(52, 5)}\n', ) print( '''If a class of 40 students must be arranged into groups of''', f'4 for group projects, there are {combinations(40, 4)} ways', '''to arrange them.\n''', ) print( '''If 10 teams are competing in a Formula One race, there''', f'are {combinations(10, 3)} ways that first, second and', '''third place can be awarded.''', )
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__ = { '''configuration_xlm_roberta''': [ '''XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''XLMRobertaConfig''', '''XLMRobertaOnnxConfig''', ], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''XLMRobertaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''XLMRobertaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''XLMRobertaForCausalLM''', '''XLMRobertaForMaskedLM''', '''XLMRobertaForMultipleChoice''', '''XLMRobertaForQuestionAnswering''', '''XLMRobertaForSequenceClassification''', '''XLMRobertaForTokenClassification''', '''XLMRobertaModel''', '''XLMRobertaPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFXLMRobertaForCausalLM''', '''TFXLMRobertaForMaskedLM''', '''TFXLMRobertaForMultipleChoice''', '''TFXLMRobertaForQuestionAnswering''', '''TFXLMRobertaForSequenceClassification''', '''TFXLMRobertaForTokenClassification''', '''TFXLMRobertaModel''', '''TFXLMRobertaPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FlaxXLMRobertaForMaskedLM''', '''FlaxXLMRobertaForCausalLM''', '''FlaxXLMRobertaForMultipleChoice''', '''FlaxXLMRobertaForQuestionAnswering''', '''FlaxXLMRobertaForSequenceClassification''', '''FlaxXLMRobertaForTokenClassification''', '''FlaxXLMRobertaModel''', '''FlaxXLMRobertaPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, XLMRobertaConfig, XLMRobertaOnnxConfig, ) try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta import XLMRobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlm_roberta_fast import XLMRobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlm_roberta import ( XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, XLMRobertaForCausalLM, XLMRobertaForMaskedLM, XLMRobertaForMultipleChoice, XLMRobertaForQuestionAnswering, XLMRobertaForSequenceClassification, XLMRobertaForTokenClassification, XLMRobertaModel, XLMRobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlm_roberta import ( TF_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLMRobertaForCausalLM, TFXLMRobertaForMaskedLM, TFXLMRobertaForMultipleChoice, TFXLMRobertaForQuestionAnswering, TFXLMRobertaForSequenceClassification, TFXLMRobertaForTokenClassification, TFXLMRobertaModel, TFXLMRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_xlm_roberta import ( FLAX_XLM_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, FlaxXLMRobertaForCausalLM, FlaxXLMRobertaForMaskedLM, FlaxXLMRobertaForMultipleChoice, FlaxXLMRobertaForQuestionAnswering, FlaxXLMRobertaForSequenceClassification, FlaxXLMRobertaForTokenClassification, FlaxXLMRobertaModel, FlaxXLMRobertaPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = '''https://openaipublic.azureedge.net/jukebox/models/''' lowerCAmelCase__ = { '''jukebox-1b-lyrics''': [ '''5b/vqvae.pth.tar''', '''5b/prior_level_0.pth.tar''', '''5b/prior_level_1.pth.tar''', '''1b_lyrics/prior_level_2.pth.tar''', ], '''jukebox-5b-lyrics''': [ '''5b/vqvae.pth.tar''', '''5b/prior_level_0.pth.tar''', '''5b/prior_level_1.pth.tar''', '''5b_lyrics/prior_level_2.pth.tar''', ], } def _A ( A__ ): """simple docstring""" if key.endswith('''.model.1.bias''' ) and len(key.split('''.''' ) ) > 10: __lowercase = key.replace('''.model.1.bias''' , '''.conv1d_1.bias''' ) elif key.endswith('''.model.1.weight''' ) and len(key.split('''.''' ) ) > 10: __lowercase = key.replace('''.model.1.weight''' , '''.conv1d_1.weight''' ) elif key.endswith('''.model.3.bias''' ) and len(key.split('''.''' ) ) > 10: __lowercase = key.replace('''.model.3.bias''' , '''.conv1d_2.bias''' ) elif key.endswith('''.model.3.weight''' ) and len(key.split('''.''' ) ) > 10: __lowercase = key.replace('''.model.3.weight''' , '''.conv1d_2.weight''' ) if "conditioner_blocks.0." in key: __lowercase = key.replace('''conditioner_blocks.0''' , '''conditioner_blocks''' ) if "prime_prior" in key: __lowercase = key.replace('''prime_prior''' , '''encoder''' ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: __lowercase = key.replace('''.emb.''' , '''.''' ) if key.endswith('''k''' ): # replace vqvae.X.k with vqvae.X.codebook return key.replace('''.k''' , '''.codebook''' ) if "y_emb." in key: return key.replace('''y_emb.''' , '''metadata_embedding.''' ) if "x_emb.emb." in key: __lowercase = key.replace('''0.x_emb.emb''' , '''embed_tokens''' ) if "prime_state_ln" in key: return key.replace('''prime_state_ln''' , '''encoder.final_layer_norm''' ) if ".ln" in key: return key.replace('''.ln''' , '''.layer_norm''' ) if "_ln" in key: return key.replace('''_ln''' , '''_layer_norm''' ) if "prime_state_proj" in key: return key.replace('''prime_state_proj''' , '''encoder.proj_in''' ) if "prime_x_out" in key: return key.replace('''prime_x_out''' , '''encoder.lm_head''' ) if "prior.x_out" in key: return key.replace('''x_out''' , '''fc_proj_out''' ) if "x_emb" in key: return key.replace('''x_emb''' , '''embed_tokens''' ) return key def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = {} import re __lowercase = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) __lowercase = re.compile( R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowercase = re.compile(R'''encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) __lowercase = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)''' ) __lowercase = re.compile( R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowercase = re.compile(R'''decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)''' ) __lowercase = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)''' ) __lowercase = re.compile( R'''conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)''' ) __lowercase = re.compile(R'''conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)''' ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(lowercase__ ): __lowercase = re_encoder_block_conv_in.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[2] ) * 2 + int(groups[3] ) __lowercase = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}" __lowercase = re_encoder_block_conv_in.sub(lowercase__ , lowercase__ ) elif re_encoder_block_resnet.fullmatch(lowercase__ ): __lowercase = re_encoder_block_resnet.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[2] ) * 2 + int(groups[3] ) __lowercase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowercase = F"encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}." __lowercase = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowercase = prefix + resnet_block __lowercase = re_encoder_block_resnet.sub(lowercase__ , lowercase__ ) elif re_encoder_block_proj_out.fullmatch(lowercase__ ): __lowercase = re_encoder_block_proj_out.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = F"encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}" __lowercase = re_encoder_block_proj_out.sub(lowercase__ , lowercase__ ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(lowercase__ ): __lowercase = re_decoder_block_conv_out.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowercase = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}" __lowercase = re_decoder_block_conv_out.sub(lowercase__ , lowercase__ ) elif re_decoder_block_resnet.fullmatch(lowercase__ ): __lowercase = re_decoder_block_resnet.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[2] ) * 2 + int(groups[3] ) - 2 __lowercase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowercase = F"decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}." __lowercase = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowercase = prefix + resnet_block __lowercase = re_decoder_block_resnet.sub(lowercase__ , lowercase__ ) elif re_decoder_block_proj_in.fullmatch(lowercase__ ): __lowercase = re_decoder_block_proj_in.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = F"decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}" __lowercase = re_decoder_block_proj_in.sub(lowercase__ , lowercase__ ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(lowercase__ ): __lowercase = re_prior_cond_conv_out.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowercase = F"conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}" __lowercase = re_prior_cond_conv_out.sub(lowercase__ , lowercase__ ) elif re_prior_cond_resnet.fullmatch(lowercase__ ): __lowercase = re_prior_cond_resnet.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = int(groups[1] ) * 2 + int(groups[2] ) - 2 __lowercase = {'''1''': 1, '''3''': 2}[groups[-2]] __lowercase = F"conditioner_blocks.upsampler.upsample_block.{block_index}." __lowercase = F"resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}" __lowercase = prefix + resnet_block __lowercase = re_prior_cond_resnet.sub(lowercase__ , lowercase__ ) elif re_prior_cond_proj_in.fullmatch(lowercase__ ): __lowercase = re_prior_cond_proj_in.match(lowercase__ ) __lowercase = regex_match.groups() __lowercase = F"conditioner_blocks.upsampler.proj_in.{groups[-1]}" __lowercase = re_prior_cond_proj_in.sub(lowercase__ , lowercase__ ) # keep original key else: __lowercase = original_key __lowercase = replace_key(lowercase__ ) if F"{key_prefix}.{key}" not in model_state_dict or key is None: print(F"failed converting {original_key} to {key}, does not match" ) # handle missmatched shape elif value.shape != model_state_dict[F"{key_prefix}.{key}"].shape: __lowercase = model_state_dict[F"{key_prefix}.{key}"] print(F"{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match" ) __lowercase = original_key __lowercase = original_key __lowercase = value return new_dict @torch.no_grad() def _A ( A__=None , A__=None ): """simple docstring""" for file in MODEL_MAPPING[model_name]: if not os.path.isfile(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" ): __lowercase = requests.get(F"{PREFIX}{file}" , allow_redirects=lowercase__ ) os.makedirs(F"{pytorch_dump_folder_path}/" , exist_ok=lowercase__ ) open(F"{pytorch_dump_folder_path}/{file.split('/' )[-1]}" , '''wb''' ).write(r.content ) __lowercase = MODEL_MAPPING[model_name.split('''/''' )[-1]] __lowercase = JukeboxConfig.from_pretrained(lowercase__ ) __lowercase = JukeboxModel(lowercase__ ) __lowercase = [] __lowercase = {} for i, dict_name in enumerate(lowercase__ ): __lowercase = torch.load(F"{pytorch_dump_folder_path}/{dict_name.split('/' )[-1]}" )['''model'''] __lowercase = {} for k in old_dic.keys(): if k.endswith('''.b''' ): __lowercase = old_dic[k] elif k.endswith('''.w''' ): __lowercase = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: __lowercase = old_dic[k] else: __lowercase = old_dic[k] __lowercase = '''vqvae''' if i == 0 else F"priors.{3 - i}" __lowercase = fix_jukebox_keys(lowercase__ , model.state_dict() , lowercase__ , lowercase__ ) weight_dict.append(lowercase__ ) __lowercase = weight_dict.pop(0 ) model.vqvae.load_state_dict(lowercase__ ) for i in range(len(lowercase__ ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(lowercase__ ).mkdir(exist_ok=lowercase__ ) with open(F"{pytorch_dump_folder_path}/mapping.json" , '''w''' ) as txtfile: json.dump(lowercase__ , lowercase__ ) print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(lowercase__ ) return weight_dict if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''jukebox-5b-lyrics''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''jukebox-5b-lyrics-converted''', type=str, help='''Path to the output PyTorch model directory.''', ) lowerCAmelCase__ = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' import unittest from pathlib import Path from tempfile import NamedTemporaryFile, TemporaryDirectory from transformers import BertConfig, BertTokenizerFast, FeatureExtractionPipeline from transformers.convert_graph_to_onnx import ( convert, ensure_valid_input, generate_identified_filename, infer_shapes, quantize, ) from transformers.testing_utils import require_tf, require_tokenizers, require_torch, slow class lowercase_ : """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : Tuple ,lowercase__ : Optional[int] ): return None class lowercase_ : """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Tuple ,lowercase__ : str ,lowercase__ : str ,lowercase__ : str ): return None class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = [ # (model_name, model_kwargs) ('''bert-base-cased''', {}), ('''gpt2''', {'''use_cache''': False}), # We don't support exporting GPT2 past keys anymore ] @require_tf @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__a ,'''tf''' ,1_2 ,**__a ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self : str ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: self._test_export(__a ,'''pt''' ,1_2 ,**__a ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): from transformers import BertModel __lowercase = ["""[UNK]""", """[SEP]""", """[CLS]""", """[PAD]""", """[MASK]""", """some""", """other""", """words"""] with NamedTemporaryFile(mode='''w+t''' ) as vocab_file: vocab_file.write('''\n'''.join(__a ) ) vocab_file.flush() __lowercase = BertTokenizerFast(vocab_file.name ) with TemporaryDirectory() as bert_save_dir: __lowercase = BertModel(BertConfig(vocab_size=len(__a ) ) ) model.save_pretrained(__a ) self._test_export(__a ,'''pt''' ,1_2 ,__a ) @require_tf @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: __lowercase = self._test_export(__a ,'''tf''' ,1_2 ,**__a ) __lowercase = quantize(Path(__a ) ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__a ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) @require_torch @slow def SCREAMING_SNAKE_CASE ( self : Any ): for model, model_kwargs in OnnxExportTestCase.MODEL_TO_TEST: __lowercase = self._test_export(__a ,'''pt''' ,1_2 ,**__a ) __lowercase = quantize(__a ) # Ensure the actual quantized model is not bigger than the original one if quantized_path.stat().st_size >= Path(__a ).stat().st_size: self.fail('''Quantized model is bigger than initial ONNX model''' ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any]=None ,**lowercase__ : Optional[int] ): try: # Compute path with TemporaryDirectory() as tempdir: __lowercase = Path(__a ).joinpath('''model.onnx''' ) # Remove folder if exists if path.parent.exists(): path.parent.rmdir() # Export convert(__a ,__a ,__a ,__a ,__a ,**__a ) return path except Exception as e: self.fail(__a ) @require_torch @require_tokenizers @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): from transformers import BertModel __lowercase = BertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) __lowercase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(__a ,__a ,'''pt''' ) @require_tf @require_tokenizers @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): from transformers import TFBertModel __lowercase = TFBertModel(BertConfig.from_pretrained('''lysandre/tiny-bert-random''' ) ) __lowercase = BertTokenizerFast.from_pretrained('''lysandre/tiny-bert-random''' ) self._test_infer_dynamic_axis(__a ,__a ,'''tf''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : str ,lowercase__ : Optional[Any] ,lowercase__ : str ): __lowercase = FeatureExtractionPipeline(__a ,__a ) __lowercase = ["""input_ids""", """token_type_ids""", """attention_mask""", """output_0""", """output_1"""] __lowercase = infer_shapes(__a ,__a ) # Assert all variables are present self.assertEqual(len(__a ) ,len(__a ) ) self.assertTrue(all(var_name in shapes for var_name in variable_names ) ) self.assertSequenceEqual(variable_names[:3] ,__a ) self.assertSequenceEqual(variable_names[3:] ,__a ) # Assert inputs are {0: batch, 1: sequence} for var_name in ["input_ids", "token_type_ids", "attention_mask"]: self.assertDictEqual(shapes[var_name] ,{0: '''batch''', 1: '''sequence'''} ) # Assert outputs are {0: batch, 1: sequence} and {0: batch} self.assertDictEqual(shapes['''output_0'''] ,{0: '''batch''', 1: '''sequence'''} ) self.assertDictEqual(shapes['''output_1'''] ,{0: '''batch'''} ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ["""input_ids""", """attention_mask""", """token_type_ids"""] __lowercase = {"""input_ids""": [1, 2, 3, 4], """attention_mask""": [0, 0, 0, 0], """token_type_ids""": [1, 1, 1, 1]} __lowercase = ensure_valid_input(FuncContiguousArgs() ,__a ,__a ) # Should have exactly the same number of args (all are valid) self.assertEqual(len(__a ) ,3 ) # Should have exactly the same input names self.assertEqual(set(__a ) ,set(__a ) ) # Parameter should be reordered according to their respective place in the function: # (input_ids, token_type_ids, attention_mask) self.assertEqual(__a ,(tokens['''input_ids'''], tokens['''token_type_ids'''], tokens['''attention_mask''']) ) # Generated args are interleaved with another args (for instance parameter "past" in GPT2) __lowercase = ensure_valid_input(FuncNonContiguousArgs() ,__a ,__a ) # Should have exactly the one arg (all before the one not provided "some_other_args") self.assertEqual(len(__a ) ,1 ) self.assertEqual(len(__a ) ,1 ) # Should have only "input_ids" self.assertEqual(inputs_args[0] ,tokens['''input_ids'''] ) self.assertEqual(ordered_input_names[0] ,'''input_ids''' ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = generate_identified_filename(Path('''/home/something/my_fake_model.onnx''' ) ,'''-test''' ) self.assertEqual('''/home/something/my_fake_model-test.onnx''' ,generated.as_posix() )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name def _A ( A__ ): """simple docstring""" if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(_UpperCamelCase ): return ext raise Exception( F"Unable to determine file format from file extension {path}. " F"Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}" ) def _A ( A__ ): """simple docstring""" __lowercase = pipeline( task=args.task , model=args.model if args.model else None , config=args.config , tokenizer=args.tokenizer , device=args.device , ) __lowercase = try_infer_format_from_ext(args.input ) if args.format == '''infer''' else args.format __lowercase = PipelineDataFormat.from_str( format=_UpperCamelCase , output_path=args.output , input_path=args.input , column=args.column if args.column else nlp.default_input_names , overwrite=args.overwrite , ) return RunCommand(_UpperCamelCase , _UpperCamelCase ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Any ,lowercase__ : Pipeline ,lowercase__ : PipelineDataFormat ): __lowercase = nlp __lowercase = reader @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''run''' ,help='''Run a pipeline through the CLI''' ) run_parser.add_argument('''--task''' ,choices=get_supported_tasks() ,help='''Task to run''' ) run_parser.add_argument('''--input''' ,type=UpperCAmelCase__ ,help='''Path to the file to use for inference''' ) run_parser.add_argument('''--output''' ,type=UpperCAmelCase__ ,help='''Path to the file that will be used post to write results.''' ) run_parser.add_argument('''--model''' ,type=UpperCAmelCase__ ,help='''Name or path to the model to instantiate.''' ) run_parser.add_argument('''--config''' ,type=UpperCAmelCase__ ,help='''Name or path to the model\'s config to instantiate.''' ) run_parser.add_argument( '''--tokenizer''' ,type=UpperCAmelCase__ ,help='''Name of the tokenizer to use. (default: same as the model name)''' ) run_parser.add_argument( '''--column''' ,type=UpperCAmelCase__ ,help='''Name of the column to use as input. (For multi columns input as QA use column1,columns2)''' ,) run_parser.add_argument( '''--format''' ,type=UpperCAmelCase__ ,default='''infer''' ,choices=PipelineDataFormat.SUPPORTED_FORMATS ,help='''Input format to read from''' ,) run_parser.add_argument( '''--device''' ,type=UpperCAmelCase__ ,default=-1 ,help='''Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)''' ,) run_parser.add_argument('''--overwrite''' ,action='''store_true''' ,help='''Allow overwriting the output file.''' ) run_parser.set_defaults(func=UpperCAmelCase__ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self._nlp, [] for entry in self._reader: __lowercase = nlp(**UpperCAmelCase__ ) if self._reader.is_multi_columns else nlp(UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ ,UpperCAmelCase__ ): outputs.append(UpperCAmelCase__ ) else: outputs += output # Saving data if self._nlp.binary_output: __lowercase = self._reader.save_binary(UpperCAmelCase__ ) logger.warning(F"Current pipeline requires output to be in binary format, saving at {binary_path}" ) else: self._reader.save(UpperCAmelCase__ )
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' import argparse import torch from transformers import BertForMaskedLM if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser( description=( '''Extraction some layers of the full BertForMaskedLM or RObertaForMaskedLM for Transfer Learned''' ''' Distillation''' ) ) parser.add_argument('''--model_type''', default='''bert''', choices=['''bert''']) parser.add_argument('''--model_name''', default='''bert-base-uncased''', type=str) parser.add_argument('''--dump_checkpoint''', default='''serialization_dir/tf_bert-base-uncased_0247911.pth''', type=str) parser.add_argument('''--vocab_transform''', action='''store_true''') lowerCAmelCase__ = parser.parse_args() if args.model_type == "bert": lowerCAmelCase__ = BertForMaskedLM.from_pretrained(args.model_name) lowerCAmelCase__ = '''bert''' else: raise ValueError('''args.model_type should be "bert".''') lowerCAmelCase__ = model.state_dict() lowerCAmelCase__ = {} for w in ["word_embeddings", "position_embeddings"]: lowerCAmelCase__ = state_dict[f'{prefix}.embeddings.{w}.weight'] for w in ["weight", "bias"]: lowerCAmelCase__ = state_dict[f'{prefix}.embeddings.LayerNorm.{w}'] lowerCAmelCase__ = 0 for teacher_idx in [0, 2, 4, 7, 9, 11]: for w in ["weight", "bias"]: lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.attention.self.query.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.attention.self.key.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.attention.self.value.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.attention.output.dense.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.attention.output.LayerNorm.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.intermediate.dense.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.output.dense.{w}' ] lowerCAmelCase__ = state_dict[ f'{prefix}.encoder.layer.{teacher_idx}.output.LayerNorm.{w}' ] std_idx += 1 lowerCAmelCase__ = state_dict['''cls.predictions.decoder.weight'''] lowerCAmelCase__ = state_dict['''cls.predictions.bias'''] if args.vocab_transform: for w in ["weight", "bias"]: lowerCAmelCase__ = state_dict[f'cls.predictions.transform.dense.{w}'] lowerCAmelCase__ = state_dict[f'cls.predictions.transform.LayerNorm.{w}'] print(f'N layers selected for distillation: {std_idx}') print(f'Number of params transferred for distillation: {len(compressed_sd.keys())}') print(f'Save transferred checkpoint to {args.dump_checkpoint}.') torch.save(compressed_sd, args.dump_checkpoint)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : Any ,lowercase__ : Any ,lowercase__ : Dict=7 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=3_0 ,lowercase__ : List[str]=4_0_0 ,lowercase__ : Union[str, Any]=True ,lowercase__ : List[Any]=None ,lowercase__ : int=True ,lowercase__ : Any=[0.5, 0.5, 0.5] ,lowercase__ : str=[0.5, 0.5, 0.5] ,lowercase__ : Dict=True ,lowercase__ : Optional[int]=1 / 2_5_5 ,lowercase__ : str=True ,): __lowercase = size if size is not None else {'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = do_normalize __lowercase = image_mean __lowercase = image_std __lowercase = do_rescale __lowercase = rescale_factor __lowercase = do_pad def SCREAMING_SNAKE_CASE ( self : str ): return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : int ,lowercase__ : Tuple=False ): if not batched: __lowercase = image_inputs[0] if isinstance(lowercase__ ,Image.Image ): __lowercase , __lowercase = image.size else: __lowercase , __lowercase = image.shape[1], image.shape[2] if w < h: __lowercase = int(self.size['''shortest_edge'''] * h / w ) __lowercase = self.size['''shortest_edge'''] elif w > h: __lowercase = self.size['''shortest_edge'''] __lowercase = int(self.size['''shortest_edge'''] * w / h ) else: __lowercase = self.size['''shortest_edge'''] __lowercase = self.size['''shortest_edge'''] else: __lowercase = [] for image in image_inputs: __lowercase , __lowercase = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) __lowercase = max(lowercase__ ,key=lambda lowercase__ : item[0] )[0] __lowercase = max(lowercase__ ,key=lambda lowercase__ : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = DeformableDetrImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = DeformableDetrImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : Any ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''image_mean''' ) ) self.assertTrue(hasattr(lowercase__ ,'''image_std''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_normalize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_rescale''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_pad''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size''' ) ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 1_8, '''longest_edge''': 1_3_3_3} ) self.assertEqual(image_processor.do_pad ,lowercase__ ) __lowercase = self.image_processing_class.from_dict( self.image_processor_dict ,size=4_2 ,max_size=8_4 ,pad_and_return_pixel_mask=lowercase__ ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2, '''longest_edge''': 8_4} ) self.assertEqual(image_processor.do_pad ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): pass def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ,batched=lowercase__ ) __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ,batched=lowercase__ ) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ) self.assertEqual( encoded_images.shape ,(1, self.image_processor_tester.num_channels, expected_height, expected_width) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values __lowercase , __lowercase = self.image_processor_tester.get_expected_values(lowercase__ ,batched=lowercase__ ) self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) ,) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_annotations.txt''' ,'''r''' ) as f: __lowercase = json.loads(f.read() ) __lowercase = {'''image_id''': 3_9_7_6_9, '''annotations''': target} # encode them __lowercase = DeformableDetrImageProcessor() __lowercase = image_processing(images=lowercase__ ,annotations=lowercase__ ,return_tensors='''pt''' ) # verify pixel values __lowercase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['''pixel_values'''].shape ,lowercase__ ) __lowercase = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] ,lowercase__ ,atol=1e-4 ) ) # verify area __lowercase = torch.tensor([5_8_8_7.9_6_0_0, 1_1_2_5_0.2_0_6_1, 4_8_9_3_5_3.8_4_3_8, 8_3_7_1_2_2.7_5_0_0, 1_4_7_9_6_7.5_1_5_6, 1_6_5_7_3_2.3_4_3_8] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] ,lowercase__ ) ) # verify boxes __lowercase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape ,lowercase__ ) __lowercase = torch.tensor([0.5_5_0_3, 0.2_7_6_5, 0.0_6_0_4, 0.2_2_1_5] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] ,lowercase__ ,atol=1e-3 ) ) # verify image_id __lowercase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] ,lowercase__ ) ) # verify is_crowd __lowercase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] ,lowercase__ ) ) # verify class_labels __lowercase = torch.tensor([7_5, 7_5, 6_3, 6_5, 1_7, 1_7] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] ,lowercase__ ) ) # verify orig_size __lowercase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] ,lowercase__ ) ) # verify size __lowercase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] ,lowercase__ ) ) @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) with open('''./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt''' ,'''r''' ) as f: __lowercase = json.loads(f.read() ) __lowercase = {'''file_name''': '''000000039769.png''', '''image_id''': 3_9_7_6_9, '''segments_info''': target} __lowercase = pathlib.Path('''./tests/fixtures/tests_samples/COCO/coco_panoptic''' ) # encode them __lowercase = DeformableDetrImageProcessor(format='''coco_panoptic''' ) __lowercase = image_processing(images=lowercase__ ,annotations=lowercase__ ,masks_path=lowercase__ ,return_tensors='''pt''' ) # verify pixel values __lowercase = torch.Size([1, 3, 8_0_0, 1_0_6_6] ) self.assertEqual(encoding['''pixel_values'''].shape ,lowercase__ ) __lowercase = torch.tensor([0.2_7_9_6, 0.3_1_3_8, 0.3_4_8_1] ) self.assertTrue(torch.allclose(encoding['''pixel_values'''][0, 0, 0, :3] ,lowercase__ ,atol=1e-4 ) ) # verify area __lowercase = torch.tensor([1_4_7_9_7_9.6_8_7_5, 1_6_5_5_2_7.0_4_6_9, 4_8_4_6_3_8.5_9_3_8, 1_1_2_9_2.9_3_7_5, 5_8_7_9.6_5_6_2, 7_6_3_4.1_1_4_7] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''area'''] ,lowercase__ ) ) # verify boxes __lowercase = torch.Size([6, 4] ) self.assertEqual(encoding['''labels'''][0]['''boxes'''].shape ,lowercase__ ) __lowercase = torch.tensor([0.2_6_2_5, 0.5_4_3_7, 0.4_6_8_8, 0.8_6_2_5] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''boxes'''][0] ,lowercase__ ,atol=1e-3 ) ) # verify image_id __lowercase = torch.tensor([3_9_7_6_9] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''image_id'''] ,lowercase__ ) ) # verify is_crowd __lowercase = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''iscrowd'''] ,lowercase__ ) ) # verify class_labels __lowercase = torch.tensor([1_7, 1_7, 6_3, 7_5, 7_5, 9_3] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''class_labels'''] ,lowercase__ ) ) # verify masks __lowercase = 8_2_2_8_7_3 self.assertEqual(encoding['''labels'''][0]['''masks'''].sum().item() ,lowercase__ ) # verify orig_size __lowercase = torch.tensor([4_8_0, 6_4_0] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''orig_size'''] ,lowercase__ ) ) # verify size __lowercase = torch.tensor([8_0_0, 1_0_6_6] ) self.assertTrue(torch.allclose(encoding['''labels'''][0]['''size'''] ,lowercase__ ) )
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' from __future__ import annotations import inspect import unittest import numpy as np from transformers import DeiTConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, TFDeiTModel, ) from transformers.models.deit.modeling_tf_deit import TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DeiTImageProcessor class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : int ,lowercase__ : str=1_3 ,lowercase__ : List[str]=3_0 ,lowercase__ : Optional[int]=2 ,lowercase__ : List[str]=3 ,lowercase__ : Tuple=True ,lowercase__ : List[str]=True ,lowercase__ : int=3_2 ,lowercase__ : Optional[Any]=2 ,lowercase__ : str=4 ,lowercase__ : Tuple=3_7 ,lowercase__ : Dict="gelu" ,lowercase__ : List[Any]=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Dict=1_0 ,lowercase__ : Optional[Any]=0.0_2 ,lowercase__ : List[str]=3 ,lowercase__ : Tuple=None ,lowercase__ : Optional[int]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_size __lowercase = num_channels __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = encoder_stride # in DeiT, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distilation tokens) __lowercase = (image_size // patch_size) ** 2 __lowercase = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return DeiTConfig( image_size=self.image_size ,patch_size=self.patch_size ,num_channels=self.num_channels ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=__UpperCamelCase ,initializer_range=self.initializer_range ,encoder_stride=self.encoder_stride ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,lowercase__ : Any ): __lowercase = TFDeiTModel(config=__UpperCamelCase ) __lowercase = model(__UpperCamelCase ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, Any] ,lowercase__ : Any ,lowercase__ : Optional[Any] ): __lowercase = TFDeiTForMaskedImageModeling(config=__UpperCamelCase ) __lowercase = model(__UpperCamelCase ) self.parent.assertEqual( result.reconstruction.shape ,(self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __lowercase = 1 __lowercase = TFDeiTForMaskedImageModeling(__UpperCamelCase ) __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(__UpperCamelCase ) self.parent.assertEqual(result.reconstruction.shape ,(self.batch_size, 1, self.image_size, self.image_size) ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ,lowercase__ : int ,lowercase__ : Any ): __lowercase = self.type_sequence_label_size __lowercase = TFDeiTForImageClassification(__UpperCamelCase ) __lowercase = model(__UpperCamelCase ,labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) # test greyscale images __lowercase = 1 __lowercase = TFDeiTForImageClassification(__UpperCamelCase ) __lowercase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __lowercase = model(__UpperCamelCase ,labels=__UpperCamelCase ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.type_sequence_label_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_tf class lowercase_ (__snake_case , __snake_case , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = ( ( TFDeiTModel, TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher, TFDeiTForMaskedImageModeling, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': TFDeiTModel, 'image-classification': (TFDeiTForImageClassification, TFDeiTForImageClassificationWithTeacher), } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : Optional[int] = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = TFDeiTModelTester(self ) __lowercase = ConfigTester(self ,config_class=__UpperCamelCase ,has_text_modality=__UpperCamelCase ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() @unittest.skip(reason='''DeiT does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(__UpperCamelCase ) self.assertIsInstance(model.get_input_embeddings() ,(tf.keras.layers.Layer) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCamelCase ,tf.keras.layers.Dense ) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(__UpperCamelCase ) __lowercase = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,__UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCamelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : Any ,lowercase__ : Any=False ): __lowercase = super()._prepare_for_class(__UpperCamelCase ,__UpperCamelCase ,return_labels=__UpperCamelCase ) if return_labels: if "labels" in inputs_dict and "labels" not in inspect.signature(model_class.call ).parameters: del inputs_dict["labels"] return inputs_dict @slow def SCREAMING_SNAKE_CASE ( self : Tuple ): for model_name in TF_DEIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = TFDeiTModel.from_pretrained(__UpperCamelCase ) self.assertIsNotNone(__UpperCamelCase ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_tf @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Tuple ): return ( DeiTImageProcessor.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ) if is_vision_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = TFDeiTForImageClassificationWithTeacher.from_pretrained('''facebook/deit-base-distilled-patch16-224''' ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=__UpperCamelCase ,return_tensors='''tf''' ) # forward pass __lowercase = model(**__UpperCamelCase ) # verify the logits __lowercase = tf.TensorShape((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,__UpperCamelCase ) __lowercase = tf.constant([-1.0_2_6_6, 0.1_9_1_2, -1.2_8_6_1] ) self.assertTrue(np.allclose(outputs.logits[0, :3] ,__UpperCamelCase ,atol=1e-4 ) )
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mvp import MvpTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} # See all MVP models at https://huggingface.co/models?filter=mvp lowerCAmelCase__ = { 'vocab_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/vocab.json', }, 'added_tokens.json': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/added_tokens.json', }, 'merges_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/merges.txt', }, 'tokenizer_file': { 'RUCAIBox/mvp': 'https://huggingface.co/RUCAIBox/mvp/resolve/main/tokenizer.json', }, } lowerCAmelCase__ = { 'RUCAIBox/mvp': 1024, } class lowercase_ (a__ ): """simple docstring""" SCREAMING_SNAKE_CASE : str = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : str = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : str = ["""input_ids""", """attention_mask"""] SCREAMING_SNAKE_CASE : List[Any] = MvpTokenizer def __init__( self : int ,lowercase__ : List[Any]=None ,lowercase__ : Optional[Any]=None ,lowercase__ : List[str]=None ,lowercase__ : List[str]="replace" ,lowercase__ : Optional[int]="<s>" ,lowercase__ : Union[str, Any]="</s>" ,lowercase__ : Any="</s>" ,lowercase__ : str="<s>" ,lowercase__ : str="<unk>" ,lowercase__ : Union[str, Any]="<pad>" ,lowercase__ : Dict="<mask>" ,lowercase__ : Dict=False ,lowercase__ : Tuple=True ,**lowercase__ : List[Any] ,): super().__init__( lowercase__ ,lowercase__ ,tokenizer_file=lowercase__ ,errors=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,sep_token=lowercase__ ,cls_token=lowercase__ ,unk_token=lowercase__ ,pad_token=lowercase__ ,mask_token=lowercase__ ,add_prefix_space=lowercase__ ,trim_offsets=lowercase__ ,**lowercase__ ,) __lowercase = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('''add_prefix_space''' ,lowercase__ ) != add_prefix_space: __lowercase = getattr(lowercase__ ,pre_tok_state.pop('''type''' ) ) __lowercase = add_prefix_space __lowercase = pre_tok_class(**lowercase__ ) __lowercase = add_prefix_space # the pre_tokenizer is already updated in the GPT2TokenizerFast `__init__` __lowercase = '''post_processor''' __lowercase = getattr(self.backend_tokenizer ,lowercase__ ,lowercase__ ) if tokenizer_component_instance: __lowercase = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: __lowercase = tuple(state['''sep'''] ) if "cls" in state: __lowercase = tuple(state['''cls'''] ) __lowercase = False if state.get('''add_prefix_space''' ,lowercase__ ) != add_prefix_space: __lowercase = add_prefix_space __lowercase = True if state.get('''trim_offsets''' ,lowercase__ ) != trim_offsets: __lowercase = trim_offsets __lowercase = True if changes_to_apply: __lowercase = getattr(lowercase__ ,state.pop('''type''' ) ) __lowercase = component_class(**lowercase__ ) setattr(self.backend_tokenizer ,lowercase__ ,lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : List[str] ): if self._mask_token is None: if self.verbose: logger.error('''Using mask_token, but it is not set yet.''' ) return None return str(self._mask_token ) @mask_token.setter def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Optional[int] ): __lowercase = AddedToken(lowercase__ ,lstrip=lowercase__ ,rstrip=lowercase__ ) if isinstance(lowercase__ ,lowercase__ ) else value __lowercase = value def SCREAMING_SNAKE_CASE ( self : Tuple ,*lowercase__ : int ,**lowercase__ : int ): __lowercase = kwargs.get('''is_split_into_words''' ,lowercase__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " '''to use it with pretokenized inputs.''' ) return super()._batch_encode_plus(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,*lowercase__ : Dict ,**lowercase__ : Dict ): __lowercase = kwargs.get('''is_split_into_words''' ,lowercase__ ) if is_split_into_words and not self.add_prefix_space: raise ValueError( F"You need to instantiate {self.__class__.__name__} with add_prefix_space=True " '''to use it with pretokenized inputs.''' ) return super()._encode_plus(*lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : Optional[Any] = None ): __lowercase = self._tokenizer.model.save(lowercase__ ,name=lowercase__ ) return tuple(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Dict ,lowercase__ : Union[str, Any]=None ): __lowercase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Dict ,lowercase__ : Dict = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [[0] * 3 for _ in range(length + 1 )] for row_length in range(length + 1 ): for tile_length in range(2 , 5 ): for tile_start in range(row_length - tile_length + 1 ): different_colour_ways_number[row_length][tile_length - 2] += ( different_colour_ways_number[row_length - tile_start - tile_length][ tile_length - 2 ] + 1 ) return sum(different_colour_ways_number[length] ) if __name__ == "__main__": print(f'{solution() = }')
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' import copy import os import cva import numpy as np from matplotlib import pyplot as plt class lowercase_ : """simple docstring""" def __init__( self : int ): __lowercase = '''''' __lowercase = '''''' __lowercase = [] __lowercase = 0 __lowercase = 2_5_6 __lowercase = 0 __lowercase = 0 __lowercase = 0 __lowercase = 0 def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ): __lowercase = cva.imread(__snake_case ,0 ) __lowercase = copy.deepcopy(self.img ) __lowercase = plt.hist(self.img.ravel() ,2_5_6 ,[0, 2_5_6] ,label='''x''' ) __lowercase = np.sum(__snake_case ) for i in range(len(__snake_case ) ): __lowercase = x[i] / self.k self.sk += prk __lowercase = (self.L - 1) * self.sk if self.rem != 0: __lowercase = int(last % last ) __lowercase = int(last + 1 if self.rem >= 0.5 else last ) self.last_list.append(__snake_case ) __lowercase = int(np.ma.count(self.img ) / self.img[1].size ) __lowercase = self.img[1].size for i in range(self.number_of_cols ): for j in range(self.number_of_rows ): __lowercase = self.img[j][i] if num != self.last_list[num]: __lowercase = self.last_list[num] cva.imwrite('''output_data/output.jpg''' ,self.img ) def SCREAMING_SNAKE_CASE ( self : int ): plt.hist(self.img.ravel() ,2_5_6 ,[0, 2_5_6] ) def SCREAMING_SNAKE_CASE ( self : str ): cva.imshow('''Output-Image''' ,self.img ) cva.imshow('''Input-Image''' ,self.original_image ) cva.waitKey(5_0_0_0 ) cva.destroyAllWindows() if __name__ == "__main__": lowerCAmelCase__ = os.path.join(os.path.basename(__file__), '''image_data/input.jpg''') lowerCAmelCase__ = ConstantStretch() stretcher.stretch(file_path) stretcher.plot_histogram() stretcher.show_image()
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' from math import factorial def _A ( A__ = 20 ): """simple docstring""" __lowercase = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... __lowercase = n // 2 return int(factorial(A__ ) / (factorial(A__ ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: lowerCAmelCase__ = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' from __future__ import annotations def _A ( A__ , A__ , A__ ): """simple docstring""" if days_between_payments <= 0: raise ValueError('''days_between_payments must be > 0''' ) if daily_interest_rate < 0: raise ValueError('''daily_interest_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * daily_interest_rate * days_between_payments def _A ( A__ , A__ , A__ , ): """simple docstring""" if number_of_compounding_periods <= 0: raise ValueError('''number_of_compounding_periods must be > 0''' ) if nominal_annual_interest_rate_percentage < 0: raise ValueError('''nominal_annual_interest_rate_percentage must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return principal * ( (1 + nominal_annual_interest_rate_percentage) ** number_of_compounding_periods - 1 ) def _A ( A__ , A__ , A__ , ): """simple docstring""" if number_of_years <= 0: raise ValueError('''number_of_years must be > 0''' ) if nominal_annual_percentage_rate < 0: raise ValueError('''nominal_annual_percentage_rate must be >= 0''' ) if principal <= 0: raise ValueError('''principal must be > 0''' ) return compound_interest( lowercase_ , nominal_annual_percentage_rate / 365 , number_of_years * 365 ) if __name__ == "__main__": import doctest doctest.testmod()
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowerCAmelCase__ = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' from math import factorial def _A ( A__ = 100 ): """simple docstring""" return sum(int(A__ ) for x in str(factorial(A__ ) ) ) if __name__ == "__main__": print(solution(int(input('''Enter the Number: ''').strip())))
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
def _A ( A__ = 200 ): """simple docstring""" __lowercase = [1, 2, 5, 10, 20, 50, 100, 200] __lowercase = [0] * (pence + 1) __lowercase = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(snake_case_ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Dict ,**lowercase__ : List[Any] ): pass def _A ( A__ ): """simple docstring""" return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. lowerCAmelCase__ = ( 'https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png' ) @is_pipeline_test @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,lowercase__ : Dict ): __lowercase = pipeline( '''document-question-answering''' ,model=UpperCamelCase__ ,tokenizer=UpperCamelCase__ ,image_processor=UpperCamelCase__ ) __lowercase = INVOICE_URL __lowercase = list(zip(*apply_tesseract(load_image(UpperCamelCase__ ) ,UpperCamelCase__ ,'''''' ) ) ) __lowercase = '''What is the placebo?''' __lowercase = [ { '''image''': load_image(UpperCamelCase__ ), '''question''': question, }, { '''image''': image, '''question''': question, }, { '''image''': image, '''question''': question, '''word_boxes''': word_boxes, }, ] return dqa_pipeline, examples def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : int ,lowercase__ : List[Any] ): __lowercase = dqa_pipeline(UpperCamelCase__ ,top_k=2 ) self.assertEqual( UpperCamelCase__ ,[ [ {'''score''': ANY(UpperCamelCase__ ), '''answer''': ANY(UpperCamelCase__ ), '''start''': ANY(UpperCamelCase__ ), '''end''': ANY(UpperCamelCase__ )}, {'''score''': ANY(UpperCamelCase__ ), '''answer''': ANY(UpperCamelCase__ ), '''start''': ANY(UpperCamelCase__ ), '''end''': ANY(UpperCamelCase__ )}, ] ] * 3 ,) @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = pipeline('''document-question-answering''' ,model='''hf-internal-testing/tiny-random-layoutlmv2''' ) __lowercase = INVOICE_URL __lowercase = '''How many cats are there?''' __lowercase = [ {'''score''': 0.0_0_0_1, '''answer''': '''oy 2312/2019''', '''start''': 3_8, '''end''': 3_9}, {'''score''': 0.0_0_0_1, '''answer''': '''oy 2312/2019 DUE''', '''start''': 3_8, '''end''': 4_0}, ] __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual(nested_simplify(UpperCamelCase__ ,decimals=4 ) ,UpperCamelCase__ ) __lowercase = dqa_pipeline({'''image''': image, '''question''': question} ,top_k=2 ) self.assertEqual(nested_simplify(UpperCamelCase__ ,decimals=4 ) ,UpperCamelCase__ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably __lowercase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual(UpperCamelCase__ ,[] ) # We can optionnally pass directly the words and bounding boxes __lowercase = '''./tests/fixtures/tests_samples/COCO/000000039769.png''' __lowercase = [] __lowercase = [] __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,words=UpperCamelCase__ ,boxes=UpperCamelCase__ ,top_k=2 ) self.assertEqual(UpperCamelCase__ ,[] ) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = pipeline( '''document-question-answering''' ,model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' ,revision='''9977165''' ,) __lowercase = INVOICE_URL __lowercase = '''What is the invoice number?''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) __lowercase = dqa_pipeline({'''image''': image, '''question''': question} ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) __lowercase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ [ {'''score''': 0.9_9_4_4, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_0_0_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ], ] * 2 ,) @slow @require_torch @require_detectrona @require_pytesseract def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = pipeline( '''document-question-answering''' ,model='''tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa''' ,revision='''9977165''' ,max_seq_len=5_0 ,) __lowercase = INVOICE_URL __lowercase = '''What is the invoice number?''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) __lowercase = dqa_pipeline({'''image''': image, '''question''': question} ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) __lowercase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ [ {'''score''': 0.9_9_7_4, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, {'''score''': 0.9_9_4_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ] * 2 ,) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' ,revision='''3dc6de3''' ,add_prefix_space=UpperCamelCase__ ) __lowercase = pipeline( '''document-question-answering''' ,model='''impira/layoutlm-document-qa''' ,tokenizer=UpperCamelCase__ ,revision='''3dc6de3''' ,) __lowercase = INVOICE_URL __lowercase = '''What is the invoice number?''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] ,) __lowercase = dqa_pipeline({'''image''': image, '''question''': question} ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] ,) __lowercase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ [ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] ] * 2 ,) __lowercase = list(zip(*apply_tesseract(load_image(UpperCamelCase__ ) ,UpperCamelCase__ ,'''''' ) ) ) # This model should also work if `image` is set to None __lowercase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.4_2_5_1, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.0_8_1_9, '''answer''': '''1110212019''', '''start''': 2_3, '''end''': 2_3}, ] ,) @slow @require_torch @require_pytesseract @require_vision def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = AutoTokenizer.from_pretrained( '''impira/layoutlm-document-qa''' ,revision='''3dc6de3''' ,add_prefix_space=UpperCamelCase__ ) __lowercase = pipeline( '''document-question-answering''' ,model='''impira/layoutlm-document-qa''' ,tokenizer=UpperCamelCase__ ,revision='''3dc6de3''' ,max_seq_len=5_0 ,) __lowercase = INVOICE_URL __lowercase = '''What is the invoice number?''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) __lowercase = dqa_pipeline( [{'''image''': image, '''question''': question}, {'''image''': image, '''question''': question}] ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ [ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ] * 2 ,) __lowercase = list(zip(*apply_tesseract(load_image(UpperCamelCase__ ) ,UpperCamelCase__ ,'''''' ) ) ) # This model should also work if `image` is set to None __lowercase = dqa_pipeline({'''image''': None, '''word_boxes''': word_boxes, '''question''': question} ,top_k=2 ) self.assertEqual( nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[ {'''score''': 0.9_9_9_9, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, {'''score''': 0.9_9_9_8, '''answer''': '''us-001''', '''start''': 1_6, '''end''': 1_6}, ] ,) @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = pipeline( '''document-question-answering''' ,model='''naver-clova-ix/donut-base-finetuned-docvqa''' ,tokenizer=AutoTokenizer.from_pretrained('''naver-clova-ix/donut-base-finetuned-docvqa''' ) ,feature_extractor='''naver-clova-ix/donut-base-finetuned-docvqa''' ,) __lowercase = INVOICE_URL __lowercase = '''What is the invoice number?''' __lowercase = dqa_pipeline(image=UpperCamelCase__ ,question=UpperCamelCase__ ,top_k=2 ) self.assertEqual(nested_simplify(UpperCamelCase__ ,decimals=4 ) ,[{'''answer''': '''us-001'''}] ) @require_tf @unittest.skip('''Document question answering not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : int ): pass
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Any ): __lowercase = num_of_nodes __lowercase = [] __lowercase = {} def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : int ,lowercase__ : List[Any] ): self.m_edges.append([u_node, v_node, weight] ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : Union[str, Any] ): if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node] ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : str ): if self.m_component[u_node] != u_node: for k in self.m_component: __lowercase = self.find_component(_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Optional[int] ): if component_size[u_node] <= component_size[v_node]: __lowercase = v_node component_size[v_node] += component_size[u_node] self.set_component(_lowerCamelCase ) elif component_size[u_node] >= component_size[v_node]: __lowercase = self.find_component(_lowerCamelCase ) component_size[u_node] += component_size[v_node] self.set_component(_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = [] __lowercase = 0 __lowercase = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes ): self.m_component.update({node: node} ) component_size.append(1 ) __lowercase = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: __lowercase , __lowercase , __lowercase = edge __lowercase = self.m_component[u] __lowercase = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): __lowercase = [u, v, w] for edge in minimum_weight_edge: if isinstance(_lowerCamelCase ,_lowerCamelCase ): __lowercase , __lowercase , __lowercase = edge __lowercase = self.m_component[u] __lowercase = self.m_component[v] if u_component != v_component: mst_weight += w self.union(_lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) print(F"Added edge [{u} - {v}]\nAdded weight: {w}\n" ) num_of_components -= 1 __lowercase = [-1] * self.m_num_of_nodes print(F"The total weight of the minimal spanning tree is: {mst_weight}" ) def _A ( ): pass if __name__ == "__main__": import doctest doctest.testmod()
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : int ,lowercase__ : List[str]=7 ,lowercase__ : Optional[Any]=3 ,lowercase__ : Optional[int]=1_8 ,lowercase__ : str=3_0 ,lowercase__ : str=4_0_0 ,lowercase__ : Dict=True ,lowercase__ : List[Any]=None ,lowercase__ : List[Any]=True ,lowercase__ : str=None ,): __lowercase = size if size is not None else {"shortest_edge": 2_0} __lowercase = crop_size if crop_size is not None else {"height": 1_8, "width": 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = do_center_crop __lowercase = crop_size def SCREAMING_SNAKE_CASE ( self : List[str] ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class lowercase_ (_A , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = MobileNetVaImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = MobileNetVaImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : str ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase__ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase__ ,'''size''' ) ) self.assertTrue(hasattr(lowercase__ ,'''do_center_crop''' ) ) self.assertTrue(hasattr(lowercase__ ,'''crop_size''' ) ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 2_0} ) self.assertEqual(image_processor.crop_size ,{'''height''': 1_8, '''width''': 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ,crop_size=8_4 ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size ,{'''height''': 8_4, '''width''': 8_4} ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,numpify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=lowercase__ ,torchify=lowercase__ ) for image in image_inputs: self.assertIsInstance(lowercase__ ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched __lowercase = image_processing(lowercase__ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,)
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' import os from pathlib import Path def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = { '''en''': '''Machine learning is great, isn\'t it?''', '''ru''': '''Машинное обучение - это здорово, не так ли?''', '''de''': '''Maschinelles Lernen ist großartig, oder?''', } # BLUE scores as follows: # "pair": [fairseq, transformers] __lowercase = { '''ru-en''': ['''[41.3](http://matrix.statmt.org/matrix/output/1907?run_id=6937)''', '''39.20'''], '''en-ru''': ['''[36.4](http://matrix.statmt.org/matrix/output/1914?run_id=6724)''', '''33.47'''], '''en-de''': ['''[43.1](http://matrix.statmt.org/matrix/output/1909?run_id=6862)''', '''42.83'''], '''de-en''': ['''[42.3](http://matrix.statmt.org/matrix/output/1902?run_id=6750)''', '''41.35'''], } __lowercase = F"{src_lang}-{tgt_lang}" __lowercase = F"\n---\nlanguage: \n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt19\n- facebook\nlicense: apache-2.0\ndatasets:\n- wmt19\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of [fairseq wmt19 transformer](https://github.com/pytorch/fairseq/blob/master/examples/wmt19/README.md) for {src_lang}-{tgt_lang}.\n\nFor more details, please see, [Facebook FAIR\'s WMT19 News Translation Task Submission](https://arxiv.org/abs/1907.06616).\n\nThe abbreviation FSMT stands for FairSeqMachineTranslation\n\nAll four models are available:\n\n* [wmt19-en-ru](https://huggingface.co/facebook/wmt19-en-ru)\n* [wmt19-ru-en](https://huggingface.co/facebook/wmt19-ru-en)\n* [wmt19-en-de](https://huggingface.co/facebook/wmt19-en-de)\n* [wmt19-de-en](https://huggingface.co/facebook/wmt19-de-en)\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"facebook/wmt19-{src_lang}-{tgt_lang}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n- The original (and this ported model) doesn\'t seem to handle well inputs with repeated sub-phrases, [content gets truncated](https://discuss.huggingface.co/t/issues-with-translating-inputs-containing-repeated-phrases/981)\n\n## Training data\n\nPretrained weights were left identical to the original model released by fairseq. For more details, please, see the [paper](https://arxiv.org/abs/1907.06616).\n\n## Eval results\n\npair | fairseq | transformers\n-------|---------|----------\n{pair} | {scores[pair][0]} | {scores[pair][1]}\n\nThe score is slightly below the score reported by `fairseq`, since `transformers`` currently doesn\'t support:\n- model ensemble, therefore the best performing checkpoint was ported (``model4.pt``).\n- re-ranking\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=15\nmkdir -p $DATA_DIR\nsacrebleu -t wmt19 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt19 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py facebook/wmt19-$PAIR $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\nnote: fairseq reports using a beam of 50, so you should get a slightly higher score if re-run with `--num_beams 50`.\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt19/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2019.tgz?1556572561)\n\n\n### BibTeX entry and citation info\n\n```bibtex\n@inproceedings{{...,\n year={{2020}},\n title={{Facebook FAIR\'s WMT19 News Translation Task Submission}},\n author={{Ng, Nathan and Yee, Kyra and Baevski, Alexei and Ott, Myle and Auli, Michael and Edunov, Sergey}},\n booktitle={{Proc. of WMT}},\n}}\n```\n\n\n## TODO\n\n- port model ensemble (fairseq uses 4 model checkpoints)\n\n" os.makedirs(a__ , exist_ok=a__ ) __lowercase = os.path.join(a__ , '''README.md''' ) print(F"Generating {path}" ) with open(a__ , '''w''' , encoding='''utf-8''' ) as f: f.write(a__ ) # make sure we are under the root of the project lowerCAmelCase__ = Path(__file__).resolve().parent.parent.parent lowerCAmelCase__ = repo_dir / '''model_cards''' for model_name in ["wmt19-ru-en", "wmt19-en-ru", "wmt19-en-de", "wmt19-de-en"]: lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = model_name.split('''-''') lowerCAmelCase__ = model_cards_dir / '''facebook''' / model_name write_model_card(model_card_dir, src_lang=src_lang, tgt_lang=tgt_lang)
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" __lowercase = [1] __lowercase , __lowercase , __lowercase = 0, 0, 0 __lowercase = ugly_nums[ia] * 2 __lowercase = ugly_nums[ia] * 3 __lowercase = ugly_nums[ia] * 5 for _ in range(1 , UpperCamelCase__ ): __lowercase = min(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) ugly_nums.append(UpperCamelCase__ ) if next_num == next_a: ia += 1 __lowercase = ugly_nums[ia] * 2 if next_num == next_a: ia += 1 __lowercase = ugly_nums[ia] * 3 if next_num == next_a: ia += 1 __lowercase = ugly_nums[ia] * 5 return ugly_nums[-1] if __name__ == "__main__": from doctest import testmod testmod(verbose=True) print(f'{ugly_numbers(200) = }')
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' import json import os import shutil import warnings from argparse import ArgumentParser, Namespace from pathlib import Path from typing import List from ..utils import logging from . import BaseTransformersCLICommand try: from cookiecutter.main import cookiecutter lowerCAmelCase__ = True except ImportError: lowerCAmelCase__ = False lowerCAmelCase__ = logging.get_logger(__name__) # pylint: disable=invalid-name def _A ( A__ ): """simple docstring""" return AddNewModelCommand(args.testing , args.testing_file , path=args.path ) class lowercase_ (__UpperCAmelCase ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''add-new-model''' ) add_new_model_parser.add_argument('''--testing''' ,action='''store_true''' ,help='''If in testing mode.''' ) add_new_model_parser.add_argument('''--testing_file''' ,type=__SCREAMING_SNAKE_CASE ,help='''Configuration file on which to run.''' ) add_new_model_parser.add_argument( '''--path''' ,type=__SCREAMING_SNAKE_CASE ,help='''Path to cookiecutter. Should only be used for testing purposes.''' ) add_new_model_parser.set_defaults(func=__SCREAMING_SNAKE_CASE ) def __init__( self : Any ,lowercase__ : bool ,lowercase__ : str ,lowercase__ : int=None ,*lowercase__ : List[Any] ): __lowercase = testing __lowercase = testing_file __lowercase = path def SCREAMING_SNAKE_CASE ( self : List[str] ): warnings.warn( '''The command `transformers-cli add-new-model` is deprecated and will be removed in v5 of Transformers. ''' '''It is not actively maintained anymore, so might give a result that won\'t pass all tests and quality ''' '''checks, you should use `transformers-cli add-new-model-like` instead.''' ) if not _has_cookiecutter: raise ImportError( '''Model creation dependencies are required to use the `add_new_model` command. Install them by running ''' '''the following at the root of your `transformers` clone:\n\n\t$ pip install -e .[modelcreation]\n''' ) # Ensure that there is no other `cookiecutter-template-xxx` directory in the current working directory __lowercase = [directory for directory in os.listdir() if '''cookiecutter-template-''' == directory[:2_2]] if len(__SCREAMING_SNAKE_CASE ) > 0: raise ValueError( '''Several directories starting with `cookiecutter-template-` in current working directory. ''' '''Please clean your directory by removing all folders starting with `cookiecutter-template-` or ''' '''change your working directory.''' ) __lowercase = ( Path(__SCREAMING_SNAKE_CASE ).parent.parent.parent.parent if self._path is None else Path(self._path ).parent.parent ) __lowercase = path_to_transformer_root / '''templates''' / '''adding_a_new_model''' # Execute cookiecutter if not self._testing: cookiecutter(str(__SCREAMING_SNAKE_CASE ) ) else: with open(self._testing_file ,'''r''' ) as configuration_file: __lowercase = json.load(__SCREAMING_SNAKE_CASE ) cookiecutter( str(path_to_cookiecutter if self._path is None else self._path ) ,no_input=__SCREAMING_SNAKE_CASE ,extra_context=__SCREAMING_SNAKE_CASE ,) __lowercase = [directory for directory in os.listdir() if '''cookiecutter-template-''' in directory[:2_2]][0] # Retrieve configuration with open(directory + '''/configuration.json''' ,'''r''' ) as configuration_file: __lowercase = json.load(__SCREAMING_SNAKE_CASE ) __lowercase = configuration['''lowercase_modelname'''] __lowercase = configuration['''generate_tensorflow_pytorch_and_flax'''] os.remove(F"{directory}/configuration.json" ) __lowercase = '''PyTorch''' in generate_tensorflow_pytorch_and_flax __lowercase = '''TensorFlow''' in generate_tensorflow_pytorch_and_flax __lowercase = '''Flax''' in generate_tensorflow_pytorch_and_flax __lowercase = F"{path_to_transformer_root}/src/transformers/models/{lowercase_model_name}" os.makedirs(__SCREAMING_SNAKE_CASE ,exist_ok=__SCREAMING_SNAKE_CASE ) os.makedirs(F"{path_to_transformer_root}/tests/models/{lowercase_model_name}" ,exist_ok=__SCREAMING_SNAKE_CASE ) # Tests require submodules as they have parent imports with open(F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/__init__.py" ,'''w''' ): pass shutil.move( F"{directory}/__init__.py" ,F"{model_dir}/__init__.py" ,) shutil.move( F"{directory}/configuration_{lowercase_model_name}.py" ,F"{model_dir}/configuration_{lowercase_model_name}.py" ,) def remove_copy_lines(lowercase__ : List[str] ): with open(__SCREAMING_SNAKE_CASE ,'''r''' ) as f: __lowercase = f.readlines() with open(__SCREAMING_SNAKE_CASE ,'''w''' ) as f: for line in lines: if "# Copied from transformers." not in line: f.write(__SCREAMING_SNAKE_CASE ) if output_pytorch: if not self._testing: remove_copy_lines(F"{directory}/modeling_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_{lowercase_model_name}.py" ,F"{model_dir}/modeling_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_{lowercase_model_name}.py" ) if output_tensorflow: if not self._testing: remove_copy_lines(F"{directory}/modeling_tf_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_tf_{lowercase_model_name}.py" ,F"{model_dir}/modeling_tf_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_tf_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_tf_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_tf_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_tf_{lowercase_model_name}.py" ) if output_flax: if not self._testing: remove_copy_lines(F"{directory}/modeling_flax_{lowercase_model_name}.py" ) shutil.move( F"{directory}/modeling_flax_{lowercase_model_name}.py" ,F"{model_dir}/modeling_flax_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/test_modeling_flax_{lowercase_model_name}.py" ,F"{path_to_transformer_root}/tests/models/{lowercase_model_name}/test_modeling_flax_{lowercase_model_name}.py" ,) else: os.remove(F"{directory}/modeling_flax_{lowercase_model_name}.py" ) os.remove(F"{directory}/test_modeling_flax_{lowercase_model_name}.py" ) shutil.move( F"{directory}/{lowercase_model_name}.md" ,F"{path_to_transformer_root}/docs/source/en/model_doc/{lowercase_model_name}.md" ,) shutil.move( F"{directory}/tokenization_{lowercase_model_name}.py" ,F"{model_dir}/tokenization_{lowercase_model_name}.py" ,) shutil.move( F"{directory}/tokenization_fast_{lowercase_model_name}.py" ,F"{model_dir}/tokenization_{lowercase_model_name}_fast.py" ,) from os import fdopen, remove from shutil import copymode, move from tempfile import mkstemp def replace(lowercase__ : str ,lowercase__ : str ,lowercase__ : List[str] ): # Create temp file __lowercase , __lowercase = mkstemp() __lowercase = False with fdopen(__SCREAMING_SNAKE_CASE ,'''w''' ) as new_file: with open(__SCREAMING_SNAKE_CASE ) as old_file: for line in old_file: new_file.write(__SCREAMING_SNAKE_CASE ) if line_to_copy_below in line: __lowercase = True for line_to_copy in lines_to_copy: new_file.write(__SCREAMING_SNAKE_CASE ) if not line_found: raise ValueError(F"Line {line_to_copy_below} was not found in file." ) # Copy the file permissions from the old file to the new file copymode(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) # Remove original file remove(__SCREAMING_SNAKE_CASE ) # Move new file move(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) def skip_units(lowercase__ : Union[str, Any] ): return ( ("generating PyTorch" in line and not output_pytorch) or ("generating TensorFlow" in line and not output_tensorflow) or ("generating Flax" in line and not output_flax) ) def replace_in_files(lowercase__ : str ): with open(__SCREAMING_SNAKE_CASE ) as datafile: __lowercase = [] __lowercase = False __lowercase = False for line in datafile: if "# To replace in: " in line and "##" not in line: __lowercase = line.split('''"''' )[1] __lowercase = skip_units(__SCREAMING_SNAKE_CASE ) elif "# Below: " in line and "##" not in line: __lowercase = line.split('''"''' )[1] __lowercase = skip_units(__SCREAMING_SNAKE_CASE ) elif "# End." in line and "##" not in line: if not skip_file and not skip_snippet: replace(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) __lowercase = [] elif "# Replace with" in line and "##" not in line: __lowercase = [] elif "##" not in line: lines_to_copy.append(__SCREAMING_SNAKE_CASE ) remove(__SCREAMING_SNAKE_CASE ) replace_in_files(F"{directory}/to_replace_{lowercase_model_name}.py" ) os.rmdir(__SCREAMING_SNAKE_CASE )
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from math import sqrt import numpy as np from sympy import symbols # Coefficient # Speed of light (m/s) lowerCAmelCase__ = 2_9979_2458 # Symbols lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = symbols('''ct x y z''') def _A ( A__ ): """simple docstring""" if velocity > c: raise ValueError('''Speed must not exceed light speed 299,792,458 [m/s]!''' ) elif velocity < 1: # Usually the speed should be much higher than 1 (c order of magnitude) raise ValueError('''Speed must be greater than or equal to 1!''' ) return velocity / c def _A ( A__ ): """simple docstring""" return 1 / sqrt(1 - beta(A__ ) ** 2 ) def _A ( A__ ): """simple docstring""" return np.array( [ [gamma(A__ ), -gamma(A__ ) * beta(A__ ), 0, 0], [-gamma(A__ ) * beta(A__ ), gamma(A__ ), 0, 0], [0, 0, 1, 0], [0, 0, 0, 1], ] ) def _A ( A__ , A__ = None ): """simple docstring""" if event is None: __lowercase = np.array([ct, x, y, z] ) # Symbolic four vector else: event[0] *= c # x0 is ct (speed of light * time) return transformation_matrix(A__ ) @ event if __name__ == "__main__": import doctest doctest.testmod() # Example of symbolic vector: lowerCAmelCase__ = transform(2997_9245) print('''Example of four vector: ''') print(f'ct\' = {four_vector[0]}') print(f'x\' = {four_vector[1]}') print(f'y\' = {four_vector[2]}') print(f'z\' = {four_vector[3]}') # Substitute symbols with numerical values lowerCAmelCase__ = {ct: c, x: 1, y: 1, z: 1} lowerCAmelCase__ = [four_vector[i].subs(sub_dict) for i in range(4)] print(f'\n{numerical_vector}')
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' import unittest import numpy as np from transformers.testing_utils import require_pytesseract, require_torch from transformers.utils import is_pytesseract_available, is_torch_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_pytesseract_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : Optional[int] ,lowercase__ : Optional[Any]=7 ,lowercase__ : Any=3 ,lowercase__ : Dict=1_8 ,lowercase__ : List[Any]=3_0 ,lowercase__ : Union[str, Any]=4_0_0 ,lowercase__ : Tuple=True ,lowercase__ : Dict=None ,lowercase__ : Optional[int]=True ,): __lowercase = size if size is not None else {'''height''': 1_8, '''width''': 1_8} __lowercase = parent __lowercase = batch_size __lowercase = num_channels __lowercase = image_size __lowercase = min_resolution __lowercase = max_resolution __lowercase = do_resize __lowercase = size __lowercase = apply_ocr def SCREAMING_SNAKE_CASE ( self : List[Any] ): return {"do_resize": self.do_resize, "size": self.size, "apply_ocr": self.apply_ocr} @require_torch @require_pytesseract class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = LayoutLMvaImageProcessor if is_pytesseract_available() else None def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = LayoutLMvaImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__A ,'''do_resize''' ) ) self.assertTrue(hasattr(__A ,'''size''' ) ) self.assertTrue(hasattr(__A ,'''apply_ocr''' ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''height''': 1_8, '''width''': 1_8} ) __lowercase = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ) self.assertEqual(image_processor.size ,{'''height''': 4_2, '''width''': 4_2} ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ) for image in image_inputs: self.assertIsInstance(__A ,Image.Image ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ) self.assertEqual( encoding.pixel_values.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) self.assertIsInstance(encoding.words ,__A ) self.assertIsInstance(encoding.boxes ,__A ) # Test batched __lowercase = image_processing(__A ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ,numpify=__A ) for image in image_inputs: self.assertIsInstance(__A ,np.ndarray ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched __lowercase = image_processing(__A ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # Initialize image_processing __lowercase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors __lowercase = prepare_image_inputs(self.image_processor_tester ,equal_resolution=__A ,torchify=__A ) for image in image_inputs: self.assertIsInstance(__A ,torch.Tensor ) # Test not batched input __lowercase = image_processing(image_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) # Test batched __lowercase = image_processing(__A ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_images.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size['''height'''], self.image_processor_tester.size['''width'''], ) ,) def SCREAMING_SNAKE_CASE ( self : str ): # with apply_OCR = True __lowercase = LayoutLMvaImageProcessor() from datasets import load_dataset __lowercase = load_dataset('''hf-internal-testing/fixtures_docvqa''' ,split='''test''' ) __lowercase = Image.open(ds[0]['''file'''] ).convert('''RGB''' ) __lowercase = image_processing(__A ,return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) ) self.assertEqual(len(encoding.words ) ,len(encoding.boxes ) ) # fmt: off # the words and boxes were obtained with Tesseract 4.1.1 __lowercase = [['''11:14''', '''to''', '''11:39''', '''a.m''', '''11:39''', '''to''', '''11:44''', '''a.m.''', '''11:44''', '''a.m.''', '''to''', '''12:25''', '''p.m.''', '''12:25''', '''to''', '''12:58''', '''p.m.''', '''12:58''', '''to''', '''4:00''', '''p.m.''', '''2:00''', '''to''', '''5:00''', '''p.m.''', '''Coffee''', '''Break''', '''Coffee''', '''will''', '''be''', '''served''', '''for''', '''men''', '''and''', '''women''', '''in''', '''the''', '''lobby''', '''adjacent''', '''to''', '''exhibit''', '''area.''', '''Please''', '''move''', '''into''', '''exhibit''', '''area.''', '''(Exhibits''', '''Open)''', '''TRRF''', '''GENERAL''', '''SESSION''', '''(PART''', '''|)''', '''Presiding:''', '''Lee''', '''A.''', '''Waller''', '''TRRF''', '''Vice''', '''President''', '''“Introductory''', '''Remarks”''', '''Lee''', '''A.''', '''Waller,''', '''TRRF''', '''Vice''', '''Presi-''', '''dent''', '''Individual''', '''Interviews''', '''with''', '''TRRF''', '''Public''', '''Board''', '''Members''', '''and''', '''Sci-''', '''entific''', '''Advisory''', '''Council''', '''Mem-''', '''bers''', '''Conducted''', '''by''', '''TRRF''', '''Treasurer''', '''Philip''', '''G.''', '''Kuehn''', '''to''', '''get''', '''answers''', '''which''', '''the''', '''public''', '''refrigerated''', '''warehousing''', '''industry''', '''is''', '''looking''', '''for.''', '''Plus''', '''questions''', '''from''', '''the''', '''floor.''', '''Dr.''', '''Emil''', '''M.''', '''Mrak,''', '''University''', '''of''', '''Cal-''', '''ifornia,''', '''Chairman,''', '''TRRF''', '''Board;''', '''Sam''', '''R.''', '''Cecil,''', '''University''', '''of''', '''Georgia''', '''College''', '''of''', '''Agriculture;''', '''Dr.''', '''Stanley''', '''Charm,''', '''Tufts''', '''University''', '''School''', '''of''', '''Medicine;''', '''Dr.''', '''Robert''', '''H.''', '''Cotton,''', '''ITT''', '''Continental''', '''Baking''', '''Company;''', '''Dr.''', '''Owen''', '''Fennema,''', '''University''', '''of''', '''Wis-''', '''consin;''', '''Dr.''', '''Robert''', '''E.''', '''Hardenburg,''', '''USDA.''', '''Questions''', '''and''', '''Answers''', '''Exhibits''', '''Open''', '''Capt.''', '''Jack''', '''Stoney''', '''Room''', '''TRRF''', '''Scientific''', '''Advisory''', '''Council''', '''Meeting''', '''Ballroom''', '''Foyer''']] # noqa: E231 __lowercase = [[[1_4_1, 5_7, 2_1_4, 6_9], [2_2_8, 5_8, 2_5_2, 6_9], [1_4_1, 7_5, 2_1_6, 8_8], [2_3_0, 7_9, 2_8_0, 8_8], [1_4_2, 2_6_0, 2_1_8, 2_7_3], [2_3_0, 2_6_1, 2_5_5, 2_7_3], [1_4_3, 2_7_9, 2_1_8, 2_9_0], [2_3_1, 2_8_2, 2_9_0, 2_9_1], [1_4_3, 3_4_2, 2_1_8, 3_5_4], [2_3_1, 3_4_5, 2_8_9, 3_5_5], [2_0_2, 3_6_2, 2_2_7, 3_7_3], [1_4_3, 3_7_9, 2_2_0, 3_9_2], [2_3_1, 3_8_2, 2_9_1, 3_9_4], [1_4_4, 7_1_4, 2_2_0, 7_2_6], [2_3_1, 7_1_5, 2_5_6, 7_2_6], [1_4_4, 7_3_2, 2_2_0, 7_4_5], [2_3_2, 7_3_6, 2_9_1, 7_4_7], [1_4_4, 7_6_9, 2_1_8, 7_8_2], [2_3_1, 7_7_0, 2_5_6, 7_8_2], [1_4_1, 7_8_8, 2_0_2, 8_0_1], [2_1_5, 7_9_1, 2_7_4, 8_0_4], [1_4_3, 8_2_6, 2_0_4, 8_3_8], [2_1_5, 8_2_6, 2_4_0, 8_3_8], [1_4_2, 8_4_4, 2_0_2, 8_5_7], [2_1_5, 8_4_7, 2_7_4, 8_5_9], [3_3_4, 5_7, 4_2_7, 6_9], [4_4_0, 5_7, 5_2_2, 6_9], [3_6_9, 7_5, 4_6_1, 8_8], [4_6_9, 7_5, 5_1_6, 8_8], [5_2_8, 7_6, 5_6_2, 8_8], [5_7_0, 7_6, 6_6_7, 8_8], [6_7_5, 7_5, 7_1_1, 8_7], [7_2_1, 7_9, 7_7_8, 8_8], [7_8_9, 7_5, 8_4_0, 8_8], [3_6_9, 9_7, 4_7_0, 1_0_7], [4_8_4, 9_4, 5_0_7, 1_0_6], [5_1_8, 9_4, 5_6_2, 1_0_7], [5_7_6, 9_4, 6_5_5, 1_1_0], [6_6_8, 9_4, 7_9_2, 1_0_9], [8_0_4, 9_5, 8_2_9, 1_0_7], [3_6_9, 1_1_3, 4_6_5, 1_2_5], [4_7_7, 1_1_6, 5_4_7, 1_2_5], [5_6_2, 1_1_3, 6_5_8, 1_2_5], [6_7_1, 1_1_6, 7_4_8, 1_2_5], [7_6_1, 1_1_3, 8_1_1, 1_2_5], [3_6_9, 1_3_1, 4_6_5, 1_4_3], [4_7_7, 1_3_3, 5_4_8, 1_4_3], [5_6_3, 1_3_0, 6_9_8, 1_4_5], [7_1_0, 1_3_0, 8_0_2, 1_4_6], [3_3_6, 1_7_1, 4_1_2, 1_8_3], [4_2_3, 1_7_1, 5_7_2, 1_8_3], [5_8_2, 1_7_0, 7_1_6, 1_8_4], [7_2_8, 1_7_1, 8_1_7, 1_8_7], [8_2_9, 1_7_1, 8_4_4, 1_8_6], [3_3_8, 1_9_7, 4_8_2, 2_1_2], [5_0_7, 1_9_6, 5_5_7, 2_0_9], [5_6_9, 1_9_6, 5_9_5, 2_0_8], [6_1_0, 1_9_6, 7_0_2, 2_0_9], [5_0_5, 2_1_4, 5_8_3, 2_2_6], [5_9_5, 2_1_4, 6_5_6, 2_2_7], [6_7_0, 2_1_5, 8_0_7, 2_2_7], [3_3_5, 2_5_9, 5_4_3, 2_7_4], [5_5_6, 2_5_9, 7_0_8, 2_7_2], [3_7_2, 2_7_9, 4_2_2, 2_9_1], [4_3_5, 2_7_9, 4_6_0, 2_9_1], [4_7_4, 2_7_9, 5_7_4, 2_9_2], [5_8_7, 2_7_8, 6_6_4, 2_9_1], [6_7_6, 2_7_8, 7_3_8, 2_9_1], [7_5_1, 2_7_9, 8_3_4, 2_9_1], [3_7_2, 2_9_8, 4_3_4, 3_1_0], [3_3_5, 3_4_1, 4_8_3, 3_5_4], [4_9_7, 3_4_1, 6_5_5, 3_5_4], [6_6_7, 3_4_1, 7_2_8, 3_5_4], [7_4_0, 3_4_1, 8_2_5, 3_5_4], [3_3_5, 3_6_0, 4_3_0, 3_7_2], [4_4_2, 3_6_0, 5_3_4, 3_7_2], [5_4_5, 3_5_9, 6_8_7, 3_7_2], [6_9_7, 3_6_0, 7_5_4, 3_7_2], [7_6_5, 3_6_0, 8_2_3, 3_7_3], [3_3_4, 3_7_8, 4_2_8, 3_9_1], [4_4_0, 3_7_8, 5_7_7, 3_9_4], [5_9_0, 3_7_8, 7_0_5, 3_9_1], [7_2_0, 3_7_8, 8_0_1, 3_9_1], [3_3_4, 3_9_7, 4_0_0, 4_0_9], [3_7_0, 4_1_6, 5_2_9, 4_2_9], [5_4_4, 4_1_6, 5_7_6, 4_3_2], [5_8_7, 4_1_6, 6_6_5, 4_2_8], [6_7_7, 4_1_6, 8_1_4, 4_2_9], [3_7_2, 4_3_5, 4_5_2, 4_5_0], [4_6_5, 4_3_4, 4_9_5, 4_4_7], [5_1_1, 4_3_4, 6_0_0, 4_4_7], [6_1_1, 4_3_6, 6_3_7, 4_4_7], [6_4_9, 4_3_6, 6_9_4, 4_5_1], [7_0_5, 4_3_8, 8_2_4, 4_4_7], [3_6_9, 4_5_3, 4_5_2, 4_6_6], [4_6_4, 4_5_4, 5_0_9, 4_6_6], [5_2_2, 4_5_3, 6_1_1, 4_6_9], [6_2_5, 4_5_3, 7_9_2, 4_6_9], [3_7_0, 4_7_2, 5_5_6, 4_8_8], [5_7_0, 4_7_2, 6_8_4, 4_8_7], [6_9_7, 4_7_2, 7_1_8, 4_8_5], [7_3_2, 4_7_2, 8_3_5, 4_8_8], [3_6_9, 4_9_0, 4_1_1, 5_0_3], [4_2_5, 4_9_0, 4_8_4, 5_0_3], [4_9_6, 4_9_0, 6_3_5, 5_0_6], [6_4_5, 4_9_0, 7_0_7, 5_0_3], [7_1_8, 4_9_1, 7_6_1, 5_0_3], [7_7_1, 4_9_0, 8_4_0, 5_0_3], [3_3_6, 5_1_0, 3_7_4, 5_2_1], [3_8_8, 5_1_0, 4_4_7, 5_2_2], [4_6_0, 5_1_0, 4_8_9, 5_2_1], [5_0_3, 5_1_0, 5_8_0, 5_2_2], [5_9_2, 5_0_9, 7_3_6, 5_2_5], [7_4_5, 5_0_9, 7_7_0, 5_2_2], [7_8_1, 5_0_9, 8_4_0, 5_2_2], [3_3_8, 5_2_8, 4_3_4, 5_4_1], [4_4_8, 5_2_8, 5_9_6, 5_4_1], [6_0_9, 5_2_7, 6_8_7, 5_4_0], [7_0_0, 5_2_8, 7_9_2, 5_4_1], [3_3_6, 5_4_6, 3_9_7, 5_5_9], [4_0_7, 5_4_6, 4_3_1, 5_5_9], [4_4_3, 5_4_6, 5_2_5, 5_6_0], [5_3_7, 5_4_6, 6_8_0, 5_6_2], [6_8_8, 5_4_6, 7_1_4, 5_5_9], [7_2_2, 5_4_6, 8_3_7, 5_6_2], [3_3_6, 5_6_5, 4_4_9, 5_8_1], [4_6_1, 5_6_5, 4_8_5, 5_7_7], [4_9_7, 5_6_5, 6_6_5, 5_8_1], [6_8_1, 5_6_5, 7_1_8, 5_7_7], [7_3_2, 5_6_5, 8_3_7, 5_8_0], [3_3_7, 5_8_4, 4_3_8, 5_9_7], [4_5_2, 5_8_3, 5_2_1, 5_9_6], [5_3_5, 5_8_4, 6_7_7, 5_9_9], [6_9_0, 5_8_3, 7_8_7, 5_9_6], [8_0_1, 5_8_3, 8_2_5, 5_9_6], [3_3_8, 6_0_2, 4_7_8, 6_1_5], [4_9_2, 6_0_2, 5_3_0, 6_1_4], [5_4_3, 6_0_2, 6_3_8, 6_1_5], [6_5_0, 6_0_2, 6_7_6, 6_1_4], [6_8_8, 6_0_2, 7_8_8, 6_1_5], [8_0_2, 6_0_2, 8_4_3, 6_1_4], [3_3_7, 6_2_1, 5_0_2, 6_3_3], [5_1_6, 6_2_1, 6_1_5, 6_3_7], [6_2_9, 6_2_1, 7_7_4, 6_3_6], [7_8_9, 6_2_1, 8_2_7, 6_3_3], [3_3_7, 6_3_9, 4_1_8, 6_5_2], [4_3_2, 6_4_0, 5_7_1, 6_5_3], [5_8_7, 6_3_9, 7_3_1, 6_5_5], [7_4_3, 6_3_9, 7_6_9, 6_5_2], [7_8_0, 6_3_9, 8_4_1, 6_5_2], [3_3_8, 6_5_8, 4_4_0, 6_7_3], [4_5_5, 6_5_8, 4_9_1, 6_7_0], [5_0_8, 6_5_8, 6_0_2, 6_7_1], [6_1_6, 6_5_8, 6_3_8, 6_7_0], [6_5_4, 6_5_8, 8_3_5, 6_7_4], [3_3_7, 6_7_7, 4_2_9, 6_8_9], [3_3_7, 7_1_4, 4_8_2, 7_2_6], [4_9_5, 7_1_4, 5_4_8, 7_2_6], [5_6_1, 7_1_4, 6_8_3, 7_2_6], [3_3_8, 7_7_0, 4_6_1, 7_8_2], [4_7_4, 7_6_9, 5_5_4, 7_8_5], [4_8_9, 7_8_8, 5_6_2, 8_0_3], [5_7_6, 7_8_8, 6_4_3, 8_0_1], [6_5_6, 7_8_7, 7_5_1, 8_0_4], [7_6_4, 7_8_8, 8_4_4, 8_0_1], [3_3_4, 8_2_5, 4_2_1, 8_3_8], [4_3_0, 8_2_4, 5_7_4, 8_3_8], [5_8_4, 8_2_4, 7_2_3, 8_4_1], [3_3_5, 8_4_4, 4_5_0, 8_5_7], [4_6_4, 8_4_3, 5_8_3, 8_6_0], [6_2_8, 8_6_2, 7_5_5, 8_7_5], [7_6_9, 8_6_1, 8_4_8, 8_7_8]]] # noqa: E231 # fmt: on self.assertListEqual(encoding.words ,__A ) self.assertListEqual(encoding.boxes ,__A ) # with apply_OCR = False __lowercase = LayoutLMvaImageProcessor(apply_ocr=__A ) __lowercase = image_processing(__A ,return_tensors='''pt''' ) self.assertEqual(encoding.pixel_values.shape ,(1, 3, 2_2_4, 2_2_4) )
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version('''>=''', '''4.25.0''')): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { 'xlm-roberta-base': 'https://huggingface.co/xlm-roberta-base/resolve/main/config.json', 'xlm-roberta-large': 'https://huggingface.co/xlm-roberta-large/resolve/main/config.json', 'xlm-roberta-large-finetuned-conll02-dutch': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll02-spanish': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll03-english': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json' ), 'xlm-roberta-large-finetuned-conll03-german': ( 'https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json' ), } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = '''xlm-roberta''' def __init__( self : List[str] ,lowercase__ : List[Any]=3_0_5_2_2 ,lowercase__ : str=7_6_8 ,lowercase__ : Optional[Any]=1_2 ,lowercase__ : List[str]=1_2 ,lowercase__ : List[str]=3_0_7_2 ,lowercase__ : Dict="gelu" ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : int=0.1 ,lowercase__ : Optional[int]=5_1_2 ,lowercase__ : Dict=2 ,lowercase__ : Union[str, Any]=0.0_2 ,lowercase__ : str=1e-1_2 ,lowercase__ : str=1 ,lowercase__ : Optional[int]=0 ,lowercase__ : List[str]=2 ,lowercase__ : int="absolute" ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Optional[Any] ,): super().__init__(pad_token_id=A__ ,bos_token_id=A__ ,eos_token_id=A__ ,**A__ ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = classifier_dropout class lowercase_ (lowerCamelCase__ ): """simple docstring""" @property def SCREAMING_SNAKE_CASE ( self : Any ): if self.task == "multiple-choice": __lowercase = {0: """batch""", 1: """choice""", 2: """sequence"""} else: __lowercase = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ('''input_ids''', dynamic_axis), ('''attention_mask''', dynamic_axis), ] )
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
from __future__ import annotations import math class lowercase_ : """simple docstring""" def __init__( self : List[Any] ,lowercase__ : Tuple ): __lowercase = size # approximate the overall size of segment tree with given value __lowercase = [0 for i in range(0 ,4 * size )] # create array to store lazy update __lowercase = [0 for i in range(0 ,4 * size )] __lowercase = [0 for i in range(0 ,4 * size )] # flag for lazy update def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[str] ): return idx * 2 def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ): return idx * 2 + 1 def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : int ,lowercase__ : Dict ,lowercase__ : Optional[int] ): if left_element == right_element: __lowercase = a[left_element - 1] else: __lowercase = (left_element + right_element) // 2 self.build(self.left(_a ) ,_a ,_a ,_a ) self.build(self.right(_a ) ,mid + 1 ,_a ,_a ) __lowercase = max( self.segment_tree[self.left(_a )] ,self.segment_tree[self.right(_a )] ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : Any ,lowercase__ : str ,lowercase__ : List[str] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ): if self.flag[idx] is True: __lowercase = self.lazy[idx] __lowercase = False if left_element != right_element: __lowercase = self.lazy[idx] __lowercase = self.lazy[idx] __lowercase = True __lowercase = True if right_element < a or left_element > b: return True if left_element >= a and right_element <= b: __lowercase = val if left_element != right_element: __lowercase = val __lowercase = val __lowercase = True __lowercase = True return True __lowercase = (left_element + right_element) // 2 self.update(self.left(_a ) ,_a ,_a ,_a ,_a ,_a ) self.update(self.right(_a ) ,mid + 1 ,_a ,_a ,_a ,_a ) __lowercase = max( self.segment_tree[self.left(_a )] ,self.segment_tree[self.right(_a )] ) return True def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[int] ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Any ): if self.flag[idx] is True: __lowercase = self.lazy[idx] __lowercase = False if left_element != right_element: __lowercase = self.lazy[idx] __lowercase = self.lazy[idx] __lowercase = True __lowercase = True if right_element < a or left_element > b: return -math.inf if left_element >= a and right_element <= b: return self.segment_tree[idx] __lowercase = (left_element + right_element) // 2 __lowercase = self.query(self.left(_a ) ,_a ,_a ,_a ,_a ) __lowercase = self.query(self.right(_a ) ,mid + 1 ,_a ,_a ,_a ) return max(_a ,_a ) def __str__( self : Any ): return str([self.query(1 ,1 ,self.size ,_a ,_a ) for i in range(1 ,self.size + 1 )] ) if __name__ == "__main__": lowerCAmelCase__ = [1, 2, -4, 7, 3, -5, 6, 11, -20, 9, 14, 15, 5, 2, -8] lowerCAmelCase__ = 15 lowerCAmelCase__ = SegmentTree(size) segt.build(1, 1, size, A) print(segt.query(1, 1, size, 4, 6)) print(segt.query(1, 1, size, 7, 11)) print(segt.query(1, 1, size, 7, 12)) segt.update(1, 1, size, 1, 3, 111) print(segt.query(1, 1, size, 1, 15)) segt.update(1, 1, size, 7, 8, 235) print(segt)
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ = { """configuration_mobilevit""": ["""MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """MobileViTConfig""", """MobileViTOnnxConfig"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ["""MobileViTFeatureExtractor"""] lowerCAmelCase__ = ["""MobileViTImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ """MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """MobileViTForImageClassification""", """MobileViTForSemanticSegmentation""", """MobileViTModel""", """MobileViTPreTrainedModel""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ """TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFMobileViTForImageClassification""", """TFMobileViTForSemanticSegmentation""", """TFMobileViTModel""", """TFMobileViTPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mobilevit import MOBILEVIT_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileViTConfig, MobileViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilevit import MobileViTFeatureExtractor from .image_processing_mobilevit import MobileViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilevit import ( MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTModel, MobileViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mobilevit import ( TF_MOBILEVIT_PRETRAINED_MODEL_ARCHIVE_LIST, TFMobileViTForImageClassification, TFMobileViTForSemanticSegmentation, TFMobileViTModel, TFMobileViTPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' import heapq import sys import numpy as np lowerCAmelCase__ = tuple[int, int] class lowercase_ : """simple docstring""" def __init__( self : Dict ): __lowercase = [] __lowercase = set() def SCREAMING_SNAKE_CASE ( self : str ): if not self.empty(): return self.elements[0][0] else: return float('''inf''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return len(self.elements ) == 0 def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : str ): if item not in self.set: heapq.heappush(self.elements ,(priority, item) ) self.set.add(lowercase__ ) else: # update # print("update", item) __lowercase = [] ((__lowercase) , (__lowercase)) = heapq.heappop(self.elements ) while x != item: temp.append((pri, x) ) ((__lowercase) , (__lowercase)) = heapq.heappop(self.elements ) temp.append((priority, item) ) for pro, xxx in temp: heapq.heappush(self.elements ,(pro, xxx) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Tuple ): if item in self.set: self.set.remove(lowercase__ ) __lowercase = [] ((__lowercase) , (__lowercase)) = heapq.heappop(self.elements ) while x != item: temp.append((pro, x) ) ((__lowercase) , (__lowercase)) = heapq.heappop(self.elements ) for prito, yyy in temp: heapq.heappush(self.elements ,(prito, yyy) ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return self.elements[0][1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ): ((__lowercase) , (__lowercase)) = heapq.heappop(self.elements ) self.set.remove(lowercase__ ) return (priority, item) def _A ( A__ , A__ ): """simple docstring""" __lowercase = np.array(A__ ) __lowercase = np.array(A__ ) return np.linalg.norm(a - b ) def _A ( A__ , A__ ): """simple docstring""" return consistent_heuristic(A__ , A__ ) // t def _A ( A__ , A__ ): """simple docstring""" return abs(p[0] - goal[0] ) + abs(p[1] - goal[1] ) def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = g_function[start] + Wa * heuristics[i](A__ , A__ ) return ans def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = np.chararray((n, n) ) for i in range(A__ ): for j in range(A__ ): __lowercase = '''*''' for i in range(A__ ): for j in range(A__ ): if (j, (n - 1) - i) in blocks: __lowercase = '''#''' __lowercase = '''-''' __lowercase = back_pointer[goal] while x != start: ((__lowercase) , (__lowercase)) = x # print(x) __lowercase = '''-''' __lowercase = back_pointer[x] __lowercase = '''-''' for i in range(A__ ): for j in range(A__ ): if (i, j) == (0, n - 1): print(grid[i][j] , end=''' ''' ) print('''<-- End position''' , end=''' ''' ) else: print(grid[i][j] , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) print('''PATH TAKEN BY THE ALGORITHM IS:-''' ) __lowercase = back_pointer[goal] while x != start: print(A__ , end=''' ''' ) __lowercase = back_pointer[x] print(A__ ) sys.exit() def _A ( A__ ): """simple docstring""" if p[0] < 0 or p[0] > n - 1: return False if p[1] < 0 or p[1] > n - 1: return False return True def _A ( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): """simple docstring""" for itera in range(A__ ): open_list[itera].remove_element(A__ ) # print("s", s) # print("j", j) ((__lowercase) , (__lowercase)) = s __lowercase = (x - 1, y) __lowercase = (x + 1, y) __lowercase = (x, y + 1) __lowercase = (x, y - 1) for neighbours in [left, right, up, down]: if neighbours not in blocks: if valid(A__ ) and neighbours not in visited: # print("neighbour", neighbours) visited.add(A__ ) __lowercase = -1 __lowercase = float('''inf''' ) if valid(A__ ) and g_function[neighbours] > g_function[s] + 1: __lowercase = g_function[s] + 1 __lowercase = s if neighbours not in close_list_anchor: open_list[0].put(A__ , key(A__ , 0 , A__ , A__ ) ) if neighbours not in close_list_inad: for var in range(1 , A__ ): if key(A__ , A__ , A__ , A__ ) <= Wa * key( A__ , 0 , A__ , A__ ): open_list[j].put( A__ , key(A__ , A__ , A__ , A__ ) ) def _A ( ): """simple docstring""" __lowercase = [] for x in range(1 , 5 ): for y in range(1 , 6 ): some_list.append((x, y) ) for x in range(15 , 20 ): some_list.append((x, 17) ) for x in range(10 , 19 ): for y in range(1 , 15 ): some_list.append((x, y) ) # L block for x in range(1 , 4 ): for y in range(12 , 19 ): some_list.append((x, y) ) for x in range(3 , 13 ): for y in range(16 , 19 ): some_list.append((x, y) ) return some_list lowerCAmelCase__ = {0: consistent_heuristic, 1: heuristic_a, 2: heuristic_a} lowerCAmelCase__ = [ (0, 1), (1, 1), (2, 1), (3, 1), (4, 1), (5, 1), (6, 1), (7, 1), (8, 1), (9, 1), (10, 1), (11, 1), (12, 1), (13, 1), (14, 1), (15, 1), (16, 1), (17, 1), (18, 1), (19, 1), ] lowerCAmelCase__ = make_common_ground() lowerCAmelCase__ = blocks_blk # hyper parameters lowerCAmelCase__ = 1 lowerCAmelCase__ = 1 lowerCAmelCase__ = 20 lowerCAmelCase__ = 3 # one consistent and two other inconsistent # start and end destination lowerCAmelCase__ = (0, 0) lowerCAmelCase__ = (n - 1, n - 1) lowerCAmelCase__ = 1 def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = {start: 0, goal: float('''inf''' )} __lowercase = {start: -1, goal: -1} __lowercase = [] __lowercase = set() for i in range(A__ ): open_list.append(PriorityQueue() ) open_list[i].put(A__ , key(A__ , A__ , A__ , A__ ) ) __lowercase = [] __lowercase = [] while open_list[0].minkey() < float('''inf''' ): for i in range(1 , A__ ): # print(open_list[0].minkey(), open_list[i].minkey()) if open_list[i].minkey() <= Wa * open_list[0].minkey(): global t t += 1 if g_function[goal] <= open_list[i].minkey(): if g_function[goal] < float('''inf''' ): do_something(A__ , A__ , A__ ) else: __lowercase , __lowercase = open_list[i].top_show() visited.add(A__ ) expand_state( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ) close_list_inad.append(A__ ) else: if g_function[goal] <= open_list[0].minkey(): if g_function[goal] < float('''inf''' ): do_something(A__ , A__ , A__ ) else: __lowercase = open_list[0].top_show() visited.add(A__ ) expand_state( A__ , 0 , A__ , A__ , A__ , A__ , A__ , A__ , ) close_list_anchor.append(A__ ) print('''No path found to goal''' ) print() for i in range(n - 1 , -1 , -1 ): for j in range(A__ ): if (j, i) in blocks: print('''#''' , end=''' ''' ) elif (j, i) in back_pointer: if (j, i) == (n - 1, n - 1): print('''*''' , end=''' ''' ) else: print('''-''' , end=''' ''' ) else: print('''*''' , end=''' ''' ) if (j, i) == (n - 1, n - 1): print('''<-- End position''' , end=''' ''' ) print() print('''^''' ) print('''Start position''' ) print() print('''# is an obstacle''' ) print('''- is the path taken by algorithm''' ) if __name__ == "__main__": multi_a_star(start, goal, n_heuristic)
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' import argparse import copy def _A ( A__ ): """simple docstring""" __lowercase = {} with open(SCREAMING_SNAKE_CASE_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: __lowercase = [] _list.append([line.split()[1], line.split()[2]] ) __lowercase = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: __lowercase = [] _list.append([line.split()[0], line.split()[2]] ) __lowercase = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def _A ( A__ , A__ ): """simple docstring""" with open(SCREAMING_SNAKE_CASE_ ) as f: __lowercase = f.read(1 ) __lowercase = start_node __lowercase = [] __lowercase = start_node __lowercase = 0 while visiting not in first_solution: __lowercase = 10000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(SCREAMING_SNAKE_CASE_ ) and k[0] not in first_solution: __lowercase = k[1] __lowercase = k[0] first_solution.append(SCREAMING_SNAKE_CASE_ ) __lowercase = distance_of_first_solution + int(SCREAMING_SNAKE_CASE_ ) __lowercase = best_node first_solution.append(SCREAMING_SNAKE_CASE_ ) __lowercase = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 __lowercase = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 10000 ) return first_solution, distance_of_first_solution def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] for n in solution[1:-1]: __lowercase = solution.index(SCREAMING_SNAKE_CASE_ ) for kn in solution[1:-1]: __lowercase = solution.index(SCREAMING_SNAKE_CASE_ ) if n == kn: continue __lowercase = copy.deepcopy(SCREAMING_SNAKE_CASE_ ) __lowercase = kn __lowercase = n __lowercase = 0 for k in _tmp[:-1]: __lowercase = _tmp[_tmp.index(SCREAMING_SNAKE_CASE_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: __lowercase = distance + int(i[1] ) _tmp.append(SCREAMING_SNAKE_CASE_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) __lowercase = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda A__ : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = 1 __lowercase = first_solution __lowercase = [] __lowercase = distance_of_first_solution __lowercase = solution while count <= iters: __lowercase = find_neighborhood(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) __lowercase = 0 __lowercase = neighborhood[index_of_best_solution] __lowercase = len(SCREAMING_SNAKE_CASE_ ) - 1 __lowercase = False while not found: __lowercase = 0 while i < len(SCREAMING_SNAKE_CASE_ ): if best_solution[i] != solution[i]: __lowercase = best_solution[i] __lowercase = solution[i] break __lowercase = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) __lowercase = True __lowercase = best_solution[:-1] __lowercase = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: __lowercase = cost __lowercase = solution else: __lowercase = index_of_best_solution + 1 __lowercase = neighborhood[index_of_best_solution] if len(SCREAMING_SNAKE_CASE_ ) >= size: tabu_list.pop(0 ) __lowercase = count + 1 return best_solution_ever, best_cost def _A ( A__=None ): """simple docstring""" __lowercase = generate_neighbours(args.File ) __lowercase = generate_first_solution( args.File , SCREAMING_SNAKE_CASE_ ) __lowercase = tabu_search( SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , args.Iterations , args.Size , ) print(F"Best solution: {best_sol}, with total distance: {best_cost}." ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser(description='''Tabu Search''') parser.add_argument( '''-f''', '''--File''', type=str, help='''Path to the file containing the data''', required=True, ) parser.add_argument( '''-i''', '''--Iterations''', type=int, help='''How many iterations the algorithm should perform''', required=True, ) parser.add_argument( '''-s''', '''--Size''', type=int, help='''Size of the tabu list''', required=True ) # Pass the arguments to main method main(parser.parse_args())
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCAmelCase__ = { '''configuration_roc_bert''': ['''ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''RoCBertConfig'''], '''tokenization_roc_bert''': ['''RoCBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: pass try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''RoCBertForCausalLM''', '''RoCBertForMaskedLM''', '''RoCBertForMultipleChoice''', '''RoCBertForPreTraining''', '''RoCBertForQuestionAnswering''', '''RoCBertForSequenceClassification''', '''RoCBertForTokenClassification''', '''RoCBertLayer''', '''RoCBertModel''', '''RoCBertPreTrainedModel''', '''load_tf_weights_in_roc_bert''', ] if TYPE_CHECKING: from .configuration_roc_bert import ROC_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RoCBertConfig from .tokenization_roc_bert import RoCBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: raise OptionalDependencyNotAvailable() try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roc_bert import ( ROC_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, RoCBertForCausalLM, RoCBertForMaskedLM, RoCBertForMultipleChoice, RoCBertForPreTraining, RoCBertForQuestionAnswering, RoCBertForSequenceClassification, RoCBertForTokenClassification, RoCBertLayer, RoCBertModel, RoCBertPreTrainedModel, load_tf_weights_in_roc_bert, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' from math import log from scipy.constants import Boltzmann, physical_constants lowerCAmelCase__ = 300 # TEMPERATURE (unit = K) def _A ( A__ , A__ , A__ , ): """simple docstring""" if donor_conc <= 0: raise ValueError('''Donor concentration should be positive''' ) elif acceptor_conc <= 0: raise ValueError('''Acceptor concentration should be positive''' ) elif intrinsic_conc <= 0: raise ValueError('''Intrinsic concentration should be positive''' ) elif donor_conc <= intrinsic_conc: raise ValueError( '''Donor concentration should be greater than intrinsic concentration''' ) elif acceptor_conc <= intrinsic_conc: raise ValueError( '''Acceptor concentration should be greater than intrinsic concentration''' ) else: return ( Boltzmann * T * log((donor_conc * acceptor_conc) / intrinsic_conc**2 ) / physical_constants["electron volt"][0] ) if __name__ == "__main__": import doctest doctest.testmod()
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_unispeech''': ['''UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''UniSpeechConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST''', '''UniSpeechForCTC''', '''UniSpeechForPreTraining''', '''UniSpeechForSequenceClassification''', '''UniSpeechModel''', '''UniSpeechPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_unispeech import UNISPEECH_PRETRAINED_CONFIG_ARCHIVE_MAP, UniSpeechConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_unispeech import ( UNISPEECH_PRETRAINED_MODEL_ARCHIVE_LIST, UniSpeechForCTC, UniSpeechForPreTraining, UniSpeechForSequenceClassification, UniSpeechModel, UniSpeechPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py lowerCAmelCase__ = '''src/diffusers''' lowerCAmelCase__ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. lowerCAmelCase__ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) lowerCAmelCase__ = spec.loader.load_module() def _A ( A__ , A__ ): """simple docstring""" return line.startswith(lowerCAmelCase__ ) or len(lowerCAmelCase__ ) <= 1 or re.search(R'''^\s*\)(\s*->.*:|:)\s*$''' , lowerCAmelCase__ ) is not None def _A ( A__ ): """simple docstring""" __lowercase = object_name.split('''.''' ) __lowercase = 0 # First let's find the module where our object lives. __lowercase = parts[i] while i < len(lowerCAmelCase__ ) and not os.path.isfile(os.path.join(lowerCAmelCase__ , F"{module}.py" ) ): i += 1 if i < len(lowerCAmelCase__ ): __lowercase = os.path.join(lowerCAmelCase__ , parts[i] ) if i >= len(lowerCAmelCase__ ): raise ValueError(F"`object_name` should begin with the name of a module of diffusers but got {object_name}." ) with open(os.path.join(lowerCAmelCase__ , F"{module}.py" ) , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: __lowercase = f.readlines() # Now let's find the class / func in the code! __lowercase = '''''' __lowercase = 0 for name in parts[i + 1 :]: while ( line_index < len(lowerCAmelCase__ ) and re.search(RF"^{indent}(class|def)\s+{name}(\(|\:)" , lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(lowerCAmelCase__ ): raise ValueError(F" {object_name} does not match any function or class in {module}." ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). __lowercase = line_index while line_index < len(lowerCAmelCase__ ) and _should_continue(lines[line_index] , lowerCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 __lowercase = lines[start_index:line_index] return "".join(lowerCAmelCase__ ) lowerCAmelCase__ = re.compile(R'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') lowerCAmelCase__ = re.compile(R'''^\s*(\S+)->(\S+)(\s+.*|$)''') lowerCAmelCase__ = re.compile(R'''<FILL\s+[^>]*>''') def _A ( A__ ): """simple docstring""" __lowercase = code.split('''\n''' ) __lowercase = 0 while idx < len(lowerCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(lowerCAmelCase__ ): return re.search(R'''^(\s*)\S''' , lines[idx] ).groups()[0] return "" def _A ( A__ ): """simple docstring""" __lowercase = len(get_indent(lowerCAmelCase__ ) ) > 0 if has_indent: __lowercase = F"class Bla:\n{code}" __lowercase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 , preview=lowerCAmelCase__ ) __lowercase = black.format_str(lowerCAmelCase__ , mode=lowerCAmelCase__ ) __lowercase , __lowercase = style_docstrings_in_code(lowerCAmelCase__ ) return result[len('''class Bla:\n''' ) :] if has_indent else result def _A ( A__ , A__=False ): """simple docstring""" with open(lowerCAmelCase__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: __lowercase = f.readlines() __lowercase = [] __lowercase = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(lowerCAmelCase__ ): __lowercase = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. __lowercase , __lowercase , __lowercase = search.groups() __lowercase = find_code_in_diffusers(lowerCAmelCase__ ) __lowercase = get_indent(lowerCAmelCase__ ) __lowercase = line_index + 1 if indent == theoretical_indent else line_index + 2 __lowercase = theoretical_indent __lowercase = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. __lowercase = True while line_index < len(lowerCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(lowerCAmelCase__ ): break __lowercase = lines[line_index] __lowercase = _should_continue(lowerCAmelCase__ , lowerCAmelCase__ ) and re.search(F"^{indent}# End copy" , lowerCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 __lowercase = lines[start_index:line_index] __lowercase = ''''''.join(lowerCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies __lowercase = [line for line in theoretical_code.split('''\n''' ) if _re_copy_warning.search(lowerCAmelCase__ ) is None] __lowercase = '''\n'''.join(lowerCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(lowerCAmelCase__ ) > 0: __lowercase = replace_pattern.replace('''with''' , '''''' ).split(''',''' ) __lowercase = [_re_replace_pattern.search(lowerCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue __lowercase , __lowercase , __lowercase = pattern.groups() __lowercase = re.sub(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ ) if option.strip() == "all-casing": __lowercase = re.sub(obja.lower() , obja.lower() , lowerCAmelCase__ ) __lowercase = re.sub(obja.upper() , obja.upper() , lowerCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line __lowercase = blackify(lines[start_index - 1] + theoretical_code ) __lowercase = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: __lowercase = lines[:start_index] + [theoretical_code] + lines[line_index:] __lowercase = start_index + 1 if overwrite and len(lowerCAmelCase__ ) > 0: # Warn the user a file has been modified. print(F"Detected changes, rewriting {filename}." ) with open(lowerCAmelCase__ , '''w''' , encoding='''utf-8''' , newline='''\n''' ) as f: f.writelines(lowerCAmelCase__ ) return diffs def _A ( A__ = False ): """simple docstring""" __lowercase = glob.glob(os.path.join(lowerCAmelCase__ , '''**/*.py''' ) , recursive=lowerCAmelCase__ ) __lowercase = [] for filename in all_files: __lowercase = is_copy_consistent(lowerCAmelCase__ , lowerCAmelCase__ ) diffs += [F"- {filename}: copy does not match {d[0]} at line {d[1]}" for d in new_diffs] if not overwrite and len(lowerCAmelCase__ ) > 0: __lowercase = '''\n'''.join(lowerCAmelCase__ ) raise Exception( '''Found the following copy inconsistencies:\n''' + diff + '''\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') lowerCAmelCase__ = parser.parse_args() check_copies(args.fix_and_overwrite)
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/unispeech-large-1500h-cv''': ( '''https://huggingface.co/microsoft/unispeech-large-1500h-cv/resolve/main/config.json''' ), # See all UniSpeech models at https://huggingface.co/models?filter=unispeech } class lowercase_ (__lowerCamelCase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = '''unispeech''' def __init__( self : Any ,lowercase__ : List[Any]=3_2 ,lowercase__ : str=7_6_8 ,lowercase__ : List[str]=1_2 ,lowercase__ : Any=1_2 ,lowercase__ : Any=3_0_7_2 ,lowercase__ : int="gelu" ,lowercase__ : str=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : List[Any]=0.0 ,lowercase__ : Dict=0.0 ,lowercase__ : str=0.1 ,lowercase__ : Optional[Any]=0.1 ,lowercase__ : str=0.0_2 ,lowercase__ : Optional[int]=1e-5 ,lowercase__ : List[str]="group" ,lowercase__ : Union[str, Any]="gelu" ,lowercase__ : Union[str, Any]=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) ,lowercase__ : Any=(5, 2, 2, 2, 2, 2, 2) ,lowercase__ : List[str]=(1_0, 3, 3, 3, 3, 2, 2) ,lowercase__ : int=False ,lowercase__ : List[str]=1_2_8 ,lowercase__ : Union[str, Any]=1_6 ,lowercase__ : Optional[Any]=False ,lowercase__ : Tuple=True ,lowercase__ : Union[str, Any]=0.0_5 ,lowercase__ : Any=1_0 ,lowercase__ : str=2 ,lowercase__ : Tuple=0.0 ,lowercase__ : Optional[int]=1_0 ,lowercase__ : int=0 ,lowercase__ : int=3_2_0 ,lowercase__ : Union[str, Any]=2 ,lowercase__ : int=0.1 ,lowercase__ : Optional[Any]=1_0_0 ,lowercase__ : Union[str, Any]=2_5_6 ,lowercase__ : Union[str, Any]=2_5_6 ,lowercase__ : Optional[Any]=0.1 ,lowercase__ : str="mean" ,lowercase__ : Optional[int]=False ,lowercase__ : Any=False ,lowercase__ : int=2_5_6 ,lowercase__ : Union[str, Any]=8_0 ,lowercase__ : Dict=0 ,lowercase__ : Optional[int]=1 ,lowercase__ : Any=2 ,lowercase__ : Optional[int]=0.5 ,**lowercase__ : Union[str, Any] ,): super().__init__(**a_ ,pad_token_id=a_ ,bos_token_id=a_ ,eos_token_id=a_ ) __lowercase = hidden_size __lowercase = feat_extract_norm __lowercase = feat_extract_activation __lowercase = list(a_ ) __lowercase = list(a_ ) __lowercase = list(a_ ) __lowercase = conv_bias __lowercase = num_conv_pos_embeddings __lowercase = num_conv_pos_embedding_groups __lowercase = len(self.conv_dim ) __lowercase = num_hidden_layers __lowercase = intermediate_size __lowercase = hidden_act __lowercase = num_attention_heads __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = feat_proj_dropout __lowercase = final_dropout __lowercase = layerdrop __lowercase = layer_norm_eps __lowercase = initializer_range __lowercase = num_ctc_classes __lowercase = vocab_size __lowercase = do_stable_layer_norm __lowercase = use_weighted_layer_sum __lowercase = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( '''Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==''' ''' `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =''' F" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`," F" `len(config.conv_kernel) = {len(self.conv_kernel )}`." ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 __lowercase = apply_spec_augment __lowercase = mask_time_prob __lowercase = mask_time_length __lowercase = mask_time_min_masks __lowercase = mask_feature_prob __lowercase = mask_feature_length __lowercase = mask_feature_min_masks # parameters for pretraining with codevector quantized representations __lowercase = num_codevectors_per_group __lowercase = num_codevector_groups __lowercase = contrastive_logits_temperature __lowercase = feat_quantizer_dropout __lowercase = num_negatives __lowercase = codevector_dim __lowercase = proj_codevector_dim __lowercase = diversity_loss_weight # ctc loss __lowercase = ctc_loss_reduction __lowercase = ctc_zero_infinity # pretraining loss __lowercase = replace_prob @property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return functools.reduce(operator.mul ,self.conv_stride ,1 )
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
import argparse import json import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ConvNextConfig, SegformerImageProcessor, UperNetConfig, UperNetForSemanticSegmentation def _A ( A__ ): """simple docstring""" __lowercase = 384 if "tiny" in model_name: __lowercase = [3, 3, 9, 3] __lowercase = [96, 192, 384, 768] if "small" in model_name: __lowercase = [3, 3, 27, 3] __lowercase = [96, 192, 384, 768] if "base" in model_name: __lowercase = [3, 3, 27, 3] __lowercase = [128, 256, 512, 1024] __lowercase = 512 if "large" in model_name: __lowercase = [3, 3, 27, 3] __lowercase = [192, 384, 768, 1536] __lowercase = 768 if "xlarge" in model_name: __lowercase = [3, 3, 27, 3] __lowercase = [256, 512, 1024, 2048] __lowercase = 1024 # set label information __lowercase = 150 __lowercase = '''huggingface/label-files''' __lowercase = '''ade20k-id2label.json''' __lowercase = json.load(open(hf_hub_download(_snake_case , _snake_case , repo_type='''dataset''' ) , '''r''' ) ) __lowercase = {int(_snake_case ): v for k, v in idalabel.items()} __lowercase = {v: k for k, v in idalabel.items()} __lowercase = ConvNextConfig( depths=_snake_case , hidden_sizes=_snake_case , out_features=['''stage1''', '''stage2''', '''stage3''', '''stage4'''] ) __lowercase = UperNetConfig( backbone_config=_snake_case , auxiliary_in_channels=_snake_case , num_labels=_snake_case , idalabel=_snake_case , labelaid=_snake_case , ) return config def _A ( A__ ): """simple docstring""" __lowercase = [] # fmt: off # stem rename_keys.append(('''backbone.downsample_layers.0.0.weight''', '''backbone.embeddings.patch_embeddings.weight''') ) rename_keys.append(('''backbone.downsample_layers.0.0.bias''', '''backbone.embeddings.patch_embeddings.bias''') ) rename_keys.append(('''backbone.downsample_layers.0.1.weight''', '''backbone.embeddings.layernorm.weight''') ) rename_keys.append(('''backbone.downsample_layers.0.1.bias''', '''backbone.embeddings.layernorm.bias''') ) # stages for i in range(len(config.backbone_config.depths ) ): for j in range(config.backbone_config.depths[i] ): rename_keys.append((F"backbone.stages.{i}.{j}.gamma", F"backbone.encoder.stages.{i}.layers.{j}.layer_scale_parameter") ) rename_keys.append((F"backbone.stages.{i}.{j}.depthwise_conv.weight", F"backbone.encoder.stages.{i}.layers.{j}.dwconv.weight") ) rename_keys.append((F"backbone.stages.{i}.{j}.depthwise_conv.bias", F"backbone.encoder.stages.{i}.layers.{j}.dwconv.bias") ) rename_keys.append((F"backbone.stages.{i}.{j}.norm.weight", F"backbone.encoder.stages.{i}.layers.{j}.layernorm.weight") ) rename_keys.append((F"backbone.stages.{i}.{j}.norm.bias", F"backbone.encoder.stages.{i}.layers.{j}.layernorm.bias") ) rename_keys.append((F"backbone.stages.{i}.{j}.pointwise_conv1.weight", F"backbone.encoder.stages.{i}.layers.{j}.pwconv1.weight") ) rename_keys.append((F"backbone.stages.{i}.{j}.pointwise_conv1.bias", F"backbone.encoder.stages.{i}.layers.{j}.pwconv1.bias") ) rename_keys.append((F"backbone.stages.{i}.{j}.pointwise_conv2.weight", F"backbone.encoder.stages.{i}.layers.{j}.pwconv2.weight") ) rename_keys.append((F"backbone.stages.{i}.{j}.pointwise_conv2.bias", F"backbone.encoder.stages.{i}.layers.{j}.pwconv2.bias") ) if i > 0: rename_keys.append((F"backbone.downsample_layers.{i}.0.weight", F"backbone.encoder.stages.{i}.downsampling_layer.0.weight") ) rename_keys.append((F"backbone.downsample_layers.{i}.0.bias", F"backbone.encoder.stages.{i}.downsampling_layer.0.bias") ) rename_keys.append((F"backbone.downsample_layers.{i}.1.weight", F"backbone.encoder.stages.{i}.downsampling_layer.1.weight") ) rename_keys.append((F"backbone.downsample_layers.{i}.1.bias", F"backbone.encoder.stages.{i}.downsampling_layer.1.bias") ) rename_keys.append((F"backbone.norm{i}.weight", F"backbone.hidden_states_norms.stage{i+1}.weight") ) rename_keys.append((F"backbone.norm{i}.bias", F"backbone.hidden_states_norms.stage{i+1}.bias") ) # decode head rename_keys.extend( [ ('''decode_head.conv_seg.weight''', '''decode_head.classifier.weight'''), ('''decode_head.conv_seg.bias''', '''decode_head.classifier.bias'''), ('''auxiliary_head.conv_seg.weight''', '''auxiliary_head.classifier.weight'''), ('''auxiliary_head.conv_seg.bias''', '''auxiliary_head.classifier.bias'''), ] ) # fmt: on return rename_keys def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = dct.pop(_snake_case ) __lowercase = val def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = { '''upernet-convnext-tiny''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_tiny_fp16_512x512_160k_ade20k/upernet_convnext_tiny_fp16_512x512_160k_ade20k_20220227_124553-cad485de.pth''', '''upernet-convnext-small''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_small_fp16_512x512_160k_ade20k/upernet_convnext_small_fp16_512x512_160k_ade20k_20220227_131208-1b1e394f.pth''', '''upernet-convnext-base''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_base_fp16_512x512_160k_ade20k/upernet_convnext_base_fp16_512x512_160k_ade20k_20220227_181227-02a24fc6.pth''', '''upernet-convnext-large''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_large_fp16_640x640_160k_ade20k/upernet_convnext_large_fp16_640x640_160k_ade20k_20220226_040532-e57aa54d.pth''', '''upernet-convnext-xlarge''': '''https://download.openmmlab.com/mmsegmentation/v0.5/convnext/upernet_convnext_xlarge_fp16_640x640_160k_ade20k/upernet_convnext_xlarge_fp16_640x640_160k_ade20k_20220226_080344-95fc38c2.pth''', } __lowercase = model_name_to_url[model_name] __lowercase = torch.hub.load_state_dict_from_url(_snake_case , map_location='''cpu''' )['''state_dict'''] __lowercase = get_upernet_config(_snake_case ) __lowercase = UperNetForSemanticSegmentation(_snake_case ) model.eval() # replace "bn" => "batch_norm" for key in state_dict.copy().keys(): __lowercase = state_dict.pop(_snake_case ) if "bn" in key: __lowercase = key.replace('''bn''' , '''batch_norm''' ) __lowercase = val # rename keys __lowercase = create_rename_keys(_snake_case ) for src, dest in rename_keys: rename_key(_snake_case , _snake_case , _snake_case ) model.load_state_dict(_snake_case ) # verify on image __lowercase = '''https://huggingface.co/datasets/hf-internal-testing/fixtures_ade20k/resolve/main/ADE_val_00000001.jpg''' __lowercase = Image.open(requests.get(_snake_case , stream=_snake_case ).raw ).convert('''RGB''' ) __lowercase = SegformerImageProcessor() __lowercase = processor(_snake_case , return_tensors='''pt''' ).pixel_values with torch.no_grad(): __lowercase = model(_snake_case ) if model_name == "upernet-convnext-tiny": __lowercase = torch.tensor( [[-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.8_1_1_0, -8.8_1_1_0, -8.6_5_2_1], [-8.7_7_4_6, -8.7_7_4_6, -8.6_1_3_0]] ) elif model_name == "upernet-convnext-small": __lowercase = torch.tensor( [[-8.8_2_3_6, -8.8_2_3_6, -8.6_7_7_1], [-8.8_2_3_6, -8.8_2_3_6, -8.6_7_7_1], [-8.7_6_3_8, -8.7_6_3_8, -8.6_2_4_0]] ) elif model_name == "upernet-convnext-base": __lowercase = torch.tensor( [[-8.8_5_5_8, -8.8_5_5_8, -8.6_9_0_5], [-8.8_5_5_8, -8.8_5_5_8, -8.6_9_0_5], [-8.7_6_6_9, -8.7_6_6_9, -8.6_0_2_1]] ) elif model_name == "upernet-convnext-large": __lowercase = torch.tensor( [[-8.6_6_6_0, -8.6_6_6_0, -8.6_2_1_0], [-8.6_6_6_0, -8.6_6_6_0, -8.6_2_1_0], [-8.6_3_1_0, -8.6_3_1_0, -8.5_9_6_4]] ) elif model_name == "upernet-convnext-xlarge": __lowercase = torch.tensor( [[-8.4_9_8_0, -8.4_9_8_0, -8.3_9_7_7], [-8.4_9_8_0, -8.4_9_8_0, -8.3_9_7_7], [-8.4_3_7_9, -8.4_3_7_9, -8.3_4_1_2]] ) print('''Logits:''' , outputs.logits[0, 0, :3, :3] ) assert torch.allclose(outputs.logits[0, 0, :3, :3] , _snake_case , atol=1e-4 ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: print(F"Saving model {model_name} to {pytorch_dump_folder_path}" ) model.save_pretrained(_snake_case ) print(F"Saving processor to {pytorch_dump_folder_path}" ) processor.save_pretrained(_snake_case ) if push_to_hub: print(F"Pushing model and processor for {model_name} to hub" ) model.push_to_hub(F"openmmlab/{model_name}" ) processor.push_to_hub(F"openmmlab/{model_name}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''upernet-convnext-tiny''', type=str, choices=[f'upernet-convnext-{size}' for size in ['''tiny''', '''small''', '''base''', '''large''', '''xlarge''']], help='''Name of the ConvNext UperNet model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCAmelCase__ = parser.parse_args() convert_upernet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import string # frequency taken from https://en.wikipedia.org/wiki/Letter_frequency lowerCAmelCase__ = { 'E': 12.70, 'T': 9.06, 'A': 8.17, 'O': 7.51, 'I': 6.97, 'N': 6.75, 'S': 6.33, 'H': 6.09, 'R': 5.99, 'D': 4.25, 'L': 4.03, 'C': 2.78, 'U': 2.76, 'M': 2.41, 'W': 2.36, 'F': 2.23, 'G': 2.02, 'Y': 1.97, 'P': 1.93, 'B': 1.29, 'V': 0.98, 'K': 0.77, 'J': 0.15, 'X': 0.15, 'Q': 0.10, 'Z': 0.07, } lowerCAmelCase__ = 'ETAOINSHRDLCUMWFGYPBVKJXQZ' lowerCAmelCase__ = 'ABCDEFGHIJKLMNOPQRSTUVWXYZ' def _A ( A__ ): """simple docstring""" __lowercase = {letter: 0 for letter in string.ascii_uppercase} for letter in message.upper(): if letter in LETTERS: letter_count[letter] += 1 return letter_count def _A ( A__ ): """simple docstring""" return x[0] def _A ( A__ ): """simple docstring""" __lowercase = get_letter_count(_A ) __lowercase = { freq: [] for letter, freq in letter_to_freq.items() } for letter in LETTERS: freq_to_letter[letter_to_freq[letter]].append(_A ) __lowercase = {} for freq in freq_to_letter: freq_to_letter[freq].sort(key=ETAOIN.find , reverse=_A ) __lowercase = ''''''.join(freq_to_letter[freq] ) __lowercase = list(freq_to_letter_str.items() ) freq_pairs.sort(key=_A , reverse=_A ) __lowercase = [freq_pair[1] for freq_pair in freq_pairs] return "".join(_A ) def _A ( A__ ): """simple docstring""" __lowercase = get_frequency_order(_A ) __lowercase = 0 for common_letter in ETAOIN[:6]: if common_letter in freq_order[:6]: match_score += 1 for uncommon_letter in ETAOIN[-6:]: if uncommon_letter in freq_order[-6:]: match_score += 1 return match_score if __name__ == "__main__": import doctest doctest.testmod()
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowerCAmelCase__ = False class lowercase_ (unittest.TestCase ): """simple docstring""" pass @nightly @require_torch_gpu class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().tearDown() gc.collect() torch.cuda.empty_cache() def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = VersatileDiffusionTextToImagePipeline.from_pretrained('''shi-labs/versatile-diffusion''' ) # remove text_unet pipe.remove_unused_weights() pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowercase = '''A painting of a squirrel eating a burger ''' __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=UpperCamelCase__ ,generator=UpperCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='''numpy''' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(UpperCamelCase__ ) __lowercase = VersatileDiffusionTextToImagePipeline.from_pretrained(UpperCamelCase__ ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowercase = generator.manual_seed(0 ) __lowercase = pipe( prompt=UpperCamelCase__ ,generator=UpperCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=2 ,output_type='''numpy''' ).images assert np.abs(image - new_image ).sum() < 1e-5, "Models don't have the same forward pass" def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = VersatileDiffusionTextToImagePipeline.from_pretrained( '''shi-labs/versatile-diffusion''' ,torch_dtype=torch.floataa ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) __lowercase = '''A painting of a squirrel eating a burger ''' __lowercase = torch.manual_seed(0 ) __lowercase = pipe( prompt=UpperCamelCase__ ,generator=UpperCamelCase__ ,guidance_scale=7.5 ,num_inference_steps=5_0 ,output_type='''numpy''' ).images __lowercase = image[0, 2_5_3:2_5_6, 2_5_3:2_5_6, -1] assert image.shape == (1, 5_1_2, 5_1_2, 3) __lowercase = np.array([0.3_3_6_7, 0.3_1_6_9, 0.2_6_5_6, 0.3_8_7_0, 0.4_7_9_0, 0.3_7_9_6, 0.4_0_0_9, 0.4_8_7_8, 0.4_7_7_8] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/conditional-detr-resnet-50''': ( '''https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json''' ), } class lowercase_ (snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE : str = '''conditional_detr''' SCREAMING_SNAKE_CASE : int = ['''past_key_values'''] SCREAMING_SNAKE_CASE : Tuple = { '''hidden_size''': '''d_model''', '''num_attention_heads''': '''encoder_attention_heads''', } def __init__( self : Optional[Any] ,lowercase__ : str=True ,lowercase__ : Optional[int]=None ,lowercase__ : int=3 ,lowercase__ : Any=3_0_0 ,lowercase__ : List[str]=6 ,lowercase__ : str=2_0_4_8 ,lowercase__ : Optional[int]=8 ,lowercase__ : str=6 ,lowercase__ : str=2_0_4_8 ,lowercase__ : str=8 ,lowercase__ : str=0.0 ,lowercase__ : str=0.0 ,lowercase__ : int=True ,lowercase__ : str="relu" ,lowercase__ : Dict=2_5_6 ,lowercase__ : str=0.1 ,lowercase__ : int=0.0 ,lowercase__ : Tuple=0.0 ,lowercase__ : Any=0.0_2 ,lowercase__ : Tuple=1.0 ,lowercase__ : Tuple=False ,lowercase__ : Union[str, Any]="sine" ,lowercase__ : Union[str, Any]="resnet50" ,lowercase__ : int=True ,lowercase__ : str=False ,lowercase__ : List[str]=2 ,lowercase__ : List[Any]=5 ,lowercase__ : Dict=2 ,lowercase__ : Tuple=1 ,lowercase__ : List[str]=1 ,lowercase__ : Optional[Any]=2 ,lowercase__ : Optional[int]=5 ,lowercase__ : Tuple=2 ,lowercase__ : Optional[int]=0.2_5 ,**lowercase__ : int ,): if backbone_config is not None and use_timm_backbone: raise ValueError('''You can\'t specify both `backbone_config` and `use_timm_backbone`.''' ) if not use_timm_backbone: if backbone_config is None: logger.info('''`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.''' ) __lowercase = CONFIG_MAPPING['resnet'](out_features=['''stage4'''] ) elif isinstance(_A ,_A ): __lowercase = backbone_config.get('''model_type''' ) __lowercase = CONFIG_MAPPING[backbone_model_type] __lowercase = config_class.from_dict(_A ) __lowercase = use_timm_backbone __lowercase = backbone_config __lowercase = num_channels __lowercase = num_queries __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = init_xavier_std __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = encoder_layers __lowercase = auxiliary_loss __lowercase = position_embedding_type __lowercase = backbone __lowercase = use_pretrained_backbone __lowercase = dilation # Hungarian matcher __lowercase = class_cost __lowercase = bbox_cost __lowercase = giou_cost # Loss coefficients __lowercase = mask_loss_coefficient __lowercase = dice_loss_coefficient __lowercase = cls_loss_coefficient __lowercase = bbox_loss_coefficient __lowercase = giou_loss_coefficient __lowercase = focal_alpha super().__init__(is_encoder_decoder=_A ,**_A ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return self.encoder_attention_heads @property def SCREAMING_SNAKE_CASE ( self : int ): return self.d_model def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: __lowercase = self.backbone_config.to_dict() __lowercase = self.__class__.model_type return output class lowercase_ (snake_case__ ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = version.parse('1.11' ) @property def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): return OrderedDict( [ ('''pixel_values''', {0: '''batch''', 1: '''num_channels''', 2: '''height''', 3: '''width'''}), ('''pixel_mask''', {0: '''batch'''}), ] ) @property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return 1e-5 @property def SCREAMING_SNAKE_CASE ( self : str ): return 1_2
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool lowerCAmelCase__ = { '''Acehnese Arabic''': '''ace_Arab''', '''Acehnese Latin''': '''ace_Latn''', '''Mesopotamian Arabic''': '''acm_Arab''', '''Ta\'izzi-Adeni Arabic''': '''acq_Arab''', '''Tunisian Arabic''': '''aeb_Arab''', '''Afrikaans''': '''afr_Latn''', '''South Levantine Arabic''': '''ajp_Arab''', '''Akan''': '''aka_Latn''', '''Amharic''': '''amh_Ethi''', '''North Levantine Arabic''': '''apc_Arab''', '''Modern Standard Arabic''': '''arb_Arab''', '''Modern Standard Arabic Romanized''': '''arb_Latn''', '''Najdi Arabic''': '''ars_Arab''', '''Moroccan Arabic''': '''ary_Arab''', '''Egyptian Arabic''': '''arz_Arab''', '''Assamese''': '''asm_Beng''', '''Asturian''': '''ast_Latn''', '''Awadhi''': '''awa_Deva''', '''Central Aymara''': '''ayr_Latn''', '''South Azerbaijani''': '''azb_Arab''', '''North Azerbaijani''': '''azj_Latn''', '''Bashkir''': '''bak_Cyrl''', '''Bambara''': '''bam_Latn''', '''Balinese''': '''ban_Latn''', '''Belarusian''': '''bel_Cyrl''', '''Bemba''': '''bem_Latn''', '''Bengali''': '''ben_Beng''', '''Bhojpuri''': '''bho_Deva''', '''Banjar Arabic''': '''bjn_Arab''', '''Banjar Latin''': '''bjn_Latn''', '''Standard Tibetan''': '''bod_Tibt''', '''Bosnian''': '''bos_Latn''', '''Buginese''': '''bug_Latn''', '''Bulgarian''': '''bul_Cyrl''', '''Catalan''': '''cat_Latn''', '''Cebuano''': '''ceb_Latn''', '''Czech''': '''ces_Latn''', '''Chokwe''': '''cjk_Latn''', '''Central Kurdish''': '''ckb_Arab''', '''Crimean Tatar''': '''crh_Latn''', '''Welsh''': '''cym_Latn''', '''Danish''': '''dan_Latn''', '''German''': '''deu_Latn''', '''Southwestern Dinka''': '''dik_Latn''', '''Dyula''': '''dyu_Latn''', '''Dzongkha''': '''dzo_Tibt''', '''Greek''': '''ell_Grek''', '''English''': '''eng_Latn''', '''Esperanto''': '''epo_Latn''', '''Estonian''': '''est_Latn''', '''Basque''': '''eus_Latn''', '''Ewe''': '''ewe_Latn''', '''Faroese''': '''fao_Latn''', '''Fijian''': '''fij_Latn''', '''Finnish''': '''fin_Latn''', '''Fon''': '''fon_Latn''', '''French''': '''fra_Latn''', '''Friulian''': '''fur_Latn''', '''Nigerian Fulfulde''': '''fuv_Latn''', '''Scottish Gaelic''': '''gla_Latn''', '''Irish''': '''gle_Latn''', '''Galician''': '''glg_Latn''', '''Guarani''': '''grn_Latn''', '''Gujarati''': '''guj_Gujr''', '''Haitian Creole''': '''hat_Latn''', '''Hausa''': '''hau_Latn''', '''Hebrew''': '''heb_Hebr''', '''Hindi''': '''hin_Deva''', '''Chhattisgarhi''': '''hne_Deva''', '''Croatian''': '''hrv_Latn''', '''Hungarian''': '''hun_Latn''', '''Armenian''': '''hye_Armn''', '''Igbo''': '''ibo_Latn''', '''Ilocano''': '''ilo_Latn''', '''Indonesian''': '''ind_Latn''', '''Icelandic''': '''isl_Latn''', '''Italian''': '''ita_Latn''', '''Javanese''': '''jav_Latn''', '''Japanese''': '''jpn_Jpan''', '''Kabyle''': '''kab_Latn''', '''Jingpho''': '''kac_Latn''', '''Kamba''': '''kam_Latn''', '''Kannada''': '''kan_Knda''', '''Kashmiri Arabic''': '''kas_Arab''', '''Kashmiri Devanagari''': '''kas_Deva''', '''Georgian''': '''kat_Geor''', '''Central Kanuri Arabic''': '''knc_Arab''', '''Central Kanuri Latin''': '''knc_Latn''', '''Kazakh''': '''kaz_Cyrl''', '''Kabiyè''': '''kbp_Latn''', '''Kabuverdianu''': '''kea_Latn''', '''Khmer''': '''khm_Khmr''', '''Kikuyu''': '''kik_Latn''', '''Kinyarwanda''': '''kin_Latn''', '''Kyrgyz''': '''kir_Cyrl''', '''Kimbundu''': '''kmb_Latn''', '''Northern Kurdish''': '''kmr_Latn''', '''Kikongo''': '''kon_Latn''', '''Korean''': '''kor_Hang''', '''Lao''': '''lao_Laoo''', '''Ligurian''': '''lij_Latn''', '''Limburgish''': '''lim_Latn''', '''Lingala''': '''lin_Latn''', '''Lithuanian''': '''lit_Latn''', '''Lombard''': '''lmo_Latn''', '''Latgalian''': '''ltg_Latn''', '''Luxembourgish''': '''ltz_Latn''', '''Luba-Kasai''': '''lua_Latn''', '''Ganda''': '''lug_Latn''', '''Luo''': '''luo_Latn''', '''Mizo''': '''lus_Latn''', '''Standard Latvian''': '''lvs_Latn''', '''Magahi''': '''mag_Deva''', '''Maithili''': '''mai_Deva''', '''Malayalam''': '''mal_Mlym''', '''Marathi''': '''mar_Deva''', '''Minangkabau Arabic ''': '''min_Arab''', '''Minangkabau Latin''': '''min_Latn''', '''Macedonian''': '''mkd_Cyrl''', '''Plateau Malagasy''': '''plt_Latn''', '''Maltese''': '''mlt_Latn''', '''Meitei Bengali''': '''mni_Beng''', '''Halh Mongolian''': '''khk_Cyrl''', '''Mossi''': '''mos_Latn''', '''Maori''': '''mri_Latn''', '''Burmese''': '''mya_Mymr''', '''Dutch''': '''nld_Latn''', '''Norwegian Nynorsk''': '''nno_Latn''', '''Norwegian Bokmål''': '''nob_Latn''', '''Nepali''': '''npi_Deva''', '''Northern Sotho''': '''nso_Latn''', '''Nuer''': '''nus_Latn''', '''Nyanja''': '''nya_Latn''', '''Occitan''': '''oci_Latn''', '''West Central Oromo''': '''gaz_Latn''', '''Odia''': '''ory_Orya''', '''Pangasinan''': '''pag_Latn''', '''Eastern Panjabi''': '''pan_Guru''', '''Papiamento''': '''pap_Latn''', '''Western Persian''': '''pes_Arab''', '''Polish''': '''pol_Latn''', '''Portuguese''': '''por_Latn''', '''Dari''': '''prs_Arab''', '''Southern Pashto''': '''pbt_Arab''', '''Ayacucho Quechua''': '''quy_Latn''', '''Romanian''': '''ron_Latn''', '''Rundi''': '''run_Latn''', '''Russian''': '''rus_Cyrl''', '''Sango''': '''sag_Latn''', '''Sanskrit''': '''san_Deva''', '''Santali''': '''sat_Olck''', '''Sicilian''': '''scn_Latn''', '''Shan''': '''shn_Mymr''', '''Sinhala''': '''sin_Sinh''', '''Slovak''': '''slk_Latn''', '''Slovenian''': '''slv_Latn''', '''Samoan''': '''smo_Latn''', '''Shona''': '''sna_Latn''', '''Sindhi''': '''snd_Arab''', '''Somali''': '''som_Latn''', '''Southern Sotho''': '''sot_Latn''', '''Spanish''': '''spa_Latn''', '''Tosk Albanian''': '''als_Latn''', '''Sardinian''': '''srd_Latn''', '''Serbian''': '''srp_Cyrl''', '''Swati''': '''ssw_Latn''', '''Sundanese''': '''sun_Latn''', '''Swedish''': '''swe_Latn''', '''Swahili''': '''swh_Latn''', '''Silesian''': '''szl_Latn''', '''Tamil''': '''tam_Taml''', '''Tatar''': '''tat_Cyrl''', '''Telugu''': '''tel_Telu''', '''Tajik''': '''tgk_Cyrl''', '''Tagalog''': '''tgl_Latn''', '''Thai''': '''tha_Thai''', '''Tigrinya''': '''tir_Ethi''', '''Tamasheq Latin''': '''taq_Latn''', '''Tamasheq Tifinagh''': '''taq_Tfng''', '''Tok Pisin''': '''tpi_Latn''', '''Tswana''': '''tsn_Latn''', '''Tsonga''': '''tso_Latn''', '''Turkmen''': '''tuk_Latn''', '''Tumbuka''': '''tum_Latn''', '''Turkish''': '''tur_Latn''', '''Twi''': '''twi_Latn''', '''Central Atlas Tamazight''': '''tzm_Tfng''', '''Uyghur''': '''uig_Arab''', '''Ukrainian''': '''ukr_Cyrl''', '''Umbundu''': '''umb_Latn''', '''Urdu''': '''urd_Arab''', '''Northern Uzbek''': '''uzn_Latn''', '''Venetian''': '''vec_Latn''', '''Vietnamese''': '''vie_Latn''', '''Waray''': '''war_Latn''', '''Wolof''': '''wol_Latn''', '''Xhosa''': '''xho_Latn''', '''Eastern Yiddish''': '''ydd_Hebr''', '''Yoruba''': '''yor_Latn''', '''Yue Chinese''': '''yue_Hant''', '''Chinese Simplified''': '''zho_Hans''', '''Chinese Traditional''': '''zho_Hant''', '''Standard Malay''': '''zsm_Latn''', '''Zulu''': '''zul_Latn''', } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = 'facebook/nllb-200-distilled-600M' SCREAMING_SNAKE_CASE : Optional[int] = ( 'This is a tool that translates text from a language to another. It takes three inputs: `text`, which should ' 'be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, ' 'which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in ' 'plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.' ) SCREAMING_SNAKE_CASE : Union[str, Any] = 'translator' SCREAMING_SNAKE_CASE : List[str] = AutoTokenizer SCREAMING_SNAKE_CASE : Optional[Any] = AutoModelForSeqaSeqLM SCREAMING_SNAKE_CASE : List[str] = LANGUAGE_CODES SCREAMING_SNAKE_CASE : Tuple = ['text', 'text', 'text'] SCREAMING_SNAKE_CASE : int = ['text'] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : Dict ,lowercase__ : List[Any] ): if src_lang not in self.lang_to_code: raise ValueError(F"{src_lang} is not a supported language." ) if tgt_lang not in self.lang_to_code: raise ValueError(F"{tgt_lang} is not a supported language." ) __lowercase = self.lang_to_code[src_lang] __lowercase = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( __a ,return_tensors='''pt''' ,src_lang=__a ,tgt_lang=__a ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Union[str, Any] ): return self.model.generate(**__a ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ): return self.post_processor.decode(outputs[0].tolist() ,skip_special_tokens=__a )
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from __future__ import annotations lowerCAmelCase__ = 10 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = max(lowerCamelCase_ ) while placement <= max_digit: # declare and initialize empty buckets __lowercase = [[] for _ in range(lowerCamelCase_ )] # split list_of_ints between the buckets for i in list_of_ints: __lowercase = int((i / placement) % RADIX ) buckets[tmp].append(lowerCamelCase_ ) # put each buckets' contents into list_of_ints __lowercase = 0 for b in range(lowerCamelCase_ ): for i in buckets[b]: __lowercase = i a += 1 # move to next placement *= RADIX return list_of_ints if __name__ == "__main__": import doctest doctest.testmod()
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) lowerCAmelCase__ = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''NllbTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = ['''NllbTokenizerFast'''] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb import NllbTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_nllb_fast import NllbTokenizerFast else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' from __future__ import annotations def _A ( A__ ): """simple docstring""" if not nums: return 0 __lowercase = nums[0] __lowercase = 0 for num in nums[1:]: __lowercase = ( max_excluding + num, max(lowerCamelCase__ , lowerCamelCase__ ), ) return max(lowerCamelCase__ , lowerCamelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' import fire from utils import calculate_rouge, save_json def _A ( A__ , A__ , A__=None , **A__ ): """simple docstring""" __lowercase = [x.strip() for x in open(a__ ).readlines()] __lowercase = [x.strip() for x in open(a__ ).readlines()][: len(a__ )] __lowercase = calculate_rouge(a__ , a__ , **a__ ) if save_path is not None: save_json(a__ , a__ , indent=a__ ) return metrics # these print nicely if __name__ == "__main__": fire.Fire(calculate_rouge_path)
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class lowercase_ : """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[int]=1_3 ,lowercase__ : Union[str, Any]=2 ,lowercase__ : List[Any]=2_4 ,lowercase__ : List[Any]=1_6 ,lowercase__ : Any=True ,lowercase__ : Dict=True ,lowercase__ : int=3_2 ,lowercase__ : Tuple=5 ,lowercase__ : List[str]=4 ,lowercase__ : Any=3_7 ,lowercase__ : int="gelu" ,lowercase__ : List[str]=0.1 ,lowercase__ : Optional[Any]=0.1 ,lowercase__ : Tuple=1_0 ,lowercase__ : str=0.0_2 ,lowercase__ : str=None ,lowercase__ : List[str]=2 ,lowercase__ : int=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = patch_size __lowercase = max_length __lowercase = num_mel_bins __lowercase = is_training __lowercase = use_labels __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = scope __lowercase = frequency_stride __lowercase = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) __lowercase = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 __lowercase = (self.max_length - self.patch_size) // self.time_stride + 1 __lowercase = frequency_out_dimension * time_out_dimension __lowercase = num_patches + 2 def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = self.get_config() return config, input_values, labels def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return ASTConfig( patch_size=self.patch_size ,max_length=self.max_length ,num_mel_bins=self.num_mel_bins ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,is_decoder=lowerCAmelCase_ ,initializer_range=self.initializer_range ,frequency_stride=self.frequency_stride ,time_stride=self.time_stride ,) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : Tuple ): __lowercase = ASTModel(config=lowerCAmelCase_ ) model.to(lowerCAmelCase_ ) model.eval() __lowercase = model(lowerCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_values''': input_values} return config, inputs_dict @require_torch class lowercase_ (__lowerCAmelCase , __lowerCAmelCase , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) SCREAMING_SNAKE_CASE : Dict = ( {'''audio-classification''': ASTForAudioClassification, '''feature-extraction''': ASTModel} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : str = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : Dict ,lowercase__ : List[Any] ): if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = ASTModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowerCAmelCase_ ,has_text_modality=lowerCAmelCase_ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() @unittest.skip(reason='''AST does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowerCAmelCase_ ) self.assertIsInstance(model.get_input_embeddings() ,(nn.Module) ) __lowercase = model.get_output_embeddings() self.assertTrue(x is None or isinstance(lowerCAmelCase_ ,nn.Linear ) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowerCAmelCase_ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''input_values'''] self.assertListEqual(arg_names[:1] ,lowerCAmelCase_ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCAmelCase_ ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = ASTModel.from_pretrained(lowerCAmelCase_ ) self.assertIsNotNone(lowerCAmelCase_ ) def _A ( ): """simple docstring""" __lowercase = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' ) __lowercase , __lowercase = torchaudio.load(snake_case__ ) return audio, sampling_rate @require_torch @require_torchaudio class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : List[Any] ): return ( ASTFeatureExtractor.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ) if is_torchaudio_available() else None ) @slow def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.default_feature_extractor __lowercase = ASTForAudioClassification.from_pretrained('''MIT/ast-finetuned-audioset-10-10-0.4593''' ).to(lowerCAmelCase_ ) __lowercase = self.default_feature_extractor __lowercase , __lowercase = prepare_audio() __lowercase = audio.squeeze().numpy() __lowercase = feature_extractor(lowerCAmelCase_ ,sampling_rate=lowerCAmelCase_ ,return_tensors='''pt''' ).to(lowerCAmelCase_ ) # forward pass with torch.no_grad(): __lowercase = model(**lowerCAmelCase_ ) # verify the logits __lowercase = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape ,lowerCAmelCase_ ) __lowercase = torch.tensor([-0.8_7_6_0, -7.0_0_4_2, -8.6_6_0_2] ).to(lowerCAmelCase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowerCAmelCase_ ,atol=1e-4 ) )
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' import json import os from dataclasses import dataclass from functools import partial from typing import Callable import flax.linen as nn import jax import jax.numpy as jnp import joblib import optax import wandb from flax import jax_utils, struct, traverse_util from flax.serialization import from_bytes, to_bytes from flax.training import train_state from flax.training.common_utils import shard from tqdm.auto import tqdm from transformers import BigBirdConfig, FlaxBigBirdForQuestionAnswering from transformers.models.big_bird.modeling_flax_big_bird import FlaxBigBirdForQuestionAnsweringModule class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : BigBirdConfig SCREAMING_SNAKE_CASE : jnp.dtype = jnp.floataa SCREAMING_SNAKE_CASE : bool = True def SCREAMING_SNAKE_CASE ( self : int ): super().setup() __lowercase = nn.Dense(5 ,dtype=self.dtype ) def __call__( self : Union[str, Any] ,*lowercase__ : int ,**lowercase__ : Dict ): __lowercase = super().__call__(*UpperCAmelCase__ ,**UpperCAmelCase__ ) __lowercase = self.cls(outputs[2] ) return outputs[:2] + (cls_out,) class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = FlaxBigBirdForNaturalQuestionsModule def _A ( A__ , A__ , A__ , A__ , A__ , A__ ): """simple docstring""" def cross_entropy(A__ , A__ , A__=None ): __lowercase = logits.shape[-1] __lowercase = (labels[..., None] == jnp.arange(A__ )[None]).astype('''f4''' ) __lowercase = jax.nn.log_softmax(A__ , axis=-1 ) __lowercase = -jnp.sum(labels * logits , axis=-1 ) if reduction is not None: __lowercase = reduction(A__ ) return loss __lowercase = partial(A__ , reduction=jnp.mean ) __lowercase = cross_entropy(A__ , A__ ) __lowercase = cross_entropy(A__ , A__ ) __lowercase = cross_entropy(A__ , A__ ) return (start_loss + end_loss + pooled_loss) / 3 @dataclass class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : str = "google/bigbird-roberta-base" SCREAMING_SNAKE_CASE : int = 3_0_0_0 SCREAMING_SNAKE_CASE : int = 1_0_5_0_0 SCREAMING_SNAKE_CASE : int = 1_2_8 SCREAMING_SNAKE_CASE : int = 3 SCREAMING_SNAKE_CASE : int = 1 SCREAMING_SNAKE_CASE : int = 5 # tx_args SCREAMING_SNAKE_CASE : float = 3e-5 SCREAMING_SNAKE_CASE : float = 0.0 SCREAMING_SNAKE_CASE : int = 2_0_0_0_0 SCREAMING_SNAKE_CASE : float = 0.0095 SCREAMING_SNAKE_CASE : str = "bigbird-roberta-natural-questions" SCREAMING_SNAKE_CASE : str = "training-expt" SCREAMING_SNAKE_CASE : str = "data/nq-training.jsonl" SCREAMING_SNAKE_CASE : str = "data/nq-validation.jsonl" def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): os.makedirs(self.base_dir ,exist_ok=UpperCAmelCase__ ) __lowercase = os.path.join(self.base_dir ,self.save_dir ) __lowercase = self.batch_size_per_device * jax.device_count() @dataclass class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : int SCREAMING_SNAKE_CASE : int = 4_0_9_6 # no dynamic padding on TPUs def __call__( self : Tuple ,lowercase__ : Optional[int] ): __lowercase = self.collate_fn(UpperCAmelCase__ ) __lowercase = jax.tree_util.tree_map(UpperCAmelCase__ ,UpperCAmelCase__ ) return batch def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.fetch_inputs(features['''input_ids'''] ) __lowercase = { '''input_ids''': jnp.array(UpperCAmelCase__ ,dtype=jnp.intaa ), '''attention_mask''': jnp.array(UpperCAmelCase__ ,dtype=jnp.intaa ), '''start_labels''': jnp.array(features['''start_token'''] ,dtype=jnp.intaa ), '''end_labels''': jnp.array(features['''end_token'''] ,dtype=jnp.intaa ), '''pooled_labels''': jnp.array(features['''category'''] ,dtype=jnp.intaa ), } return batch def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : list ): __lowercase = [self._fetch_inputs(UpperCAmelCase__ ) for ids in input_ids] return zip(*UpperCAmelCase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : list ): __lowercase = [1 for _ in range(len(UpperCAmelCase__ ) )] while len(UpperCAmelCase__ ) < self.max_length: input_ids.append(self.pad_id ) attention_mask.append(0 ) return input_ids, attention_mask def _A ( A__ , A__ , A__=None ): """simple docstring""" if seed is not None: __lowercase = dataset.shuffle(seed=A__ ) for i in range(len(A__ ) // batch_size ): __lowercase = dataset[i * batch_size : (i + 1) * batch_size] yield dict(A__ ) @partial(jax.pmap , axis_name='''batch''' ) def _A ( A__ , A__ , **A__ ): """simple docstring""" def loss_fn(A__ ): __lowercase = model_inputs.pop('''start_labels''' ) __lowercase = model_inputs.pop('''end_labels''' ) __lowercase = model_inputs.pop('''pooled_labels''' ) __lowercase = state.apply_fn(**A__ , params=A__ , dropout_rng=A__ , train=A__ ) __lowercase = outputs return state.loss_fn( A__ , A__ , A__ , A__ , A__ , A__ , ) __lowercase = jax.random.split(A__ ) __lowercase = jax.value_and_grad(A__ ) __lowercase = grad_fn(state.params ) __lowercase = jax.lax.pmean({'''loss''': loss} , axis_name='''batch''' ) __lowercase = jax.lax.pmean(A__ , '''batch''' ) __lowercase = state.apply_gradients(grads=A__ ) return state, metrics, new_drp_rng @partial(jax.pmap , axis_name='''batch''' ) def _A ( A__ , **A__ ): """simple docstring""" __lowercase = model_inputs.pop('''start_labels''' ) __lowercase = model_inputs.pop('''end_labels''' ) __lowercase = model_inputs.pop('''pooled_labels''' ) __lowercase = state.apply_fn(**A__ , params=state.params , train=A__ ) __lowercase = outputs __lowercase = state.loss_fn(A__ , A__ , A__ , A__ , A__ , A__ ) __lowercase = jax.lax.pmean({'''loss''': loss} , axis_name='''batch''' ) return metrics class lowercase_ (train_state.TrainState ): """simple docstring""" SCREAMING_SNAKE_CASE : Callable = struct.field(pytree_node=lowerCamelCase__ ) @dataclass class lowercase_ : """simple docstring""" SCREAMING_SNAKE_CASE : Args SCREAMING_SNAKE_CASE : Callable SCREAMING_SNAKE_CASE : Callable SCREAMING_SNAKE_CASE : Callable SCREAMING_SNAKE_CASE : Callable SCREAMING_SNAKE_CASE : wandb SCREAMING_SNAKE_CASE : Callable = None def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : List[str]=None ): __lowercase = model.params __lowercase = TrainState.create( apply_fn=model.__call__ ,params=UpperCAmelCase__ ,tx=UpperCAmelCase__ ,loss_fn=UpperCAmelCase__ ,) if ckpt_dir is not None: __lowercase = restore_checkpoint(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowercase = { '''lr''': args.lr, '''init_lr''': args.init_lr, '''warmup_steps''': args.warmup_steps, '''num_train_steps''': num_train_steps, '''weight_decay''': args.weight_decay, } __lowercase = build_tx(**UpperCAmelCase__ ) __lowercase = train_state.TrainState( step=UpperCAmelCase__ ,apply_fn=model.__call__ ,params=UpperCAmelCase__ ,tx=UpperCAmelCase__ ,opt_state=UpperCAmelCase__ ,) __lowercase = args __lowercase = data_collator __lowercase = lr __lowercase = params __lowercase = jax_utils.replicate(UpperCAmelCase__ ) return state def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : Any ): __lowercase = self.args __lowercase = len(UpperCAmelCase__ ) // args.batch_size __lowercase = jax.random.PRNGKey(0 ) __lowercase = jax.random.split(UpperCAmelCase__ ,jax.device_count() ) for epoch in range(args.max_epochs ): __lowercase = jnp.array(0 ,dtype=jnp.floataa ) __lowercase = get_batched_dataset(UpperCAmelCase__ ,args.batch_size ,seed=UpperCAmelCase__ ) __lowercase = 0 for batch in tqdm(UpperCAmelCase__ ,total=UpperCAmelCase__ ,desc=F"Running EPOCH-{epoch}" ): __lowercase = self.data_collator(UpperCAmelCase__ ) __lowercase = self.train_step_fn(UpperCAmelCase__ ,UpperCAmelCase__ ,**UpperCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics['''loss'''] ) i += 1 if i % args.logging_steps == 0: __lowercase = jax_utils.unreplicate(state.step ) __lowercase = running_loss.item() / i __lowercase = self.scheduler_fn(state_step - 1 ) __lowercase = self.evaluate(UpperCAmelCase__ ,UpperCAmelCase__ ) __lowercase = { '''step''': state_step.item(), '''eval_loss''': eval_loss.item(), '''tr_loss''': tr_loss, '''lr''': lr.item(), } tqdm.write(str(UpperCAmelCase__ ) ) self.logger.log(UpperCAmelCase__ ,commit=UpperCAmelCase__ ) if i % args.save_steps == 0: self.save_checkpoint(args.save_dir + F"-e{epoch}-s{i}" ,state=UpperCAmelCase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : Tuple ): __lowercase = get_batched_dataset(UpperCAmelCase__ ,self.args.batch_size ) __lowercase = len(UpperCAmelCase__ ) // self.args.batch_size __lowercase = jnp.array(0 ,dtype=jnp.floataa ) __lowercase = 0 for batch in tqdm(UpperCAmelCase__ ,total=UpperCAmelCase__ ,desc='''Evaluating ... ''' ): __lowercase = self.data_collator(UpperCAmelCase__ ) __lowercase = self.val_step_fn(UpperCAmelCase__ ,**UpperCAmelCase__ ) running_loss += jax_utils.unreplicate(metrics['''loss'''] ) i += 1 return running_loss / i def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ): __lowercase = jax_utils.unreplicate(UpperCAmelCase__ ) print(F"SAVING CHECKPOINT IN {save_dir}" ,end=''' ... ''' ) self.model_save_fn(UpperCAmelCase__ ,params=state.params ) with open(os.path.join(UpperCAmelCase__ ,'''opt_state.msgpack''' ) ,'''wb''' ) as f: f.write(to_bytes(state.opt_state ) ) joblib.dump(self.args ,os.path.join(UpperCAmelCase__ ,'''args.joblib''' ) ) joblib.dump(self.data_collator ,os.path.join(UpperCAmelCase__ ,'''data_collator.joblib''' ) ) with open(os.path.join(UpperCAmelCase__ ,'''training_state.json''' ) ,'''w''' ) as f: json.dump({'''step''': state.step.item()} ,UpperCAmelCase__ ) print('''DONE''' ) def _A ( A__ , A__ ): """simple docstring""" print(F"RESTORING CHECKPOINT FROM {save_dir}" , end=''' ... ''' ) with open(os.path.join(A__ , '''flax_model.msgpack''' ) , '''rb''' ) as f: __lowercase = from_bytes(state.params , f.read() ) with open(os.path.join(A__ , '''opt_state.msgpack''' ) , '''rb''' ) as f: __lowercase = from_bytes(state.opt_state , f.read() ) __lowercase = joblib.load(os.path.join(A__ , '''args.joblib''' ) ) __lowercase = joblib.load(os.path.join(A__ , '''data_collator.joblib''' ) ) with open(os.path.join(A__ , '''training_state.json''' ) , '''r''' ) as f: __lowercase = json.load(A__ ) __lowercase = training_state['''step'''] print('''DONE''' ) return params, opt_state, step, args, data_collator def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = num_train_steps - warmup_steps __lowercase = optax.linear_schedule(init_value=A__ , end_value=A__ , transition_steps=A__ ) __lowercase = optax.linear_schedule(init_value=A__ , end_value=1e-7 , transition_steps=A__ ) __lowercase = optax.join_schedules(schedules=[warmup_fn, decay_fn] , boundaries=[warmup_steps] ) return lr def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" def weight_decay_mask(A__ ): __lowercase = traverse_util.flatten_dict(A__ ) __lowercase = {k: (v[-1] != '''bias''' and v[-2:] != ('''LayerNorm''', '''scale''')) for k, v in params.items()} return traverse_util.unflatten_dict(A__ ) __lowercase = scheduler_fn(A__ , A__ , A__ , A__ ) __lowercase = optax.adamw(learning_rate=A__ , weight_decay=A__ , mask=A__ ) return tx, lr
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import datetime import platform import subprocess from typing import Optional, Tuple, Union import numpy as np def _A ( A__ , A__ ): """simple docstring""" __lowercase = F"{sampling_rate}" __lowercase = "1" __lowercase = "f32le" __lowercase = [ "ffmpeg", "-i", "pipe:0", "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-hide_banner", "-loglevel", "quiet", "pipe:1", ] try: with subprocess.Popen(A__ , stdin=subprocess.PIPE , stdout=subprocess.PIPE ) as ffmpeg_process: __lowercase = ffmpeg_process.communicate(A__ ) except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to load audio files from filename''' ) from error __lowercase = output_stream[0] __lowercase = np.frombuffer(A__ , np.floataa ) if audio.shape[0] == 0: raise ValueError('''Malformed soundfile''' ) return audio def _A ( A__ , A__ , A__ = "f32le" , ): """simple docstring""" __lowercase = F"{sampling_rate}" __lowercase = "1" if format_for_conversion == "s16le": __lowercase = 2 elif format_for_conversion == "f32le": __lowercase = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) __lowercase = platform.system() if system == "Linux": __lowercase = "alsa" __lowercase = "default" elif system == "Darwin": __lowercase = "avfoundation" __lowercase = ":0" elif system == "Windows": __lowercase = "dshow" __lowercase = "default" __lowercase = [ "ffmpeg", "-f", format_, "-i", input_, "-ac", ac, "-ar", ar, "-f", format_for_conversion, "-fflags", "nobuffer", "-hide_banner", "-loglevel", "quiet", "pipe:1", ] __lowercase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample __lowercase = _ffmpeg_stream(A__ , A__ ) for item in iterator: yield item def _A ( A__ , A__ , A__ = None , A__ = None , A__ = "f32le" , ): """simple docstring""" if stream_chunk_s is not None: __lowercase = stream_chunk_s else: __lowercase = chunk_length_s __lowercase = ffmpeg_microphone(A__ , A__ , format_for_conversion=A__ ) if format_for_conversion == "s16le": __lowercase = np.intaa __lowercase = 2 elif format_for_conversion == "f32le": __lowercase = np.floataa __lowercase = 4 else: raise ValueError(F"Unhandled format `{format_for_conversion}`. Please use `s16le` or `f32le`" ) if stride_length_s is None: __lowercase = chunk_length_s / 6 __lowercase = int(round(sampling_rate * chunk_length_s ) ) * size_of_sample if isinstance(A__ , (int, float) ): __lowercase = [stride_length_s, stride_length_s] __lowercase = int(round(sampling_rate * stride_length_s[0] ) ) * size_of_sample __lowercase = int(round(sampling_rate * stride_length_s[1] ) ) * size_of_sample __lowercase = datetime.datetime.now() __lowercase = datetime.timedelta(seconds=A__ ) for item in chunk_bytes_iter(A__ , A__ , stride=(stride_left, stride_right) , stream=A__ ): # Put everything back in numpy scale __lowercase = np.frombuffer(item['''raw'''] , dtype=A__ ) __lowercase = ( item["stride"][0] // size_of_sample, item["stride"][1] // size_of_sample, ) __lowercase = sampling_rate audio_time += delta if datetime.datetime.now() > audio_time + 10 * delta: # We're late !! SKIP continue yield item def _A ( A__ , A__ , A__ , A__ = False ): """simple docstring""" __lowercase = b"" __lowercase = stride if stride_left + stride_right >= chunk_len: raise ValueError( F"Stride needs to be strictly smaller than chunk_len: ({stride_left}, {stride_right}) vs {chunk_len}" ) __lowercase = 0 for raw in iterator: acc += raw if stream and len(A__ ) < chunk_len: __lowercase = (_stride_left, 0) yield {"raw": acc[:chunk_len], "stride": stride, "partial": True} else: while len(A__ ) >= chunk_len: # We are flushing the accumulator __lowercase = (_stride_left, stride_right) __lowercase = {"raw": acc[:chunk_len], "stride": stride} if stream: __lowercase = False yield item __lowercase = stride_left __lowercase = acc[chunk_len - stride_left - stride_right :] # Last chunk if len(A__ ) > stride_left: __lowercase = {"raw": acc, "stride": (_stride_left, 0)} if stream: __lowercase = False yield item def _A ( A__ , A__ ): """simple docstring""" __lowercase = 2**24 # 16Mo try: with subprocess.Popen(A__ , stdout=subprocess.PIPE , bufsize=A__ ) as ffmpeg_process: while True: __lowercase = ffmpeg_process.stdout.read(A__ ) if raw == b"": break yield raw except FileNotFoundError as error: raise ValueError('''ffmpeg was not found but is required to stream audio files from filename''' ) from error
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' # XXX: we want transformers master here - in the absense of conftest manipulating sys.path: # hack it in for now: import sys from pathlib import Path lowerCAmelCase__ = Path(__file__).resolve().parents[3] / '''src''' sys.path.insert(1, str(git_repo_path)) import dataclasses # noqa import io # noqa import itertools # noqa import json # noqa import os # noqa import unittest # noqa from copy import deepcopy # noqa from parameterized import parameterized # noqa from transformers import TrainingArguments, is_torch_available # noqa from transformers.deepspeed import is_deepspeed_available # noqa from transformers.file_utils import WEIGHTS_NAME # noqa from transformers.testing_utils import ( # noqa CaptureLogger, ExtendSysPath, TestCasePlus, execute_subprocess_async, get_gpu_count, mockenv_context, require_deepspeed, require_torch_gpu, require_torch_multi_gpu, slow, ) from transformers.trainer_utils import set_seed # noqa set_seed(42) lowerCAmelCase__ = {'''base''': '''patrickvonplaten/wav2vec2_tiny_random''', '''robust''': '''patrickvonplaten/wav2vec2_tiny_random_robust'''} lowerCAmelCase__ = '''zero2''' lowerCAmelCase__ = '''zero3''' lowerCAmelCase__ = [ZEROa, ZEROa] def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = parameterized.to_safe_name('''_'''.join(str(__a ) for x in param.args ) ) return F"{func.__name__}_{param_based_name}" # Cartesian-product of zero stages with models to test lowerCAmelCase__ = list(itertools.product(stages, models.keys())) @slow @require_deepspeed @require_torch_gpu class lowercase_ (__lowerCamelCase ): """simple docstring""" @parameterized.expand(SCREAMING_SNAKE_CASE_ ,name_func=SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ ,model=SCREAMING_SNAKE_CASE_ ,distributed=SCREAMING_SNAKE_CASE_ ,fpaa=SCREAMING_SNAKE_CASE_ ,) @require_torch_multi_gpu @parameterized.expand(SCREAMING_SNAKE_CASE_ ,name_func=SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : int ,lowercase__ : Dict ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ ,model=SCREAMING_SNAKE_CASE_ ,distributed=SCREAMING_SNAKE_CASE_ ,fpaa=SCREAMING_SNAKE_CASE_ ,) @parameterized.expand(SCREAMING_SNAKE_CASE_ ,name_func=SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ ,model=SCREAMING_SNAKE_CASE_ ,distributed=SCREAMING_SNAKE_CASE_ ,fpaa=SCREAMING_SNAKE_CASE_ ,) @require_torch_multi_gpu @parameterized.expand(SCREAMING_SNAKE_CASE_ ,name_func=SCREAMING_SNAKE_CASE_ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ): self.run_and_check( stage=SCREAMING_SNAKE_CASE_ ,model=SCREAMING_SNAKE_CASE_ ,distributed=SCREAMING_SNAKE_CASE_ ,fpaa=SCREAMING_SNAKE_CASE_ ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Tuple ): # XXX: run_asr is premature and doesn't save any results # so all we check for now is that the process didn't fail pass def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : Any = 1_0 ,lowercase__ : str = True ,lowercase__ : Union[str, Any] = True ,lowercase__ : Union[str, Any] = True ,): __lowercase = models[model] __lowercase = self.run_trainer( stage=SCREAMING_SNAKE_CASE_ ,model_name=SCREAMING_SNAKE_CASE_ ,eval_steps=SCREAMING_SNAKE_CASE_ ,num_train_epochs=1 ,distributed=SCREAMING_SNAKE_CASE_ ,fpaa=SCREAMING_SNAKE_CASE_ ,) self.do_checks(SCREAMING_SNAKE_CASE_ ) return output_dir def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Tuple ,lowercase__ : int ,lowercase__ : str = 1_0 ,lowercase__ : int = 1 ,lowercase__ : Any = True ,lowercase__ : Tuple = True ,): __lowercase = self.get_auto_remove_tmp_dir('''./xxx''' ,after=SCREAMING_SNAKE_CASE_ ) __lowercase = F"\n --model_name_or_path {model_name}\n --dataset_name hf-internal-testing/librispeech_asr_dummy\n --dataset_config_name clean\n --train_split_name validation\n --validation_split_name validation\n --output_dir {output_dir}\n --num_train_epochs {str(SCREAMING_SNAKE_CASE_ )}\n --per_device_train_batch_size 2\n --per_device_eval_batch_size 2\n --evaluation_strategy steps\n --learning_rate 5e-4\n --warmup_steps 8\n --orthography timit\n --preprocessing_num_workers 1\n --group_by_length\n --freeze_feature_extractor\n --report_to none\n --save_steps 0\n --eval_steps {eval_steps}\n --report_to none\n ".split() if fpaa: args.extend(['''--fp16'''] ) # currently ds_config_wav2vec2_zero.json requires "zero_optimization.find_unused_parameters": true, # hence the separate config files __lowercase = F"--deepspeed {self.test_file_dir_str}/ds_config_wav2vec2_{stage}.json".split() __lowercase = [F"{self.examples_dir_str}/research_projects/wav2vec2/run_asr.py"] __lowercase = self.get_launcher(SCREAMING_SNAKE_CASE_ ) __lowercase = launcher + script + args + ds_args # keep for quick debug # print(" ".join([f"\nPYTHONPATH={self.src_dir_str}"] +cmd)); die execute_subprocess_async(SCREAMING_SNAKE_CASE_ ,env=self.get_env() ) return output_dir def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Any=False ): # 1. explicitly set --num_nodes=1 just in case these tests end up run on a multi-node setup # - it won't be able to handle that # 2. for now testing with just 2 gpus max (since some quality tests may give different # results with mode gpus because we use very little data) __lowercase = min(2 ,get_gpu_count() ) if distributed else 1 return F"deepspeed --num_nodes 1 --num_gpus {num_gpus}".split()
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
import math from typing import Dict, Iterable, List, Optional, Tuple, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import normalize, rescale, resize, to_channel_dimension_format from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, get_image_size, is_torch_available, is_torch_tensor, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_torch_available(): import torch if is_vision_available(): import PIL lowerCAmelCase__ = logging.get_logger(__name__) def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" def constraint_to_multiple_of(A__ , A__ , A__=0 , A__=None ): __lowercase = round(val / multiple ) * multiple if max_val is not None and x > max_val: __lowercase = math.floor(val / multiple ) * multiple if x < min_val: __lowercase = math.ceil(val / multiple ) * multiple return x __lowercase = (output_size, output_size) if isinstance(A__ , A__ ) else output_size __lowercase = get_image_size(A__ ) __lowercase = output_size # determine new height and width __lowercase = output_height / input_height __lowercase = output_width / input_width if keep_aspect_ratio: # scale as little as possible if abs(1 - scale_width ) < abs(1 - scale_height ): # fit width __lowercase = scale_width else: # fit height __lowercase = scale_height __lowercase = constraint_to_multiple_of(scale_height * input_height , multiple=A__ ) __lowercase = constraint_to_multiple_of(scale_width * input_width , multiple=A__ ) return (new_height, new_width) class lowercase_ (_snake_case ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = ["""pixel_values"""] def __init__( self : List[str] ,lowercase__ : List[Any] = True ,lowercase__ : Any = None ,lowercase__ : List[Any] = PILImageResampling.BILINEAR ,lowercase__ : str = False ,lowercase__ : Tuple = 1 ,lowercase__ : int = True ,lowercase__ : Dict = 1 / 2_5_5 ,lowercase__ : Tuple = True ,lowercase__ : str = None ,lowercase__ : List[str] = None ,**lowercase__ : List[str] ,): super().__init__(**snake_case_ ) __lowercase = size if size is not None else {"height": 3_8_4, "width": 3_8_4} __lowercase = get_size_dict(snake_case_ ) __lowercase = do_resize __lowercase = size __lowercase = keep_aspect_ratio __lowercase = ensure_multiple_of __lowercase = resample __lowercase = do_rescale __lowercase = rescale_factor __lowercase = do_normalize __lowercase = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN __lowercase = image_std if image_std is not None else IMAGENET_STANDARD_STD def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Dict = False ,lowercase__ : Tuple = 1 ,lowercase__ : List[Any] = PILImageResampling.BICUBIC ,lowercase__ : Union[str, Any] = None ,**lowercase__ : Optional[Any] ,): __lowercase = get_size_dict(snake_case_ ) if "height" not in size or "width" not in size: raise ValueError(F"The size dictionary must contain the keys \'height\' and \'width\'. Got {size.keys()}" ) __lowercase = get_resize_output_image_size( snake_case_ ,output_size=(size['''height'''], size['''width''']) ,keep_aspect_ratio=snake_case_ ,multiple=snake_case_ ,) return resize(snake_case_ ,size=snake_case_ ,resample=snake_case_ ,data_format=snake_case_ ,**snake_case_ ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : Dict ,lowercase__ : str ,lowercase__ : Tuple = None ,**lowercase__ : Any ,): return rescale(snake_case_ ,scale=snake_case_ ,data_format=snake_case_ ,**snake_case_ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : Dict = None ,**lowercase__ : List[str] ,): return normalize(snake_case_ ,mean=snake_case_ ,std=snake_case_ ,data_format=snake_case_ ,**snake_case_ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] = None ,lowercase__ : Union[str, Any] = None ,lowercase__ : List[str] = None ,lowercase__ : List[str] = None ,lowercase__ : Optional[Any] = None ,lowercase__ : Dict = None ,lowercase__ : Union[str, Any] = None ,lowercase__ : List[Any] = None ,lowercase__ : Any = None ,lowercase__ : Optional[int] = None ,lowercase__ : Dict = None ,lowercase__ : Dict = ChannelDimension.FIRST ,**lowercase__ : List[str] ,): __lowercase = do_resize if do_resize is not None else self.do_resize __lowercase = size if size is not None else self.size __lowercase = get_size_dict(snake_case_ ) __lowercase = keep_aspect_ratio if keep_aspect_ratio is not None else self.keep_aspect_ratio __lowercase = ensure_multiple_of if ensure_multiple_of is not None else self.ensure_multiple_of __lowercase = resample if resample is not None else self.resample __lowercase = do_rescale if do_rescale is not None else self.do_rescale __lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase = do_normalize if do_normalize is not None else self.do_normalize __lowercase = image_mean if image_mean is not None else self.image_mean __lowercase = image_std if image_std is not None else self.image_std __lowercase = make_list_of_images(snake_case_ ) if not valid_images(snake_case_ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None or resample is None: raise ValueError('''Size and resample must be specified if do_resize is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # All transformations expect numpy arrays. __lowercase = [to_numpy_array(snake_case_ ) for image in images] if do_resize: __lowercase = [self.resize(image=snake_case_ ,size=snake_case_ ,resample=snake_case_ ) for image in images] if do_rescale: __lowercase = [self.rescale(image=snake_case_ ,scale=snake_case_ ) for image in images] if do_normalize: __lowercase = [self.normalize(image=snake_case_ ,mean=snake_case_ ,std=snake_case_ ) for image in images] __lowercase = [to_channel_dimension_format(snake_case_ ,snake_case_ ) for image in images] __lowercase = {"pixel_values": images} return BatchFeature(data=snake_case_ ,tensor_type=snake_case_ ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple = None ): __lowercase = outputs.logits # Resize logits and compute semantic segmentation maps if target_sizes is not None: if len(snake_case_ ) != len(snake_case_ ): raise ValueError( '''Make sure that you pass in as many target sizes as the batch dimension of the logits''' ) if is_torch_tensor(snake_case_ ): __lowercase = target_sizes.numpy() __lowercase = [] for idx in range(len(snake_case_ ) ): __lowercase = torch.nn.functional.interpolate( logits[idx].unsqueeze(dim=0 ) ,size=target_sizes[idx] ,mode='''bilinear''' ,align_corners=snake_case_ ) __lowercase = resized_logits[0].argmax(dim=0 ) semantic_segmentation.append(snake_case_ ) else: __lowercase = logits.argmax(dim=1 ) __lowercase = [semantic_segmentation[i] for i in range(semantic_segmentation.shape[0] )] return semantic_segmentation
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' import warnings from ...utils import logging from .image_processing_deformable_detr import DeformableDetrImageProcessor lowerCAmelCase__ = logging.get_logger(__name__) class lowercase_ (lowercase__ ): def __init__( self : List[str] ,*lowercase__ : Tuple ,**lowercase__ : str ): warnings.warn( '''The class DeformableDetrFeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use DeformableDetrImageProcessor instead.''' ,UpperCAmelCase__ ,) super().__init__(*UpperCAmelCase__ ,**UpperCAmelCase__ )
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' import pickle import unittest import torch from accelerate import Accelerator from accelerate.state import AcceleratorState from accelerate.test_utils import require_cpu @require_cpu class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = torch.nn.Linear(1_0 ,1_0 ) __lowercase = torch.optim.SGD(model.parameters() ,0.1 ) __lowercase = Accelerator() __lowercase = accelerator.prepare(_lowerCamelCase ) try: pickle.loads(pickle.dumps(_lowerCamelCase ) ) except Exception as e: self.fail(F"Accelerated optimizer pickling failed with {e}" ) AcceleratorState._reset_state()
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from ...utils.dataclasses import ( ComputeEnvironment, DistributedType, DynamoBackend, PrecisionType, SageMakerDistributedType, ) from ..menu import BulletMenu lowerCAmelCase__ = [ 'EAGER', 'AOT_EAGER', 'INDUCTOR', 'NVFUSER', 'AOT_NVFUSER', 'AOT_CUDAGRAPHS', 'OFI', 'FX2TRT', 'ONNXRT', 'IPEX', ] def _A ( A__ , A__=None , A__=None , A__=None ): """simple docstring""" __lowercase = True while ask_again: __lowercase = input(lowerCamelCase_ ) try: if default is not None and len(lowerCamelCase_ ) == 0: return default return convert_value(lowerCamelCase_ ) if convert_value is not None else result except Exception: if error_message is not None: print(lowerCamelCase_ ) def _A ( A__ , A__=[] , A__=None , A__=0 ): """simple docstring""" __lowercase = BulletMenu(lowerCamelCase_ , lowerCamelCase_ ) __lowercase = menu.run(default_choice=lowerCamelCase_ ) return convert_value(lowerCamelCase_ ) if convert_value is not None else result def _A ( A__ ): """simple docstring""" __lowercase = int(lowerCamelCase_ ) return ComputeEnvironment(['''LOCAL_MACHINE''', '''AMAZON_SAGEMAKER'''][value] ) def _A ( A__ ): """simple docstring""" __lowercase = int(lowerCamelCase_ ) return DistributedType(['''NO''', '''MULTI_CPU''', '''MULTI_XPU''', '''MULTI_GPU''', '''MULTI_NPU''', '''TPU'''][value] ) def _A ( A__ ): """simple docstring""" __lowercase = int(lowerCamelCase_ ) return DynamoBackend(DYNAMO_BACKENDS[value] ).value def _A ( A__ ): """simple docstring""" __lowercase = int(lowerCamelCase_ ) return PrecisionType(['''no''', '''fp16''', '''bf16''', '''fp8'''][value] ) def _A ( A__ ): """simple docstring""" __lowercase = int(lowerCamelCase_ ) return SageMakerDistributedType(['''NO''', '''DATA_PARALLEL''', '''MODEL_PARALLEL'''][value] ) def _A ( A__ ): """simple docstring""" return {"yes": True, "no": False}[value.lower()] class lowercase_ (argparse.RawDescriptionHelpFormatter ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = super()._format_usage(__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ,__lowerCamelCase ) __lowercase = usage.replace('''<command> [<args>] ''' ,'''''' ) return usage
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' def _A ( ): """simple docstring""" return 1 def _A ( A__ ): """simple docstring""" return 0 if x < 0 else two_pence(x - 2 ) + one_pence() def _A ( A__ ): """simple docstring""" return 0 if x < 0 else five_pence(x - 5 ) + two_pence(UpperCamelCase__ ) def _A ( A__ ): """simple docstring""" return 0 if x < 0 else ten_pence(x - 10 ) + five_pence(UpperCamelCase__ ) def _A ( A__ ): """simple docstring""" return 0 if x < 0 else twenty_pence(x - 20 ) + ten_pence(UpperCamelCase__ ) def _A ( A__ ): """simple docstring""" return 0 if x < 0 else fifty_pence(x - 50 ) + twenty_pence(UpperCamelCase__ ) def _A ( A__ ): """simple docstring""" return 0 if x < 0 else one_pound(x - 100 ) + fifty_pence(UpperCamelCase__ ) def _A ( A__ ): """simple docstring""" return 0 if x < 0 else two_pound(x - 200 ) + one_pound(UpperCamelCase__ ) def _A ( A__ = 200 ): """simple docstring""" return two_pound(UpperCamelCase__ ) if __name__ == "__main__": print(solution(int(input().strip())))
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' import argparse import json import gdown import numpy as np import torch from huggingface_hub import hf_hub_download from transformers import ( VideoMAEConfig, VideoMAEForPreTraining, VideoMAEForVideoClassification, VideoMAEImageProcessor, ) def _A ( A__ ): """simple docstring""" __lowercase = VideoMAEConfig() set_architecture_configs(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if "finetuned" not in model_name: __lowercase = False if "finetuned" in model_name: __lowercase = "huggingface/label-files" if "kinetics" in model_name: __lowercase = 400 __lowercase = "kinetics400-id2label.json" elif "ssv2" in model_name: __lowercase = 174 __lowercase = "something-something-v2-id2label.json" else: raise ValueError('''Model name should either contain \'kinetics\' or \'ssv2\' in case it\'s fine-tuned.''' ) __lowercase = json.load(open(hf_hub_download(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , repo_type='''dataset''' ) , '''r''' ) ) __lowercase = {int(__SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} __lowercase = idalabel __lowercase = {v: k for k, v in idalabel.items()} return config def _A ( A__ , A__ ): """simple docstring""" if "small" in model_name: __lowercase = 384 __lowercase = 1536 __lowercase = 12 __lowercase = 16 __lowercase = 12 __lowercase = 3 __lowercase = 192 __lowercase = 768 elif "large" in model_name: __lowercase = 1024 __lowercase = 4096 __lowercase = 24 __lowercase = 16 __lowercase = 12 __lowercase = 8 __lowercase = 512 __lowercase = 2048 elif "huge" in model_name: __lowercase = 1280 __lowercase = 5120 __lowercase = 32 __lowercase = 16 __lowercase = 12 __lowercase = 8 __lowercase = 640 __lowercase = 2560 elif "base" not in model_name: raise ValueError('''Model name should include either \"small\", \"base\", \"large\", or \"huge\"''' ) def _A ( A__ ): """simple docstring""" if "encoder." in name: __lowercase = name.replace('''encoder.''' , '''''' ) if "cls_token" in name: __lowercase = name.replace('''cls_token''' , '''videomae.embeddings.cls_token''' ) if "decoder_pos_embed" in name: __lowercase = name.replace('''decoder_pos_embed''' , '''decoder.decoder_pos_embed''' ) if "pos_embed" in name and "decoder" not in name: __lowercase = name.replace('''pos_embed''' , '''videomae.embeddings.position_embeddings''' ) if "patch_embed.proj" in name: __lowercase = name.replace('''patch_embed.proj''' , '''videomae.embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: __lowercase = name.replace('''patch_embed.norm''' , '''videomae.embeddings.norm''' ) if "decoder.blocks" in name: __lowercase = name.replace('''decoder.blocks''' , '''decoder.decoder_layers''' ) if "blocks" in name: __lowercase = name.replace('''blocks''' , '''videomae.encoder.layer''' ) if "attn.proj" in name: __lowercase = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name and "bias" not in name: __lowercase = name.replace('''attn''' , '''attention.self''' ) if "attn" in name: __lowercase = name.replace('''attn''' , '''attention.attention''' ) if "norm1" in name: __lowercase = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: __lowercase = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: __lowercase = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: __lowercase = name.replace('''mlp.fc2''' , '''output.dense''' ) if "decoder_embed" in name: __lowercase = name.replace('''decoder_embed''' , '''decoder.decoder_embed''' ) if "decoder_norm" in name: __lowercase = name.replace('''decoder_norm''' , '''decoder.decoder_norm''' ) if "decoder_pred" in name: __lowercase = name.replace('''decoder_pred''' , '''decoder.decoder_pred''' ) if "norm.weight" in name and "decoder" not in name and "fc" not in name: __lowercase = name.replace('''norm.weight''' , '''videomae.layernorm.weight''' ) if "norm.bias" in name and "decoder" not in name and "fc" not in name: __lowercase = name.replace('''norm.bias''' , '''videomae.layernorm.bias''' ) if "head" in name and "decoder" not in name: __lowercase = name.replace('''head''' , '''classifier''' ) return name def _A ( A__ , A__ ): """simple docstring""" for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(__SCREAMING_SNAKE_CASE ) if key.startswith('''encoder.''' ): __lowercase = key.replace('''encoder.''' , '''''' ) if "qkv" in key: __lowercase = key.split('''.''' ) if key.startswith('''decoder.blocks''' ): __lowercase = config.decoder_hidden_size __lowercase = int(key_split[2] ) __lowercase = "decoder.decoder_layers." if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[dim : dim * 2, :] __lowercase = val[-dim:, :] else: __lowercase = config.hidden_size __lowercase = int(key_split[1] ) __lowercase = "videomae.encoder.layer." if "weight" in key: __lowercase = val[:dim, :] __lowercase = val[dim : dim * 2, :] __lowercase = val[-dim:, :] else: __lowercase = val return orig_state_dict def _A ( ): """simple docstring""" __lowercase = hf_hub_download( repo_id='''hf-internal-testing/spaghetti-video''' , filename='''eating_spaghetti.npy''' , repo_type='''dataset''' ) __lowercase = np.load(__SCREAMING_SNAKE_CASE ) return list(__SCREAMING_SNAKE_CASE ) def _A ( A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = get_videomae_config(__SCREAMING_SNAKE_CASE ) if "finetuned" in model_name: __lowercase = VideoMAEForVideoClassification(__SCREAMING_SNAKE_CASE ) else: __lowercase = VideoMAEForPreTraining(__SCREAMING_SNAKE_CASE ) # download original checkpoint, hosted on Google Drive __lowercase = "pytorch_model.bin" gdown.cached_download(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , quiet=__SCREAMING_SNAKE_CASE ) __lowercase = torch.load(__SCREAMING_SNAKE_CASE , map_location='''cpu''' ) if "model" in files: __lowercase = files["model"] else: __lowercase = files["module"] __lowercase = convert_state_dict(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) model.load_state_dict(__SCREAMING_SNAKE_CASE ) model.eval() # verify model on basic input __lowercase = VideoMAEImageProcessor(image_mean=[0.5, 0.5, 0.5] , image_std=[0.5, 0.5, 0.5] ) __lowercase = prepare_video() __lowercase = image_processor(__SCREAMING_SNAKE_CASE , return_tensors='''pt''' ) if "finetuned" not in model_name: __lowercase = hf_hub_download(repo_id='''hf-internal-testing/bool-masked-pos''' , filename='''bool_masked_pos.pt''' ) __lowercase = torch.load(__SCREAMING_SNAKE_CASE ) __lowercase = model(**__SCREAMING_SNAKE_CASE ) __lowercase = outputs.logits __lowercase = [ "videomae-small-finetuned-kinetics", "videomae-small-finetuned-ssv2", # Kinetics-400 checkpoints (short = pretrained only for 800 epochs instead of 1600) "videomae-base-short", "videomae-base-short-finetuned-kinetics", "videomae-base", "videomae-base-finetuned-kinetics", "videomae-large", "videomae-large-finetuned-kinetics", "videomae-huge-finetuned-kinetics", # Something-Something-v2 checkpoints (short = pretrained only for 800 epochs instead of 2400) "videomae-base-short-ssv2", "videomae-base-short-finetuned-ssv2", "videomae-base-ssv2", "videomae-base-finetuned-ssv2", ] # NOTE: logits were tested with image_mean and image_std equal to [0.5, 0.5, 0.5] and [0.5, 0.5, 0.5] if model_name == "videomae-small-finetuned-kinetics": __lowercase = torch.Size([1, 400] ) __lowercase = torch.tensor([-0.9_2_9_1, -0.4_0_6_1, -0.9_3_0_7] ) elif model_name == "videomae-small-finetuned-ssv2": __lowercase = torch.Size([1, 174] ) __lowercase = torch.tensor([0.2_6_7_1, -0.4_6_8_9, -0.8_2_3_5] ) elif model_name == "videomae-base": __lowercase = torch.Size([1, 1408, 1536] ) __lowercase = torch.tensor([[0.7_7_3_9, 0.7_9_6_8, 0.7_0_8_9], [0.6_7_0_1, 0.7_4_8_7, 0.6_2_0_9], [0.4_2_8_7, 0.5_1_5_8, 0.4_7_7_3]] ) elif model_name == "videomae-base-short": __lowercase = torch.Size([1, 1408, 1536] ) __lowercase = torch.tensor([[0.7_9_9_4, 0.9_6_1_2, 0.8_5_0_8], [0.7_4_0_1, 0.8_9_5_8, 0.8_3_0_2], [0.5_8_6_2, 0.7_4_6_8, 0.7_3_2_5]] ) # we verified the loss both for normalized and unnormalized targets for this one __lowercase = torch.tensor([0.5_1_4_2] ) if config.norm_pix_loss else torch.tensor([0.6_4_6_9] ) elif model_name == "videomae-large": __lowercase = torch.Size([1, 1408, 1536] ) __lowercase = torch.tensor([[0.7_1_4_9, 0.7_9_9_7, 0.6_9_6_6], [0.6_7_6_8, 0.7_8_6_9, 0.6_9_4_8], [0.5_1_3_9, 0.6_2_2_1, 0.5_6_0_5]] ) elif model_name == "videomae-large-finetuned-kinetics": __lowercase = torch.Size([1, 400] ) __lowercase = torch.tensor([0.0_7_7_1, 0.0_0_1_1, -0.3_6_2_5] ) elif model_name == "videomae-huge-finetuned-kinetics": __lowercase = torch.Size([1, 400] ) __lowercase = torch.tensor([0.2_4_3_3, 0.1_6_3_2, -0.4_8_9_4] ) elif model_name == "videomae-base-short-finetuned-kinetics": __lowercase = torch.Size([1, 400] ) __lowercase = torch.tensor([0.6_5_8_8, 0.0_9_9_0, -0.2_4_9_3] ) elif model_name == "videomae-base-finetuned-kinetics": __lowercase = torch.Size([1, 400] ) __lowercase = torch.tensor([0.3_6_6_9, -0.0_6_8_8, -0.2_4_2_1] ) elif model_name == "videomae-base-short-ssv2": __lowercase = torch.Size([1, 1408, 1536] ) __lowercase = torch.tensor([[0.4_7_1_2, 0.5_2_9_6, 0.5_7_8_6], [0.2_2_7_8, 0.2_7_2_9, 0.4_0_2_6], [0.0_3_5_2, 0.0_7_3_0, 0.2_5_0_6]] ) elif model_name == "videomae-base-short-finetuned-ssv2": __lowercase = torch.Size([1, 174] ) __lowercase = torch.tensor([-0.0_5_3_7, -0.1_5_3_9, -0.3_2_6_6] ) elif model_name == "videomae-base-ssv2": __lowercase = torch.Size([1, 1408, 1536] ) __lowercase = torch.tensor([[0.8_1_3_1, 0.8_7_2_7, 0.8_5_4_6], [0.7_3_6_6, 0.9_3_7_7, 0.8_8_7_0], [0.5_9_3_5, 0.8_8_7_4, 0.8_5_6_4]] ) elif model_name == "videomae-base-finetuned-ssv2": __lowercase = torch.Size([1, 174] ) __lowercase = torch.tensor([0.1_9_6_1, -0.8_3_3_7, -0.6_3_8_9] ) else: raise ValueError(F"Model name not supported. Should be one of {model_names}" ) # verify logits assert logits.shape == expected_shape if "finetuned" in model_name: assert torch.allclose(logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) else: print('''Logits:''' , logits[0, :3, :3] ) assert torch.allclose(logits[0, :3, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) print('''Logits ok!''' ) # verify loss, if applicable if model_name == "videomae-base-short": __lowercase = outputs.loss assert torch.allclose(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , atol=1e-4 ) print('''Loss ok!''' ) if pytorch_dump_folder_path is not None: print(F"Saving model and image processor to {pytorch_dump_folder_path}" ) image_processor.save_pretrained(__SCREAMING_SNAKE_CASE ) model.save_pretrained(__SCREAMING_SNAKE_CASE ) if push_to_hub: print('''Pushing to the hub...''' ) model.push_to_hub(__SCREAMING_SNAKE_CASE , organization='''nielsr''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://drive.google.com/u/1/uc?id=1tEhLyskjb755TJ65ptsrafUG2llSwQE1&amp;export=download&amp;confirm=t&amp;uuid=aa3276eb-fb7e-482a-adec-dc7171df14c4''', type=str, help=( '''URL of the original PyTorch checkpoint (on Google Drive) you\'d like to convert. Should be a direct''' ''' download link.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''/Users/nielsrogge/Documents/VideoMAE/Test''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--model_name''', default='''videomae-base''', type=str, help='''Name of the model.''') parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) lowerCAmelCase__ = parser.parse_args() convert_videomae_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { "RUCAIBox/mvp": "https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json", } class lowercase_ (_UpperCAmelCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = '''mvp''' SCREAMING_SNAKE_CASE : Union[str, Any] = ['''past_key_values'''] SCREAMING_SNAKE_CASE : List[str] = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : List[Any] ,lowercase__ : Optional[Any]=5_0_2_6_7 ,lowercase__ : Union[str, Any]=1_0_2_4 ,lowercase__ : List[str]=1_2 ,lowercase__ : Dict=4_0_9_6 ,lowercase__ : int=1_6 ,lowercase__ : Optional[Any]=1_2 ,lowercase__ : Optional[int]=4_0_9_6 ,lowercase__ : Tuple=1_6 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Any=0.0 ,lowercase__ : Optional[Any]="gelu" ,lowercase__ : Tuple=1_0_2_4 ,lowercase__ : str=0.1 ,lowercase__ : List[str]=0.0 ,lowercase__ : Union[str, Any]=0.0 ,lowercase__ : Union[str, Any]=0.0_2 ,lowercase__ : str=0.0 ,lowercase__ : Optional[int]=False ,lowercase__ : Any=True ,lowercase__ : List[str]=1 ,lowercase__ : int=0 ,lowercase__ : Dict=2 ,lowercase__ : int=True ,lowercase__ : Optional[int]=2 ,lowercase__ : Any=2 ,lowercase__ : Optional[int]=False ,lowercase__ : Any=1_0_0 ,lowercase__ : str=8_0_0 ,**lowercase__ : Optional[Any] ,): __lowercase = vocab_size __lowercase = max_position_embeddings __lowercase = d_model __lowercase = encoder_ffn_dim __lowercase = encoder_layers __lowercase = encoder_attention_heads __lowercase = decoder_ffn_dim __lowercase = decoder_layers __lowercase = decoder_attention_heads __lowercase = dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = activation_function __lowercase = init_std __lowercase = encoder_layerdrop __lowercase = decoder_layerdrop __lowercase = classifier_dropout __lowercase = use_cache __lowercase = encoder_layers __lowercase = scale_embedding # scale factor will be sqrt(d_model) if True __lowercase = use_prompt __lowercase = prompt_length __lowercase = prompt_mid_dim super().__init__( pad_token_id=A_ ,bos_token_id=A_ ,eos_token_id=A_ ,is_encoder_decoder=A_ ,decoder_start_token_id=A_ ,forced_eos_token_id=A_ ,**A_ ,) if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' ,A_ ): __lowercase = self.bos_token_id warnings.warn( F"Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. " '''The config can simply be saved and uploaded again to be fixed.''' )
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' lowerCAmelCase__ = [4, 1, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCAmelCase__ = [3, 7, 7, 4, 2, 6, 4, 1, 5, 3, 7, 5] lowerCAmelCase__ = { 0: 'Sunday', 1: 'Monday', 2: 'Tuesday', 3: 'Wednesday', 4: 'Thursday', 5: 'Friday', 6: 'Saturday', } def _A ( A__ , A__ , A__ ): """simple docstring""" assert len(str(_lowerCamelCase ) ) > 2, "year should be in YYYY format" assert 1 <= month <= 12, "month should be between 1 to 12" assert 1 <= day <= 31, "day should be between 1 to 31" # Doomsday algorithm: __lowercase = year // 100 __lowercase = (5 * (century % 4) + 2) % 7 __lowercase = year % 100 __lowercase = centurian % 12 __lowercase = ( (centurian // 12) + centurian_m + (centurian_m // 4) + century_anchor ) % 7 __lowercase = ( DOOMSDAY_NOT_LEAP[month - 1] if (year % 4 != 0) or (centurian == 0 and (year % 400) == 0) else DOOMSDAY_LEAP[month - 1] ) __lowercase = (dooms_day + day - day_anchor) % 7 return WEEK_DAY_NAMES[week_day] if __name__ == "__main__": import doctest doctest.testmod()
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import argparse import intel_extension_for_pytorch as ipex import torch from diffusers import DPMSolverMultistepScheduler, StableDiffusionPipeline lowerCAmelCase__ = argparse.ArgumentParser('''Stable Diffusion script with intel optimization''', add_help=False) parser.add_argument('''--dpm''', action='''store_true''', help='''Enable DPMSolver or not''') parser.add_argument('''--steps''', default=None, type=int, help='''Num inference steps''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = "cpu" lowerCAmelCase__ = "a lovely <dicoo> in red dress and hat, in the snowly and brightly night, with many brighly buildings" lowerCAmelCase__ = "path-to-your-trained-model" lowerCAmelCase__ = StableDiffusionPipeline.from_pretrained(model_id) if args.dpm: lowerCAmelCase__ = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config) lowerCAmelCase__ = pipe.to(device) # to channels last lowerCAmelCase__ = pipe.unet.to(memory_format=torch.channels_last) lowerCAmelCase__ = pipe.vae.to(memory_format=torch.channels_last) lowerCAmelCase__ = pipe.text_encoder.to(memory_format=torch.channels_last) if pipe.requires_safety_checker: lowerCAmelCase__ = pipe.safety_checker.to(memory_format=torch.channels_last) # optimize with ipex lowerCAmelCase__ = torch.randn(2, 4, 64, 64) lowerCAmelCase__ = torch.rand(1) * 999 lowerCAmelCase__ = torch.randn(2, 77, 768) lowerCAmelCase__ = (sample, timestep, encoder_hidden_status) try: lowerCAmelCase__ = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True, sample_input=input_example) except Exception: lowerCAmelCase__ = ipex.optimize(pipe.unet.eval(), dtype=torch.bfloataa, inplace=True) lowerCAmelCase__ = ipex.optimize(pipe.vae.eval(), dtype=torch.bfloataa, inplace=True) lowerCAmelCase__ = ipex.optimize(pipe.text_encoder.eval(), dtype=torch.bfloataa, inplace=True) if pipe.requires_safety_checker: lowerCAmelCase__ = ipex.optimize(pipe.safety_checker.eval(), dtype=torch.bfloataa, inplace=True) # compute lowerCAmelCase__ = 666 lowerCAmelCase__ = torch.Generator(device).manual_seed(seed) lowerCAmelCase__ = {"generator": generator} if args.steps is not None: lowerCAmelCase__ = args.steps with torch.cpu.amp.autocast(enabled=True, dtype=torch.bfloataa): lowerCAmelCase__ = pipe(prompt, **generate_kwargs).images[0] # save image image.save('''generated.png''')
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
import multiprocessing import os from typing import BinaryIO, Optional, Union import fsspec from .. import Dataset, Features, NamedSplit, config from ..formatting import query_table from ..packaged_modules.json.json import Json from ..utils import logging from ..utils.typing import NestedDataStructureLike, PathLike from .abc import AbstractDatasetReader class lowercase_ (A_ ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Any ,lowercase__ : int = None ,lowercase__ : Dict = None ,lowercase__ : Any = None ,lowercase__ : Any = False ,lowercase__ : Union[str, Any] = False ,lowercase__ : Union[str, Any] = None ,lowercase__ : List[str] = None ,**lowercase__ : Tuple ,): super().__init__( lowercase__ ,split=lowercase__ ,features=lowercase__ ,cache_dir=lowercase__ ,keep_in_memory=lowercase__ ,streaming=lowercase__ ,num_proc=lowercase__ ,**lowercase__ ,) __lowercase = field __lowercase = path_or_paths if isinstance(lowercase__ ,lowercase__ ) else {self.split: path_or_paths} __lowercase = Json( cache_dir=lowercase__ ,data_files=lowercase__ ,features=lowercase__ ,field=lowercase__ ,**lowercase__ ,) def SCREAMING_SNAKE_CASE ( self : List[str] ): # Build iterable dataset if self.streaming: __lowercase = self.builder.as_streaming_dataset(split=self.split ) # Build regular (map-style) dataset else: __lowercase = None __lowercase = None __lowercase = None __lowercase = None self.builder.download_and_prepare( download_config=lowercase__ ,download_mode=lowercase__ ,verification_mode=lowercase__ ,base_path=lowercase__ ,num_proc=self.num_proc ,) __lowercase = self.builder.as_dataset( split=self.split ,verification_mode=lowercase__ ,in_memory=self.keep_in_memory ) return dataset class lowercase_ : """simple docstring""" def __init__( self : Dict ,lowercase__ : Optional[int] ,lowercase__ : Tuple ,lowercase__ : Tuple = None ,lowercase__ : Union[str, Any] = None ,**lowercase__ : Union[str, Any] ,): if num_proc is not None and num_proc <= 0: raise ValueError(F"num_proc {num_proc} must be an integer > 0." ) __lowercase = dataset __lowercase = path_or_buf __lowercase = batch_size if batch_size else config.DEFAULT_MAX_BATCH_SIZE __lowercase = num_proc __lowercase = """utf-8""" __lowercase = to_json_kwargs def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.to_json_kwargs.pop('''path_or_buf''' ,lowercase__ ) __lowercase = self.to_json_kwargs.pop('''orient''' ,'''records''' ) __lowercase = self.to_json_kwargs.pop('''lines''' ,True if orient == '''records''' else False ) __lowercase = self.to_json_kwargs.pop('''index''' ,False if orient in ['''split''', '''table'''] else True ) __lowercase = self.to_json_kwargs.pop('''compression''' ,lowercase__ ) if compression not in [None, "infer", "gzip", "bz2", "xz"]: raise NotImplementedError(F"`datasets` currently does not support {compression} compression" ) if isinstance(self.path_or_buf ,(str, bytes, os.PathLike) ): with fsspec.open(self.path_or_buf ,'''wb''' ,compression=lowercase__ ) as buffer: __lowercase = self._write(file_obj=lowercase__ ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**self.to_json_kwargs ) else: if compression: raise NotImplementedError( F"The compression parameter is not supported when writing to a buffer, but compression={compression}" ''' was passed. Please provide a local path instead.''' ) __lowercase = self._write( file_obj=self.path_or_buf ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**self.to_json_kwargs ) return written def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : int ): __lowercase = args __lowercase = query_table( table=self.dataset.data ,key=slice(lowercase__ ,offset + self.batch_size ) ,indices=self.dataset._indices ,) __lowercase = batch.to_pandas().to_json( path_or_buf=lowercase__ ,orient=lowercase__ ,lines=lowercase__ ,index=lowercase__ ,**lowercase__ ) if not json_str.endswith('''\n''' ): json_str += "\n" return json_str.encode(self.encoding ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : int ,**lowercase__ : List[str] ,): __lowercase = 0 if self.num_proc is None or self.num_proc == 1: for offset in logging.tqdm( range(0 ,len(self.dataset ) ,self.batch_size ) ,unit='''ba''' ,disable=not logging.is_progress_bar_enabled() ,desc='''Creating json from Arrow format''' ,): __lowercase = self._batch_json((offset, orient, lines, index, to_json_kwargs) ) written += file_obj.write(lowercase__ ) else: __lowercase = len(self.dataset ), self.batch_size with multiprocessing.Pool(self.num_proc ) as pool: for json_str in logging.tqdm( pool.imap( self._batch_json ,[(offset, orient, lines, index, to_json_kwargs) for offset in range(0 ,lowercase__ ,lowercase__ )] ,) ,total=(num_rows // batch_size) + 1 if num_rows % batch_size else num_rows // batch_size ,unit='''ba''' ,disable=not logging.is_progress_bar_enabled() ,desc='''Creating json from Arrow format''' ,): written += file_obj.write(lowercase__ ) return written
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = { '''microsoft/git-base''': '''https://huggingface.co/microsoft/git-base/resolve/main/config.json''', } class lowercase_ (__A ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = """git_vision_model""" def __init__( self : Dict ,lowercase__ : Optional[Any]=7_6_8 ,lowercase__ : Tuple=3_0_7_2 ,lowercase__ : Optional[int]=1_2 ,lowercase__ : Optional[int]=1_2 ,lowercase__ : Optional[int]=3 ,lowercase__ : Any=2_2_4 ,lowercase__ : List[Any]=1_6 ,lowercase__ : Any="quick_gelu" ,lowercase__ : Any=1e-5 ,lowercase__ : List[Any]=0.0 ,lowercase__ : int=0.0_2 ,**lowercase__ : Tuple ,): super().__init__(**UpperCamelCase__ ) __lowercase = hidden_size __lowercase = intermediate_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = num_channels __lowercase = patch_size __lowercase = image_size __lowercase = initializer_range __lowercase = attention_dropout __lowercase = layer_norm_eps __lowercase = hidden_act @classmethod def SCREAMING_SNAKE_CASE ( cls : int ,lowercase__ : Optional[int] ,**lowercase__ : List[Any] ): cls._set_token_in_kwargs(UpperCamelCase__ ) __lowercase = cls.get_config_dict(UpperCamelCase__ ,**UpperCamelCase__ ) # get the vision config dict if we are loading from GITConfig if config_dict.get('''model_type''' ) == "git": __lowercase = config_dict['vision_config'] if "model_type" in config_dict and hasattr(cls ,'''model_type''' ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(UpperCamelCase__ ,**UpperCamelCase__ ) class lowercase_ (__A ): """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = """git""" def __init__( self : Optional[Any] ,lowercase__ : List[str]=None ,lowercase__ : Optional[Any]=3_0_5_2_2 ,lowercase__ : Tuple=7_6_8 ,lowercase__ : Tuple=6 ,lowercase__ : Tuple=1_2 ,lowercase__ : List[str]=3_0_7_2 ,lowercase__ : Union[str, Any]="gelu" ,lowercase__ : int=0.1 ,lowercase__ : List[str]=0.1 ,lowercase__ : int=1_0_2_4 ,lowercase__ : Any=0.0_2 ,lowercase__ : str=1e-1_2 ,lowercase__ : Dict=0 ,lowercase__ : Optional[int]="absolute" ,lowercase__ : str=True ,lowercase__ : str=False ,lowercase__ : Union[str, Any]=1_0_1 ,lowercase__ : List[str]=1_0_2 ,lowercase__ : Any=None ,**lowercase__ : str ,): super().__init__(bos_token_id=UpperCamelCase__ ,eos_token_id=UpperCamelCase__ ,pad_token_id=UpperCamelCase__ ,**UpperCamelCase__ ) if vision_config is None: __lowercase = {} logger.info('''vision_config is None. initializing the GitVisionConfig with default values.''' ) __lowercase = GitVisionConfig(**UpperCamelCase__ ) __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = hidden_act __lowercase = intermediate_size __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = initializer_range __lowercase = layer_norm_eps __lowercase = position_embedding_type __lowercase = use_cache __lowercase = tie_word_embeddings __lowercase = num_image_with_embedding __lowercase = bos_token_id __lowercase = eos_token_id def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = copy.deepcopy(self.__dict__ ) __lowercase = self.vision_config.to_dict() __lowercase = self.__class__.model_type return output
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : str ,lowercase__ : List[Any]=3 ,lowercase__ : Union[str, Any]=3_2 ,lowercase__ : str=3 ,lowercase__ : Any=1_0 ,lowercase__ : List[Any]=[1_0, 2_0, 3_0, 4_0] ,lowercase__ : int=[1, 1, 2, 1] ,lowercase__ : Optional[Any]=True ,lowercase__ : Any=True ,lowercase__ : str="relu" ,lowercase__ : Tuple=3 ,lowercase__ : Union[str, Any]=None ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = num_channels __lowercase = embeddings_size __lowercase = hidden_sizes __lowercase = depths __lowercase = is_training __lowercase = use_labels __lowercase = hidden_act __lowercase = num_labels __lowercase = scope __lowercase = len(lowerCamelCase_ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = self.get_config() return config, pixel_values def SCREAMING_SNAKE_CASE ( self : Any ): return RegNetConfig( num_channels=self.num_channels ,embeddings_size=self.embeddings_size ,hidden_sizes=self.hidden_sizes ,depths=self.depths ,hidden_act=self.hidden_act ,num_labels=self.num_labels ,image_size=self.image_size ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Optional[int] ,lowercase__ : List[str] ): __lowercase = FlaxRegNetModel(config=lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape ,(self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) ,) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : str ,lowercase__ : Optional[int] ): __lowercase = self.num_labels __lowercase = FlaxRegNetForImageClassification(config=lowerCamelCase_ ) __lowercase = model(lowerCamelCase_ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_flax class lowercase_ (a__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : str = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : List[Any] = False def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = FlaxRegNetModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowerCamelCase_ ,has_text_modality=lowerCamelCase_ ) def SCREAMING_SNAKE_CASE ( self : Dict ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowerCamelCase_ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowerCamelCase_ ) @unittest.skip(reason='''RegNet does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): pass @unittest.skip(reason='''RegNet does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowerCamelCase_ ) __lowercase = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowerCamelCase_ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): def check_hidden_states_output(lowercase__ : Dict ,lowercase__ : Dict ,lowercase__ : Optional[Any] ): __lowercase = model_class(lowerCamelCase_ ) __lowercase = model(**self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) ) __lowercase = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states __lowercase = self.model_tester.num_stages self.assertEqual(len(lowerCamelCase_ ) ,expected_num_stages + 1 ) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowerCamelCase_ ,lowerCamelCase_ ,lowerCamelCase_ ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): __lowercase = self._prepare_for_class(lowerCamelCase_ ,lowerCamelCase_ ) __lowercase = model_class(lowerCamelCase_ ) @jax.jit def model_jitted(lowercase__ : Optional[int] ,**lowercase__ : Union[str, Any] ): return model(pixel_values=lowerCamelCase_ ,**lowerCamelCase_ ) with self.subTest('''JIT Enabled''' ): __lowercase = model_jitted(**lowerCamelCase_ ).to_tuple() with self.subTest('''JIT Disabled''' ): with jax.disable_jit(): __lowercase = model_jitted(**lowerCamelCase_ ).to_tuple() self.assertEqual(len(lowerCamelCase_ ) ,len(lowerCamelCase_ ) ) for jitted_output, output in zip(lowerCamelCase_ ,lowerCamelCase_ ): self.assertEqual(jitted_output.shape ,output.shape ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_flax class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return AutoImageProcessor.from_pretrained('''facebook/regnet-y-040''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = FlaxRegNetForImageClassification.from_pretrained('''facebook/regnet-y-040''' ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowerCamelCase_ ,return_tensors='''np''' ) __lowercase = model(**lowerCamelCase_ ) # verify the logits __lowercase = (1, 1_0_0_0) self.assertEqual(outputs.logits.shape ,lowerCamelCase_ ) __lowercase = jnp.array([-0.4_1_8_0, -1.5_0_5_1, -3.4_8_3_6] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] ,lowerCamelCase_ ,atol=1e-4 ) )
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' import logging import re import pytorch_quantization import pytorch_quantization.nn as quant_nn import torch from pytorch_quantization import calib from pytorch_quantization.tensor_quant import QuantDescriptor lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = 50 # max width of layer names lowerCAmelCase__ = 70 # max width of quantizer names def _A ( A__ ): """simple docstring""" __lowercase = parser.add_argument_group('''quant_trainer arguments''' ) group.add_argument('''--wprec''' , type=A__ , default=8 , help='''weight precision''' ) group.add_argument('''--aprec''' , type=A__ , default=8 , help='''activation precision''' ) group.add_argument('''--quant-per-tensor''' , action='''store_true''' , help='''per tensor weight scaling''' ) group.add_argument('''--quant-disable''' , action='''store_true''' , help='''disable all quantizers''' ) group.add_argument('''--quant-disable-embeddings''' , action='''store_true''' , help='''disable all embeddings quantizers''' ) group.add_argument('''--quant-disable-keyword''' , type=A__ , nargs='''+''' , help='''disable quantizers by keyword''' ) group.add_argument('''--quant-disable-layer-module''' , type=A__ , help='''disable quantizers by keyword under layer.''' ) group.add_argument('''--quant-enable-layer-module''' , type=A__ , help='''enable quantizers by keyword under layer''' ) group.add_argument('''--calibrator''' , default='''max''' , help='''which quantization range calibrator to use''' ) group.add_argument('''--percentile''' , default=A__ , type=A__ , help='''percentile for PercentileCalibrator''' ) group.add_argument('''--fuse-qkv''' , action='''store_true''' , help='''use the same scale factor for qkv''' ) group.add_argument('''--clip-gelu''' , metavar='''N''' , type=A__ , help='''clip gelu output maximum value to N''' ) group.add_argument( '''--recalibrate-weights''' , action='''store_true''' , help=( '''recalibrate weight amaxes by taking the max of the weights.''' ''' amaxes will be computed with the current quantization granularity (axis).''' ) , ) def _A ( A__ ): """simple docstring""" if args.calibrator == "max": __lowercase = '''max''' elif args.calibrator == "percentile": if args.percentile is None: raise ValueError('''Specify --percentile when using percentile calibrator''' ) __lowercase = '''histogram''' elif args.calibrator == "mse": __lowercase = '''histogram''' else: raise ValueError(F"Invalid calibrator {args.calibrator}" ) __lowercase = QuantDescriptor(num_bits=args.aprec , calib_method=A__ ) __lowercase = QuantDescriptor(num_bits=args.wprec , axis=(None if args.quant_per_tensor else (0,)) ) quant_nn.QuantLinear.set_default_quant_desc_input(A__ ) quant_nn.QuantLinear.set_default_quant_desc_weight(A__ ) def _A ( A__ , A__ , A__=False , A__=False ): """simple docstring""" logger.info('''Configuring Model for Quantization''' ) logger.info(F"using quantization package {pytorch_quantization.__file__}" ) if not calib: if args.quant_disable_embeddings: set_quantizer_by_name(A__ , ['''embeddings'''] , which='''weight''' , _disabled=A__ ) if args.quant_disable: set_quantizer_by_name(A__ , [''''''] , _disabled=A__ ) if args.quant_disable_keyword: set_quantizer_by_name(A__ , args.quant_disable_keyword , _disabled=A__ ) if args.quant_disable_layer_module: set_quantizer_by_name(A__ , [R'''layer.\d+.''' + args.quant_disable_layer_module] , _disabled=A__ ) if args.quant_enable_layer_module: set_quantizer_by_name(A__ , [R'''layer.\d+.''' + args.quant_enable_layer_module] , _disabled=A__ ) if args.recalibrate_weights: recalibrate_weights(A__ ) if args.fuse_qkv: fuse_qkv(A__ , A__ ) if args.clip_gelu: clip_gelu(A__ , args.clip_gelu ) # if args.local_rank in [-1, 0] and not calib: print_quant_summary(A__ ) def _A ( A__ ): """simple docstring""" logger.info('''Enabling Calibration''' ) for name, module in model.named_modules(): if name.endswith('''_quantizer''' ): if module._calibrator is not None: module.disable_quant() module.enable_calib() else: module.disable() logger.info(F"{name:80}: {module}" ) def _A ( A__ , A__ ): """simple docstring""" logger.info('''Loading calibrated amax''' ) for name, module in model.named_modules(): if name.endswith('''_quantizer''' ): if module._calibrator is not None: if isinstance(module._calibrator , calib.MaxCalibrator ): module.load_calib_amax() else: module.load_calib_amax('''percentile''' , percentile=args.percentile ) module.enable_quant() module.disable_calib() else: module.enable() model.cuda() print_quant_summary(A__ ) def _A ( A__ , A__ ): """simple docstring""" def fusea(A__ , A__ , A__ ): for mod in [qq, qk, qv]: if not hasattr(A__ , '''_amax''' ): print(''' WARNING: NO AMAX BUFFER''' ) return __lowercase = qq._amax.detach().item() __lowercase = qk._amax.detach().item() __lowercase = qv._amax.detach().item() __lowercase = max(A__ , A__ , A__ ) qq._amax.fill_(A__ ) qk._amax.fill_(A__ ) qv._amax.fill_(A__ ) logger.info(F" q={q:5.2f} k={k:5.2f} v={v:5.2f} -> {amax:5.2f}" ) for name, mod in model.named_modules(): if name.endswith('''.attention.self''' ): logger.info(F"FUSE_QKV: {name:{name_width}}" ) fusea(mod.matmul_q_input_quantizer , mod.matmul_k_input_quantizer , mod.matmul_v_input_quantizer ) if args.quant_per_tensor: fusea(mod.query._weight_quantizer , mod.key._weight_quantizer , mod.value._weight_quantizer ) def _A ( A__ , A__ ): """simple docstring""" for name, mod in model.named_modules(): if name.endswith('''.output.dense''' ) and not name.endswith('''attention.output.dense''' ): __lowercase = mod._input_quantizer._amax.data.detach().item() mod._input_quantizer._amax.data.detach().clamp_(max=A__ ) __lowercase = mod._input_quantizer._amax.data.detach().item() logger.info(F"CLIP_GELU: {name:{name_width}} amax: {amax_init:5.2f} -> {amax:5.2f}" ) def _A ( A__ ): """simple docstring""" for name, mod in model.named_modules(): if hasattr(A__ , '''_weight_quantizer''' ) and mod._weight_quantizer.axis is not None: __lowercase = mod.weight.shape[0] __lowercase = mod._weight_quantizer._amax.detach() __lowercase = torch.ones(A__ , dtype=amax.dtype , device=amax.device ) * amax print(F"expanding {name} {amax} -> {mod._weight_quantizer._amax}" ) def _A ( A__ ): """simple docstring""" for name, mod in model.named_modules(): if hasattr(A__ , '''_weight_quantizer''' ): if not hasattr(mod.weight_quantizer , '''_amax''' ): print('''RECALIB: {name:{name_width}} WARNING: NO AMAX BUFFER''' ) continue # determine which axes to reduce across # e.g. a 4D tensor quantized per axis 0 should reduce over (1,2,3) __lowercase = set() if mod._weight_quantizer.axis is None else set(mod._weight_quantizer.axis ) __lowercase = set(range(len(mod.weight.size() ) ) ) - axis_set __lowercase = pytorch_quantization.utils.reduce_amax(mod.weight , axis=A__ , keepdims=A__ ).detach() logger.info(F"RECALIB: {name:{name_width}} {mod._weight_quantizer._amax.flatten()} -> {amax.flatten()}" ) __lowercase = amax def _A ( A__ , A__=25 , A__=180 , A__=None ): """simple docstring""" if ignore is None: __lowercase = [] elif not isinstance(A__ , A__ ): __lowercase = [ignore] __lowercase = 0 for name, mod in model.named_modules(): if not hasattr(A__ , '''weight''' ): continue __lowercase = max(A__ , len(A__ ) ) for name, mod in model.named_modules(): __lowercase = getattr(A__ , '''_input_quantizer''' , A__ ) __lowercase = getattr(A__ , '''_weight_quantizer''' , A__ ) if not hasattr(A__ , '''weight''' ): continue if type(A__ ) in ignore: continue if [True for s in ignore if type(A__ ) is str and s in name]: continue __lowercase = F"Act:{input_q.extra_repr()}" __lowercase = F"Wgt:{weight_q.extra_repr()}" __lowercase = F"{name:{name_width}} {act_str} {wgt_str}" if len(A__ ) <= line_width: logger.info(A__ ) else: logger.info(F"{name:{name_width}} {act_str}" ) logger.info(F"{' ':{name_width}} {wgt_str}" ) def _A ( A__ ): """simple docstring""" __lowercase = 0 for name, mod in model.named_modules(): if isinstance(A__ , pytorch_quantization.nn.TensorQuantizer ): print(F"{name:80} {mod}" ) count += 1 print(F"{count} TensorQuantizers found in model" ) def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = getattr(A__ , A__ , A__ ) if quantizer_mod is not None: assert hasattr(A__ , A__ ) setattr(A__ , A__ , A__ ) else: logger.warning(F"{name} has no {quantizer}" ) def _A ( A__ , A__ , A__="both" , **A__ ): """simple docstring""" __lowercase = F"Warning: changing {which} quantizers of {name:{qname_width}}" for k, v in kwargs.items(): s += F" {k}={v}" if which in ["input", "both"]: set_quantizer(A__ , A__ , '''_input_quantizer''' , A__ , A__ ) if which in ["weight", "both"]: set_quantizer(A__ , A__ , '''_weight_quantizer''' , A__ , A__ ) logger.info(A__ ) def _A ( A__ , A__ , **A__ ): """simple docstring""" for name, mod in model.named_modules(): if hasattr(A__ , '''_input_quantizer''' ) or hasattr(A__ , '''_weight_quantizer''' ): for n in names: if re.search(A__ , A__ ): set_quantizers(A__ , A__ , **A__ ) elif name.endswith('''_quantizer''' ): for n in names: if re.search(A__ , A__ ): __lowercase = F"Warning: changing {name:{name_width}}" for k, v in kwargs.items(): s += F" {k}={v}" setattr(A__ , A__ , A__ ) logger.info(A__ )
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' import warnings warnings.warn( '''memory_utils has been reorganized to utils.memory. Import `find_executable_batchsize` from the main `__init__`: ''' '''`from accelerate import find_executable_batch_size` to avoid this warning.''', FutureWarning, )
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import tempfile import numpy as np import torch from transformers import AutoTokenizer, TaEncoderModel from diffusers import DDPMScheduler, UNetaDConditionModel from diffusers.models.attention_processor import AttnAddedKVProcessor from diffusers.pipelines.deepfloyd_if import IFWatermarker from diffusers.utils.testing_utils import torch_device from ..test_pipelines_common import to_np class lowercase_ : """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): torch.manual_seed(0 ) __lowercase = TaEncoderModel.from_pretrained('''hf-internal-testing/tiny-random-t5''' ) torch.manual_seed(0 ) __lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-t5''' ) torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( sample_size=3_2 ,layers_per_block=1 ,block_out_channels=[3_2, 6_4] ,down_block_types=[ '''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D''', ] ,mid_block_type='''UNetMidBlock2DSimpleCrossAttn''' ,up_block_types=['''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''] ,in_channels=3 ,out_channels=6 ,cross_attention_dim=3_2 ,encoder_hid_dim=3_2 ,attention_head_dim=8 ,addition_embed_type='''text''' ,addition_embed_type_num_heads=2 ,cross_attention_norm='''group_norm''' ,resnet_time_scale_shift='''scale_shift''' ,act_fn='''gelu''' ,) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) __lowercase = DDPMScheduler( num_train_timesteps=1_0_0_0 ,beta_schedule='''squaredcos_cap_v2''' ,beta_start=0.0_0_0_1 ,beta_end=0.0_2 ,thresholding=_UpperCAmelCase ,dynamic_thresholding_ratio=0.9_5 ,sample_max_value=1.0 ,prediction_type='''epsilon''' ,variance_type='''learned_range''' ,) torch.manual_seed(0 ) __lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def SCREAMING_SNAKE_CASE ( self : Any ): torch.manual_seed(0 ) __lowercase = TaEncoderModel.from_pretrained('''hf-internal-testing/tiny-random-t5''' ) torch.manual_seed(0 ) __lowercase = AutoTokenizer.from_pretrained('''hf-internal-testing/tiny-random-t5''' ) torch.manual_seed(0 ) __lowercase = UNetaDConditionModel( sample_size=3_2 ,layers_per_block=[1, 2] ,block_out_channels=[3_2, 6_4] ,down_block_types=[ '''ResnetDownsampleBlock2D''', '''SimpleCrossAttnDownBlock2D''', ] ,mid_block_type='''UNetMidBlock2DSimpleCrossAttn''' ,up_block_types=['''SimpleCrossAttnUpBlock2D''', '''ResnetUpsampleBlock2D'''] ,in_channels=6 ,out_channels=6 ,cross_attention_dim=3_2 ,encoder_hid_dim=3_2 ,attention_head_dim=8 ,addition_embed_type='''text''' ,addition_embed_type_num_heads=2 ,cross_attention_norm='''group_norm''' ,resnet_time_scale_shift='''scale_shift''' ,act_fn='''gelu''' ,class_embed_type='''timestep''' ,mid_block_scale_factor=1.4_1_4 ,time_embedding_act_fn='''gelu''' ,time_embedding_dim=3_2 ,) unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests torch.manual_seed(0 ) __lowercase = DDPMScheduler( num_train_timesteps=1_0_0_0 ,beta_schedule='''squaredcos_cap_v2''' ,beta_start=0.0_0_0_1 ,beta_end=0.0_2 ,thresholding=_UpperCAmelCase ,dynamic_thresholding_ratio=0.9_5 ,sample_max_value=1.0 ,prediction_type='''epsilon''' ,variance_type='''learned_range''' ,) torch.manual_seed(0 ) __lowercase = DDPMScheduler( num_train_timesteps=1_0_0_0 ,beta_schedule='''squaredcos_cap_v2''' ,beta_start=0.0_0_0_1 ,beta_end=0.0_2 ,) torch.manual_seed(0 ) __lowercase = IFWatermarker() return { "text_encoder": text_encoder, "tokenizer": tokenizer, "unet": unet, "scheduler": scheduler, "image_noising_scheduler": image_noising_scheduler, "watermarker": watermarker, "safety_checker": None, "feature_extractor": None, } def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) __lowercase = self.get_dummy_inputs(_UpperCAmelCase ) __lowercase = inputs['''prompt'''] __lowercase = inputs['''generator'''] __lowercase = inputs['''num_inference_steps'''] __lowercase = inputs['''output_type'''] if "image" in inputs: __lowercase = inputs['''image'''] else: __lowercase = None if "mask_image" in inputs: __lowercase = inputs['''mask_image'''] else: __lowercase = None if "original_image" in inputs: __lowercase = inputs['''original_image'''] else: __lowercase = None __lowercase = pipe.encode_prompt(_UpperCAmelCase ) # inputs with prompt converted to embeddings __lowercase = { '''prompt_embeds''': prompt_embeds, '''negative_prompt_embeds''': negative_prompt_embeds, '''generator''': generator, '''num_inference_steps''': num_inference_steps, '''output_type''': output_type, } if image is not None: __lowercase = image if mask_image is not None: __lowercase = mask_image if original_image is not None: __lowercase = original_image # set all optional components to None for optional_component in pipe._optional_components: setattr(_UpperCAmelCase ,_UpperCAmelCase ,_UpperCAmelCase ) __lowercase = pipe(**_UpperCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_UpperCAmelCase ) __lowercase = self.pipeline_class.from_pretrained(_UpperCAmelCase ) pipe_loaded.to(_UpperCAmelCase ) pipe_loaded.set_progress_bar_config(disable=_UpperCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests for optional_component in pipe._optional_components: self.assertTrue( getattr(_UpperCAmelCase ,_UpperCAmelCase ) is None ,F"`{optional_component}` did not stay set to None after loading." ,) __lowercase = self.get_dummy_inputs(_UpperCAmelCase ) __lowercase = inputs['''generator'''] __lowercase = inputs['''num_inference_steps'''] __lowercase = inputs['''output_type'''] # inputs with prompt converted to embeddings __lowercase = { '''prompt_embeds''': prompt_embeds, '''negative_prompt_embeds''': negative_prompt_embeds, '''generator''': generator, '''num_inference_steps''': num_inference_steps, '''output_type''': output_type, } if image is not None: __lowercase = image if mask_image is not None: __lowercase = mask_image if original_image is not None: __lowercase = original_image __lowercase = pipe_loaded(**_UpperCAmelCase )[0] __lowercase = np.abs(to_np(_UpperCAmelCase ) - to_np(_UpperCAmelCase ) ).max() self.assertLess(_UpperCAmelCase ,1e-4 ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.get_dummy_components() __lowercase = self.pipeline_class(**_UpperCAmelCase ) pipe.to(_UpperCAmelCase ) pipe.set_progress_bar_config(disable=_UpperCAmelCase ) __lowercase = self.get_dummy_inputs(_UpperCAmelCase ) __lowercase = pipe(**_UpperCAmelCase )[0] with tempfile.TemporaryDirectory() as tmpdir: pipe.save_pretrained(_UpperCAmelCase ) __lowercase = self.pipeline_class.from_pretrained(_UpperCAmelCase ) pipe_loaded.to(_UpperCAmelCase ) pipe_loaded.set_progress_bar_config(disable=_UpperCAmelCase ) pipe_loaded.unet.set_attn_processor(AttnAddedKVProcessor() ) # For reproducibility tests __lowercase = self.get_dummy_inputs(_UpperCAmelCase ) __lowercase = pipe_loaded(**_UpperCAmelCase )[0] __lowercase = np.abs(to_np(_UpperCAmelCase ) - to_np(_UpperCAmelCase ) ).max() self.assertLess(_UpperCAmelCase ,1e-4 )
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' import json import os import unittest from transformers.models.roc_bert.tokenization_roc_bert import ( VOCAB_FILES_NAMES, RoCBertBasicTokenizer, RoCBertTokenizer, RoCBertWordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class lowercase_ (_a , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = RoCBertTokenizer SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : Union[str, Any] = True SCREAMING_SNAKE_CASE : List[Any] = filter_non_english def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): super().setUp() __lowercase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''[PAD]''', '''[MASK]''', '''你''', '''好''', '''是''', '''谁''', '''a''', '''b''', '''c''', '''d'''] __lowercase = {} __lowercase = {} for i, value in enumerate(_A ): __lowercase = i __lowercase = i __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''vocab_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''word_shape_file'''] ) __lowercase = os.path.join(self.tmpdirname ,VOCAB_FILES_NAMES['''word_pronunciation_file'''] ) with open(self.vocab_file ,'''w''' ,encoding='''utf-8''' ) as vocab_writer: vocab_writer.write(''''''.join([x + '''\n''' for x in vocab_tokens] ) ) with open(self.word_shape_file ,'''w''' ,encoding='''utf-8''' ) as word_shape_writer: json.dump(_A ,_A ,ensure_ascii=_A ) with open(self.word_pronunciation_file ,'''w''' ,encoding='''utf-8''' ) as word_pronunciation_writer: json.dump(_A ,_A ,ensure_ascii=_A ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.tokenizer_class(self.vocab_file ,self.word_shape_file ,self.word_pronunciation_file ) __lowercase = tokenizer.tokenize('''你好[SEP]你是谁''' ) self.assertListEqual(_A ,['''你''', '''好''', '''[SEP]''', '''你''', '''是''', '''谁'''] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(_A ) ,[5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_shape_ids(_A ) ,[5, 6, 2, 5, 7, 8] ) self.assertListEqual(tokenizer.convert_tokens_to_pronunciation_ids(_A ) ,[5, 6, 2, 5, 7, 8] ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = RoCBertBasicTokenizer() self.assertListEqual(tokenizer.tokenize('''ah\u535A\u63A8zz''' ) ,['''ah''', '''\u535A''', '''\u63A8''', '''zz'''] ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) ,['''hello''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) ,['''hello'''] ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ,strip_accents=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) ,['''hällo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) ,['''h\u00E9llo'''] ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ,strip_accents=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) ,['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) ,['''hello'''] ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) ,['''hallo''', '''!''', '''how''', '''are''', '''you''', '''?'''] ) self.assertListEqual(tokenizer.tokenize('''H\u00E9llo''' ) ,['''hello'''] ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? ''' ) ,['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ,strip_accents=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) ,['''HäLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ,strip_accents=_A ) self.assertListEqual( tokenizer.tokenize(''' \tHäLLo!how \n Are yoU? ''' ) ,['''HaLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?'''] ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = RoCBertBasicTokenizer(do_lower_case=_A ,never_split=['''[UNK]'''] ) self.assertListEqual( tokenizer.tokenize(''' \tHeLLo!how \n Are yoU? [UNK]''' ) ,['''HeLLo''', '''!''', '''how''', '''Are''', '''yoU''', '''?''', '''[UNK]'''] ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ['''[UNK]''', '''[CLS]''', '''[SEP]''', '''want''', '''##want''', '''##ed''', '''wa''', '''un''', '''runn''', '''##ing'''] __lowercase = {} for i, token in enumerate(_A ): __lowercase = i __lowercase = RoCBertWordpieceTokenizer(vocab=_A ,unk_token='''[UNK]''' ) self.assertListEqual(tokenizer.tokenize('''''' ) ,[] ) self.assertListEqual(tokenizer.tokenize('''unwanted running''' ) ,['''un''', '''##want''', '''##ed''', '''runn''', '''##ing'''] ) self.assertListEqual(tokenizer.tokenize('''unwantedX running''' ) ,['''[UNK]''', '''runn''', '''##ing'''] ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): self.assertTrue(_is_whitespace(''' ''' ) ) self.assertTrue(_is_whitespace('''\t''' ) ) self.assertTrue(_is_whitespace('''\r''' ) ) self.assertTrue(_is_whitespace('''\n''' ) ) self.assertTrue(_is_whitespace('''\u00A0''' ) ) self.assertFalse(_is_whitespace('''A''' ) ) self.assertFalse(_is_whitespace('''-''' ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.assertTrue(_is_control('''\u0005''' ) ) self.assertFalse(_is_control('''A''' ) ) self.assertFalse(_is_control(''' ''' ) ) self.assertFalse(_is_control('''\t''' ) ) self.assertFalse(_is_control('''\r''' ) ) def SCREAMING_SNAKE_CASE ( self : str ): self.assertTrue(_is_punctuation('''-''' ) ) self.assertTrue(_is_punctuation('''$''' ) ) self.assertTrue(_is_punctuation('''`''' ) ) self.assertTrue(_is_punctuation('''.''' ) ) self.assertFalse(_is_punctuation('''A''' ) ) self.assertFalse(_is_punctuation(''' ''' ) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.get_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(_A ) for t in ['''Test''', '''\xad''', '''test''']] ,[['''[UNK]'''], [], ['''[UNK]''']] ) if self.test_rust_tokenizer: __lowercase = self.get_rust_tokenizer() self.assertListEqual( [rust_tokenizer.tokenize(_A ) for t in ['''Test''', '''\xad''', '''test''']] ,[['''[UNK]'''], [], ['''[UNK]''']] ) def SCREAMING_SNAKE_CASE ( self : Tuple ): for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = self.rust_tokenizer_class.from_pretrained(_A ,**_A ) __lowercase = F"A, naïve {tokenizer_r.mask_token} AllenNLP sentence." __lowercase = tokenizer_r.encode_plus( _A ,return_attention_mask=_A ,return_token_type_ids=_A ,return_offsets_mapping=_A ,add_special_tokens=_A ,) __lowercase = tokenizer_r.do_lower_case if hasattr(_A ,'''do_lower_case''' ) else False __lowercase = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''A'''), ((1, 2), ''','''), ((3, 5), '''na'''), ((5, 6), '''##ï'''), ((6, 8), '''##ve'''), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), '''Allen'''), ((2_1, 2_3), '''##NL'''), ((2_3, 2_4), '''##P'''), ((2_5, 3_3), '''sentence'''), ((3_3, 3_4), '''.'''), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), '''a'''), ((1, 2), ''','''), ((3, 8), '''naive'''), ((9, 1_5), tokenizer_r.mask_token), ((1_6, 2_1), '''allen'''), ((2_1, 2_3), '''##nl'''), ((2_3, 2_4), '''##p'''), ((2_5, 3_3), '''sentence'''), ((3_3, 3_4), '''.'''), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] ,tokenizer_r.convert_ids_to_tokens(tokens['''input_ids'''] ) ) self.assertEqual([e[0] for e in expected_results] ,tokens['''offset_mapping'''] ) def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = ['''的''', '''人''', '''有'''] __lowercase = ''''''.join(_A ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F"{tokenizer.__class__.__name__} ({pretrained_name})" ): __lowercase = True __lowercase = self.tokenizer_class.from_pretrained(_A ,**_A ) __lowercase = self.rust_tokenizer_class.from_pretrained(_A ,**_A ) __lowercase = tokenizer_p.encode(_A ,add_special_tokens=_A ) __lowercase = tokenizer_r.encode(_A ,add_special_tokens=_A ) __lowercase = tokenizer_r.convert_ids_to_tokens(_A ) __lowercase = tokenizer_p.convert_ids_to_tokens(_A ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(_A ,_A ) self.assertListEqual(_A ,_A ) __lowercase = False __lowercase = self.rust_tokenizer_class.from_pretrained(_A ,**_A ) __lowercase = self.tokenizer_class.from_pretrained(_A ,**_A ) __lowercase = tokenizer_r.encode(_A ,add_special_tokens=_A ) __lowercase = tokenizer_p.encode(_A ,add_special_tokens=_A ) __lowercase = tokenizer_r.convert_ids_to_tokens(_A ) __lowercase = tokenizer_p.convert_ids_to_tokens(_A ) # it is expected that only the first Chinese character is not preceded by "##". __lowercase = [ F"##{token}" if idx != 0 else token for idx, token in enumerate(_A ) ] self.assertListEqual(_A ,_A ) self.assertListEqual(_A ,_A ) @slow def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.tokenizer_class(self.vocab_file ,self.word_shape_file ,self.word_pronunciation_file ) __lowercase = tokenizer.encode('''你好''' ,add_special_tokens=_A ) __lowercase = tokenizer.encode('''你是谁''' ,add_special_tokens=_A ) __lowercase = tokenizer.build_inputs_with_special_tokens(_A ) __lowercase = tokenizer.build_inputs_with_special_tokens(_A ,_A ) assert encoded_sentence == [1] + text + [2] assert encoded_pair == [1] + text + [2] + text_a + [2] def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.get_tokenizers(do_lower_case=_A ) for tokenizer in tokenizers: with self.subTest(F"{tokenizer.__class__.__name__}" ): __lowercase = '''你好,你是谁''' __lowercase = tokenizer.tokenize(_A ) __lowercase = tokenizer.convert_tokens_to_ids(_A ) __lowercase = tokenizer.convert_tokens_to_shape_ids(_A ) __lowercase = tokenizer.convert_tokens_to_pronunciation_ids(_A ) __lowercase = tokenizer.prepare_for_model( _A ,_A ,_A ,add_special_tokens=_A ) __lowercase = tokenizer.encode_plus(_A ,add_special_tokens=_A ) self.assertEqual(_A ,_A )
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' def _A ( A__ = 200 ): """simple docstring""" __lowercase = [1, 2, 5, 10, 20, 50, 100, 200] __lowercase = [0] * (pence + 1) __lowercase = 1 # base case: 1 way to make 0 pence for coin in coins: for i in range(UpperCAmelCase__ , pence + 1 , 1 ): number_of_ways[i] += number_of_ways[i - coin] return number_of_ways[pence] if __name__ == "__main__": assert solution(200) == 7_3682
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger lowerCAmelCase__ = get_logger(__name__) class lowercase_ (enum.Enum ): """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = """all_checks""" SCREAMING_SNAKE_CASE : Optional[Any] = """basic_checks""" SCREAMING_SNAKE_CASE : Optional[int] = """no_checks""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" def _A ( A__ , A__ , A__=None ): """simple docstring""" if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) ) if len(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) > 0: raise UnexpectedDownloadedFile(str(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) ) __lowercase = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] __lowercase = ''' for ''' + verification_name if verification_name is not None else '''''' if len(_UpperCamelCase ) > 0: raise NonMatchingChecksumError( F"Checksums didn't match{for_verification_name}:\n" F"{bad_urls}\n" '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" class lowercase_ (lowerCamelCase__ ): """simple docstring""" def _A ( A__ , A__ ): """simple docstring""" if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) > 0: raise ExpectedMoreSplits(str(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) ) if len(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) > 0: raise UnexpectedSplits(str(set(_UpperCamelCase ) - set(_UpperCamelCase ) ) ) __lowercase = [ {'''expected''': expected_splits[name], '''recorded''': recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_UpperCamelCase ) > 0: raise NonMatchingSplitsSizesError(str(_UpperCamelCase ) ) logger.info('''All the splits matched successfully.''' ) def _A ( A__ , A__ = True ): """simple docstring""" if record_checksum: __lowercase = shaaaa() with open(_UpperCamelCase , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , b'''''' ): m.update(_UpperCamelCase ) __lowercase = m.hexdigest() else: __lowercase = None return {"num_bytes": os.path.getsize(_UpperCamelCase ), "checksum": checksum} def _A ( A__ ): """simple docstring""" if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' from tempfile import TemporaryDirectory from unittest import TestCase from unittest.mock import MagicMock, patch from transformers import AutoModel, TFAutoModel from transformers.onnx import FeaturesManager from transformers.testing_utils import SMALL_MODEL_IDENTIFIER, require_tf, require_torch @require_torch @require_tf class lowercase_ (__UpperCAmelCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Tuple ): __lowercase = SMALL_MODEL_IDENTIFIER __lowercase = '''pt''' __lowercase = '''tf''' def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : Any ): __lowercase = AutoModel.from_pretrained(self.test_model ) model_pt.save_pretrained(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : str ): __lowercase = TFAutoModel.from_pretrained(self.test_model ,from_pt=lowercase__ ) model_tf.save_pretrained(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = '''mock_framework''' # Framework provided - return whatever the user provides __lowercase = FeaturesManager.determine_framework(self.test_model ,lowercase__ ) self.assertEqual(lowercase__ ,lowercase__ ) # Local checkpoint and framework provided - return provided framework # PyTorch checkpoint with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(lowercase__ ) __lowercase = FeaturesManager.determine_framework(lowercase__ ,lowercase__ ) self.assertEqual(lowercase__ ,lowercase__ ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(lowercase__ ) __lowercase = FeaturesManager.determine_framework(lowercase__ ,lowercase__ ) self.assertEqual(lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): with TemporaryDirectory() as local_pt_ckpt: self._setup_pt_ckpt(lowercase__ ) __lowercase = FeaturesManager.determine_framework(lowercase__ ) self.assertEqual(lowercase__ ,self.framework_pt ) # TensorFlow checkpoint with TemporaryDirectory() as local_tf_ckpt: self._setup_tf_ckpt(lowercase__ ) __lowercase = FeaturesManager.determine_framework(lowercase__ ) self.assertEqual(lowercase__ ,self.framework_tf ) # Invalid local checkpoint with TemporaryDirectory() as local_invalid_ckpt: with self.assertRaises(lowercase__ ): __lowercase = FeaturesManager.determine_framework(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = MagicMock(return_value=lowercase__ ) with patch('''transformers.onnx.features.is_tf_available''' ,lowercase__ ): __lowercase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(lowercase__ ,self.framework_pt ) # PyTorch not in environment -> use TensorFlow __lowercase = MagicMock(return_value=lowercase__ ) with patch('''transformers.onnx.features.is_torch_available''' ,lowercase__ ): __lowercase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(lowercase__ ,self.framework_tf ) # Both in environment -> use PyTorch __lowercase = MagicMock(return_value=lowercase__ ) __lowercase = MagicMock(return_value=lowercase__ ) with patch('''transformers.onnx.features.is_tf_available''' ,lowercase__ ), patch( '''transformers.onnx.features.is_torch_available''' ,lowercase__ ): __lowercase = FeaturesManager.determine_framework(self.test_model ) self.assertEqual(lowercase__ ,self.framework_pt ) # Both not in environment -> raise error __lowercase = MagicMock(return_value=lowercase__ ) __lowercase = MagicMock(return_value=lowercase__ ) with patch('''transformers.onnx.features.is_tf_available''' ,lowercase__ ), patch( '''transformers.onnx.features.is_torch_available''' ,lowercase__ ): with self.assertRaises(lowercase__ ): __lowercase = FeaturesManager.determine_framework(self.test_model )
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) __lowercase = y[k] + ( (step_size / 2) * (ode_func(A__ , y[k] ) + ode_func(x + step_size , A__ )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available lowerCAmelCase__ = { 'configuration_data2vec_audio': ['DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecAudioConfig'], 'configuration_data2vec_text': [ 'DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecTextConfig', 'Data2VecTextOnnxConfig', ], 'configuration_data2vec_vision': [ 'DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Data2VecVisionConfig', 'Data2VecVisionOnnxConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ 'DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecAudioForAudioFrameClassification', 'Data2VecAudioForCTC', 'Data2VecAudioForSequenceClassification', 'Data2VecAudioForXVector', 'Data2VecAudioModel', 'Data2VecAudioPreTrainedModel', ] lowerCAmelCase__ = [ 'DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecTextForCausalLM', 'Data2VecTextForMaskedLM', 'Data2VecTextForMultipleChoice', 'Data2VecTextForQuestionAnswering', 'Data2VecTextForSequenceClassification', 'Data2VecTextForTokenClassification', 'Data2VecTextModel', 'Data2VecTextPreTrainedModel', ] lowerCAmelCase__ = [ 'DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST', 'Data2VecVisionForImageClassification', 'Data2VecVisionForMaskedImageModeling', 'Data2VecVisionForSemanticSegmentation', 'Data2VecVisionModel', 'Data2VecVisionPreTrainedModel', ] if is_tf_available(): lowerCAmelCase__ = [ 'TFData2VecVisionForImageClassification', 'TFData2VecVisionForSemanticSegmentation', 'TFData2VecVisionModel', 'TFData2VecVisionPreTrainedModel', ] if TYPE_CHECKING: from .configuration_dataavec_audio import DATA2VEC_AUDIO_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecAudioConfig from .configuration_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecTextConfig, DataaVecTextOnnxConfig, ) from .configuration_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_CONFIG_ARCHIVE_MAP, DataaVecVisionConfig, DataaVecVisionOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_dataavec_audio import ( DATA2VEC_AUDIO_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecAudioForAudioFrameClassification, DataaVecAudioForCTC, DataaVecAudioForSequenceClassification, DataaVecAudioForXVector, DataaVecAudioModel, DataaVecAudioPreTrainedModel, ) from .modeling_dataavec_text import ( DATA2VEC_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecTextForCausalLM, DataaVecTextForMaskedLM, DataaVecTextForMultipleChoice, DataaVecTextForQuestionAnswering, DataaVecTextForSequenceClassification, DataaVecTextForTokenClassification, DataaVecTextModel, DataaVecTextPreTrainedModel, ) from .modeling_dataavec_vision import ( DATA2VEC_VISION_PRETRAINED_MODEL_ARCHIVE_LIST, DataaVecVisionForImageClassification, DataaVecVisionForMaskedImageModeling, DataaVecVisionForSemanticSegmentation, DataaVecVisionModel, DataaVecVisionPreTrainedModel, ) if is_tf_available(): from .modeling_tf_dataavec_vision import ( TFDataaVecVisionForImageClassification, TFDataaVecVisionForSemanticSegmentation, TFDataaVecVisionModel, TFDataaVecVisionPreTrainedModel, ) else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' # limitations under the License. from typing import Optional, Tuple, Union import torch from diffusers import DiffusionPipeline, ImagePipelineOutput class lowercase_ (lowercase_ ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : int ): super().__init__() self.register_modules(unet=lowercase__ ,scheduler=lowercase__ ) @torch.no_grad() def __call__( self : str ,lowercase__ : List[str] = 1 ,lowercase__ : int = None ,lowercase__ : List[Any] = 5_0 ,lowercase__ : str = "pil" ,lowercase__ : List[Any] = True ,**lowercase__ : Optional[Any] ,): __lowercase = torch.randn( (batch_size, self.unet.config.in_channels, self.unet.config.sample_size, self.unet.config.sample_size) ,generator=lowercase__ ,) __lowercase = image.to(self.device ) # set step values self.scheduler.set_timesteps(lowercase__ ) for t in self.progress_bar(self.scheduler.timesteps ): # 1. predict noise model_output __lowercase = self.unet(lowercase__ ,lowercase__ ).sample # 2. predict previous mean of image x_t-1 and add variance depending on eta # eta corresponds to η in paper and should be between [0, 1] # do x_t -> x_t-1 __lowercase = self.scheduler.step(lowercase__ ,lowercase__ ,lowercase__ ).prev_sample __lowercase = (image / 2 + 0.5).clamp(0 ,1 ) __lowercase = image.cpu().permute(0 ,2 ,3 ,1 ).numpy() if output_type == "pil": __lowercase = self.numpy_to_pil(lowercase__ ) if not return_dict: return (image,), "This is a local test" return ImagePipelineOutput(images=lowercase__ ), "This is a local test"
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0
import unittest from parameterized import parameterized from transformers import LlamaConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaTokenizer class lowercase_ : """simple docstring""" def __init__( self : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : Any=1_3 ,lowercase__ : Optional[Any]=7 ,lowercase__ : Tuple=True ,lowercase__ : Optional[Any]=True ,lowercase__ : Optional[int]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=9_9 ,lowercase__ : List[Any]=3_2 ,lowercase__ : Tuple=5 ,lowercase__ : Tuple=4 ,lowercase__ : Optional[int]=3_7 ,lowercase__ : Union[str, Any]="gelu" ,lowercase__ : Tuple=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Optional[Any]=5_1_2 ,lowercase__ : Tuple=1_6 ,lowercase__ : List[str]=2 ,lowercase__ : Union[str, Any]=0.0_2 ,lowercase__ : Optional[Any]=3 ,lowercase__ : List[str]=4 ,lowercase__ : str=None ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_labels __lowercase = num_choices __lowercase = scope def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def SCREAMING_SNAKE_CASE ( self : Tuple ): return LlamaConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : Any ,lowercase__ : Optional[int] ,lowercase__ : Union[str, Any] ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ): __lowercase = LlamaModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[Any] ,lowercase__ : Any ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Optional[Any] ,): __lowercase = True __lowercase = LlamaModel(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,encoder_hidden_states=lowercase__ ,encoder_attention_mask=lowercase__ ,) __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,encoder_hidden_states=lowercase__ ,) __lowercase = model(lowercase__ ,attention_mask=lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : Dict ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = LlamaForCausalLM(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : Dict ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Optional[int] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Any ,): __lowercase = True __lowercase = True __lowercase = LlamaForCausalLM(config=lowercase__ ) model.to(lowercase__ ) model.eval() # first forward pass __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,encoder_hidden_states=lowercase__ ,encoder_attention_mask=lowercase__ ,use_cache=lowercase__ ,) __lowercase = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __lowercase = ids_tensor((self.batch_size, 3) ,config.vocab_size ) __lowercase = ids_tensor((self.batch_size, 3) ,vocab_size=2 ) # append to next input_ids and __lowercase = torch.cat([input_ids, next_tokens] ,dim=-1 ) __lowercase = torch.cat([input_mask, next_mask] ,dim=-1 ) __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,encoder_hidden_states=lowercase__ ,encoder_attention_mask=lowercase__ ,output_hidden_states=lowercase__ ,)['''hidden_states'''][0] __lowercase = model( lowercase__ ,attention_mask=lowercase__ ,encoder_hidden_states=lowercase__ ,encoder_attention_mask=lowercase__ ,past_key_values=lowercase__ ,output_hidden_states=lowercase__ ,)['''hidden_states'''][0] # select random slice __lowercase = ids_tensor((1,) ,output_from_past.shape[-1] ).item() __lowercase = output_from_no_past[:, -3:, random_slice_idx].detach() __lowercase = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(lowercase__ ,lowercase__ ,atol=1e-3 ) ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = (LlamaModel, LlamaForCausalLM, LlamaForSequenceClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : List[Any] = (LlamaForCausalLM,) if is_torch_available() else () SCREAMING_SNAKE_CASE : int = ( { """feature-extraction""": LlamaModel, """text-classification""": LlamaForSequenceClassification, """text-generation""": LlamaForCausalLM, """zero-shot""": LlamaForSequenceClassification, } if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : List[str] = False SCREAMING_SNAKE_CASE : Optional[int] = False def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = LlamaModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: __lowercase = type self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = input_dict['''input_ids'''] __lowercase = input_ids.ne(1 ).to(lowercase__ ) __lowercase = ids_tensor([self.model_tester.batch_size] ,self.model_tester.type_sequence_label_size ) __lowercase = LlamaForSequenceClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = '''single_label_classification''' __lowercase = input_dict['''input_ids'''] __lowercase = input_ids.ne(1 ).to(lowercase__ ) __lowercase = ids_tensor([self.model_tester.batch_size] ,self.model_tester.type_sequence_label_size ) __lowercase = LlamaForSequenceClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = 3 __lowercase = '''multi_label_classification''' __lowercase = input_dict['''input_ids'''] __lowercase = input_ids.ne(1 ).to(lowercase__ ) __lowercase = ids_tensor( [self.model_tester.batch_size, config.num_labels] ,self.model_tester.type_sequence_label_size ).to(torch.float ) __lowercase = LlamaForSequenceClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,attention_mask=lowercase__ ,labels=lowercase__ ) self.assertEqual(result.logits.shape ,(self.model_tester.batch_size, self.model_tester.num_labels) ) @unittest.skip('''LLaMA buffers include complex numbers, which breaks this test''' ) def SCREAMING_SNAKE_CASE ( self : List[str] ): pass @parameterized.expand([('''linear''',), ('''dynamic''',)] ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ,lowercase__ : str ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() __lowercase = ids_tensor([1, 1_0] ,config.vocab_size ) __lowercase = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] ,config.vocab_size ) set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights __lowercase = LlamaModel(lowercase__ ) original_model.to(lowercase__ ) original_model.eval() __lowercase = original_model(lowercase__ ).last_hidden_state __lowercase = original_model(lowercase__ ).last_hidden_state set_seed(4_2 ) # Fixed seed at init time so the two models get the same random weights __lowercase = {'''type''': scaling_type, '''factor''': 1_0.0} __lowercase = LlamaModel(lowercase__ ) scaled_model.to(lowercase__ ) scaled_model.eval() __lowercase = scaled_model(lowercase__ ).last_hidden_state __lowercase = scaled_model(lowercase__ ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(lowercase__ ,lowercase__ ,atol=1e-5 ) ) else: self.assertFalse(torch.allclose(lowercase__ ,lowercase__ ,atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(lowercase__ ,lowercase__ ,atol=1e-5 ) ) @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] __lowercase = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-7b-hf''' ,device_map='''auto''' ) __lowercase = model(torch.tensor([input_ids] ) ) # Expected mean on dim = -1 __lowercase = torch.tensor([[-6.6_5_5_0, -4.1_2_2_7, -4.9_8_5_9, -3.2_4_0_6, 0.8_2_6_2, -3.0_0_3_3, 1.2_9_6_4, -3.3_6_9_9]] ) torch.testing.assert_close(out.mean(-1 ) ,lowercase__ ,atol=1e-2 ,rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off __lowercase = torch.tensor([-1_2.8_2_8_1, -7.4_4_5_3, -0.4_6_3_9, -8.0_6_2_5, -7.2_5_0_0, -8.0_0_0_0, -6.4_8_8_3, -7.7_6_9_5, -7.8_4_3_8, -7.0_3_1_2, -6.2_1_8_8, -7.1_3_2_8, -1.8_4_9_6, 1.9_9_6_1, -8.6_2_5_0, -6.7_2_2_7, -1_2.8_2_8_1, -6.9_4_9_2, -7.0_7_4_2, -7.7_8_5_2, -7.5_8_2_0, -7.9_0_6_2, -6.9_3_7_5, -7.9_8_0_5, -8.3_4_3_8, -8.1_5_6_2, -8.0_4_6_9, -7.6_2_5_0, -7.7_4_2_2, -7.3_3_9_8,] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] ,lowercase__ ,atol=1e-5 ,rtol=1e-5 ) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] __lowercase = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-hf''' ,device_map='''auto''' ) __lowercase = model(torch.tensor(lowercase__ ) ) # Expected mean on dim = -1 __lowercase = torch.tensor([[-2.0_6_2_2, -1.2_7_9_4, -1.1_6_3_8, -0.9_7_8_8, -1.4_6_0_3, -1.0_2_3_8, -1.7_8_9_3, -1.4_4_1_1]] ) torch.testing.assert_close(out.mean(-1 ) ,lowercase__ ,atol=1e-2 ,rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off __lowercase = torch.tensor([-8.1_4_0_6, -8.0_5_4_7, 2.7_4_6_1, -1.2_3_4_4, -0.1_4_4_8, -1.8_2_6_2, -1.0_0_2_0, -1.8_1_5_4, -1.6_8_9_5, -1.8_5_1_6, -2.3_5_7_4, -0.9_2_7_7, 3.7_5_9_8, 6.5_7_4_2, -1.2_9_9_8, -0.1_1_7_7, -8.1_4_0_6, -2.9_6_8_8, -2.9_1_9_9, -3.1_6_9_9, -3.5_2_5_4, -2.3_5_5_5, -2.7_9_8_8, -3.4_1_4_1, -2.8_2_6_2, -4.5_1_9_5, -3.3_3_7_9, -3.3_1_6_4, -2.7_8_3_2, -3.0_2_7_3] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] ,lowercase__ ,atol=1e-5 ,rtol=1e-5 ) @unittest.skip('''Logits are not exactly the same, once we fix the instabalities somehow, will update!''' ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] __lowercase = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''' ,device_map='''auto''' ) __lowercase = model(torch.tensor(lowercase__ ) ) # Expected mean on dim = -1 __lowercase = torch.tensor([[-0.8_5_6_2, -1.8_5_2_0, -0.7_5_5_1, -0.4_1_6_2, -1.5_1_6_1, -1.2_0_3_8, -2.4_8_2_3, -2.3_2_5_4]] ) torch.testing.assert_close(out.mean(-1 ) ,lowercase__ ,atol=1e-2 ,rtol=1e-2 ) # slicing logits[0, 0, 0:30] # fmt: off __lowercase = torch.tensor([-2.2_2_2_7, 4.8_8_2_8, 0.9_0_2_3, -0.4_5_7_8, -0.7_8_7_1, -0.1_0_3_3, -0.6_2_2_1, -0.5_7_8_6, -0.7_8_0_3, -1.0_6_7_4, -1.2_9_2_0, -0.1_5_7_0, 0.8_0_0_8, 2.0_7_2_3, -0.9_4_9_7, 0.2_7_7_1, -2.2_2_2_7, -0.7_6_1_2, -1.4_3_4_6, -1.2_0_6_1, -1.6_4_2_6, -0.3_0_0_0, -0.7_1_3_9, -1.1_9_3_4, -1.8_6_9_1, -1.6_9_7_3, -1.5_9_4_7, -1.2_7_0_5, -0.3_5_2_3, -0.5_5_1_3] ) # fmt: on torch.testing.assert_close(out.mean(-1 ) ,lowercase__ ,atol=1e-2 ,rtol=1e-2 ) @unittest.skip( '''Logits are not exactly the same, once we fix the instabalities somehow, will update! Also it is gonna be a `too_slow` test''' ) @slow def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = [1, 3_0_6, 4_6_5_8, 2_7_8, 6_5_9_3, 3_1_0, 2_8_3_4, 3_3_8] __lowercase = LlamaForCausalLM.from_pretrained('''meta-llama/Llama-2-70b-hf''' ,device_map='''auto''' ) __lowercase = model(torch.tensor(lowercase__ ) ) __lowercase = torch.tensor( [[-4.2_3_2_7, -3.3_3_6_0, -4.6_6_6_5, -4.7_6_3_1, -1.8_1_8_0, -3.4_1_7_0, -1.4_2_1_1, -3.1_8_1_0]] ,dtype=torch.floataa ) torch.testing.assert_close(out.mean(-1 ) ,lowercase__ ,atol=1e-2 ,rtol=1e-2 ) # fmt: off __lowercase = torch.tensor([-9.4_9_2_2, -3.9_5_5_1, 1.7_9_9_8, -5.6_7_5_8, -5.1_0_5_5, -5.8_9_8_4, -4.8_3_2_0, -6.8_0_8_6, -6.5_3_9_1, -5.6_1_7_2, -5.5_8_2_0, -5.5_3_5_2, 1.7_8_8_1, 3.6_2_8_9, -6.5_1_1_7, -3.4_7_8_5, -9.5_0_0_0, -6.0_3_5_2, -6.8_1_2_5, -6.0_1_9_5, -6.6_8_3_6, -5.4_7_2_7, -6.2_8_1_2, -6.0_3_9_1, -7.3_3_9_8, -7.4_2_9_7, -7.4_8_4_4, -6.5_8_2_0, -5.8_7_8_9, -5.5_3_1_2] ) # fmt: on torch.testing.assert_close(out[0, 0, :3_0] ,lowercase__ ,atol=1e-5 ,rtol=1e-5 ) @unittest.skip('''Model is curently gated''' ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = '''Simply put, the theory of relativity states that 1) the laws of physics are the same everywhere in the universe and 2) the passage of time and the length of objects can vary depending on the observer\'s frame of reference.\n\nThe first part of the theory, that the laws of physics are the same everywhere, is known as the "princi''' __lowercase = '''Simply put, the theory of relativity states that ''' __lowercase = LlamaTokenizer.from_pretrained('''meta-llama/Llama-2-13b-chat-hf''' ) __lowercase = tokenizer.encode(lowercase__ ,return_tensors='''pt''' ) __lowercase = LlamaForCausalLM.from_pretrained( '''meta-llama/Llama-2-13b-chat-hf''' ,device_map='''sequential''' ,use_safetensors=lowercase__ ) # greedy generation outputs __lowercase = model.generate(lowercase__ ,max_new_tokens=6_4 ,top_p=lowercase__ ,temperature=1 ,do_sample=lowercase__ ) __lowercase = tokenizer.decode(generated_ids[0] ,skip_special_tokens=lowercase__ ) self.assertEqual(lowercase__ ,lowercase__ )
721
'''simple docstring''' from scipy.stats import spearmanr import datasets lowerCAmelCase__ = ''' The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. ''' lowerCAmelCase__ = ''' Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {\'spearmanr\': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric("spearmanr") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results[\'spearmanr\']) -0.7 >>> print(round(results[\'spearmanr_pvalue\'], 2)) 0.19 ''' lowerCAmelCase__ = R'''\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowercase_ (datasets.Metric ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): return datasets.MetricInfo( description=_DESCRIPTION ,citation=_CITATION ,inputs_description=_KWARGS_DESCRIPTION ,features=datasets.Features( { '''predictions''': datasets.Value('''float''' ), '''references''': datasets.Value('''float''' ), } ) ,reference_urls=['''https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html'''] ,) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : Union[str, Any]=False ): __lowercase = spearmanr(lowercase__ ,lowercase__ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
624
0
'''simple docstring''' import argparse import json import os import re import shutil import torch from transformers import BioGptConfig, BioGptForCausalLM from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES from transformers.tokenization_utils_base import TOKENIZER_CONFIG_FILE from transformers.utils import WEIGHTS_NAME, logging logging.set_verbosity_warning() lowerCAmelCase__ = 2 class lowercase_ : def __init__( self : str ,*, # begin keyword-only arguments lowercase__ : Optional[int]="<s>" ,lowercase__ : Any="<pad>" ,lowercase__ : str="</s>" ,lowercase__ : Optional[Any]="<unk>" ,lowercase__ : Optional[Any]=None ,): __lowercase , __lowercase , __lowercase , __lowercase = bos, unk, pad, eos __lowercase = [] __lowercase = [] __lowercase = {} __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) __lowercase = self.add_symbol(lowercase__ ) if extra_special_symbols: for s in extra_special_symbols: self.add_symbol(lowercase__ ) __lowercase = len(self.symbols ) def __eq__( self : Dict ,lowercase__ : int ): return self.indices == other.indices def __getitem__( self : str ,lowercase__ : List[Any] ): if idx < len(self.symbols ): return self.symbols[idx] return self.unk_word def __len__( self : List[str] ): return len(self.symbols ) def __contains__( self : List[str] ,lowercase__ : Optional[int] ): return sym in self.indices @classmethod def SCREAMING_SNAKE_CASE ( cls : str ,lowercase__ : Tuple ): __lowercase = cls() d.add_from_file(lowercase__ ) return d def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[str] ,lowercase__ : Optional[int]=1 ,lowercase__ : Dict=False ): if word in self.indices and not overwrite: __lowercase = self.indices[word] __lowercase = self.count[idx] + n return idx else: __lowercase = len(self.symbols ) __lowercase = idx self.symbols.append(lowercase__ ) self.count.append(lowercase__ ) return idx def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ): return 0 def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : int ): if isinstance(lowercase__ ,lowercase__ ): try: with open(lowercase__ ,'''r''' ,encoding='''utf-8''' ) as fd: self.add_from_file(lowercase__ ) except FileNotFoundError as fnfe: raise fnfe except UnicodeError: raise Exception('''Incorrect encoding detected in {}, please rebuild the dataset'''.format(lowercase__ ) ) return __lowercase = f.readlines() __lowercase = self._load_meta(lowercase__ ) for line in lines[indices_start_line:]: try: __lowercase , __lowercase = line.rstrip().rsplit(''' ''' ,1 ) if field == "#fairseq:overwrite": __lowercase = True __lowercase , __lowercase = line.rsplit(''' ''' ,1 ) else: __lowercase = False __lowercase = int(lowercase__ ) __lowercase = line if word in self and not overwrite: raise RuntimeError( '''Duplicate word found when loading Dictionary: \'{}\'. ''' '''Duplicate words can overwrite earlier ones by adding the ''' '''#fairseq:overwrite flag at the end of the corresponding row ''' '''in the dictionary file. If using the Camembert model, please ''' '''download an updated copy of the model file.'''.format(lowercase__ ) ) self.add_symbol(lowercase__ ,n=lowercase__ ,overwrite=lowercase__ ) except ValueError: raise ValueError('''Incorrect dictionary format, expected \'<token> <cnt> [flags]\'''' ) def _A ( A__ ): """simple docstring""" __lowercase = dict((re.sub(R'''@@$''' , '''''' , A__ ), v) if k.endswith('''@@''' ) else (re.sub(R'''$''' , '''</w>''' , A__ ), v) for k, v in d.items() ) __lowercase = '''<s> <pad> </s> <unk>'''.split() # restore the special tokens for k in keep_keys: del da[F"{k}</w>"] __lowercase = d[k] # restore return da def _A ( A__ , A__ ): """simple docstring""" if not os.path.exists(A__ ): raise ValueError(F"path {biogpt_checkpoint_path} does not exist!" ) os.makedirs(A__ , exist_ok=A__ ) print(F"Writing results to {pytorch_dump_folder_path}" ) # handle various types of models __lowercase = os.path.join(A__ , '''checkpoint.pt''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {checkpoint_file} does not exist!" ) __lowercase = torch.load(A__ , map_location='''cpu''' ) __lowercase = chkpt['''cfg''']['''model'''] # dicts __lowercase = os.path.join(A__ , '''dict.txt''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {dict_file} does not exist!" ) __lowercase = Dictionary.load(A__ ) __lowercase = rewrite_dict_keys(src_dict.indices ) __lowercase = len(A__ ) __lowercase = os.path.join(A__ , VOCAB_FILES_NAMES['''vocab_file'''] ) print(F"Generating {src_vocab_file} of {src_vocab_size} records" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # merges_file (bpecodes) __lowercase = os.path.join(A__ , '''bpecodes''' ) if not os.path.isfile(A__ ): raise ValueError(F"path to the file {bpecodes_file} does not exist!" ) __lowercase = os.path.join(A__ , VOCAB_FILES_NAMES['''merges_file'''] ) shutil.copyfile(A__ , A__ ) # model config __lowercase = os.path.join(A__ , '''config.json''' ) __lowercase = { '''activation_dropout''': args['''activation_dropout'''], '''architectures''': ['''BioGptForCausalLM'''], '''attention_probs_dropout_prob''': args['''attention_dropout'''], '''bos_token_id''': 0, '''eos_token_id''': 2, '''hidden_act''': args['''activation_fn'''], '''hidden_dropout_prob''': args['''dropout'''], '''hidden_size''': args['''decoder_embed_dim'''], '''initializer_range''': 0.0_2, '''intermediate_size''': args['''decoder_ffn_embed_dim'''], '''layer_norm_eps''': 1e-12, '''layerdrop''': args['''decoder_layerdrop'''], '''max_position_embeddings''': args['''max_target_positions'''], '''model_type''': '''biogpt''', '''num_attention_heads''': args['''decoder_attention_heads'''], '''num_hidden_layers''': args['''decoder_layers'''], '''pad_token_id''': 1, '''scale_embedding''': not args['''no_scale_embedding'''], '''tie_word_embeddings''': args['''share_decoder_input_output_embed'''], '''vocab_size''': src_vocab_size, } # good hparam defaults to start with print(F"Generating {biogpt_model_config_file}" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # tokenizer config __lowercase = os.path.join(A__ , A__ ) __lowercase = { '''bos_token''': '''<s>''', '''eos_token''': '''</s>''', '''model_max_length''': 1024, '''pad_token''': '''<pad>''', '''special_tokens_map_file''': None, '''tokenizer_class''': '''BioGptTokenizer''', '''unk_token''': '''<unk>''', } print(F"Generating {biogpt_tokenizer_config_file}" ) with open(A__ , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(A__ , ensure_ascii=A__ , indent=A__ ) ) # model __lowercase = chkpt['''model'''] # remove unneeded keys __lowercase = [ '''decoder.version''', ] for k in ignore_keys: model_state_dict.pop(A__ , A__ ) __lowercase = list(model_state_dict.keys() ) for layer_name in layer_names: if layer_name.endswith('''output_projection.weight''' ): __lowercase = model_state_dict.pop(A__ ) else: __lowercase = model_state_dict.pop(A__ ) __lowercase = BioGptConfig.from_pretrained(A__ ) __lowercase = BioGptForCausalLM(A__ ) # check that it loads ok model_new.load_state_dict(A__ ) # save __lowercase = os.path.join(A__ , A__ ) print(F"Generating {pytorch_weights_dump_path}" ) torch.save(A__ , A__ ) print('''Conversion is done!''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--biogpt_checkpoint_path''', default=None, type=str, required=True, help=( '''Path to the official PyTorch checkpoint file which is expected to reside in the dump dir with dicts,''' ''' bpecodes, etc.''' ), ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_biogpt_checkpoint_to_pytorch(args.biogpt_checkpoint_path, args.pytorch_dump_folder_path)
700
'''simple docstring''' import random from typing import Any def _A ( A__ ): """simple docstring""" for _ in range(len(A__ ) ): __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase = random.randint(0 , len(A__ ) - 1 ) __lowercase , __lowercase = data[b], data[a] return data if __name__ == "__main__": lowerCAmelCase__ = [0, 1, 2, 3, 4, 5, 6, 7] lowerCAmelCase__ = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
624
0
'''simple docstring''' import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class lowercase_ (unittest.TestCase ): """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict=1_3 ,lowercase__ : Any=7 ,lowercase__ : Optional[Any]=True ,lowercase__ : Tuple=True ,lowercase__ : List[str]=True ,lowercase__ : Optional[int]=True ,lowercase__ : int=9_9 ,lowercase__ : List[str]=3_2 ,lowercase__ : Any=5 ,lowercase__ : Union[str, Any]=4 ,lowercase__ : Tuple=3_7 ,lowercase__ : int="gelu" ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Union[str, Any]=0.1 ,lowercase__ : Optional[int]=5_1_2 ,lowercase__ : Union[str, Any]=1_6 ,lowercase__ : Optional[Any]=2 ,lowercase__ : Optional[int]=0.0_2 ,lowercase__ : Union[str, Any]=4 ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_attention_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = type_sequence_label_size __lowercase = initializer_range __lowercase = num_choices def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_attention_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = RobertaConfig( vocab_size=self.vocab_size ,hidden_size=self.hidden_size ,num_hidden_layers=self.num_hidden_layers ,num_attention_heads=self.num_attention_heads ,intermediate_size=self.intermediate_size ,hidden_act=self.hidden_act ,hidden_dropout_prob=self.hidden_dropout_prob ,attention_probs_dropout_prob=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,is_decoder=lowercase__ ,initializer_range=self.initializer_range ,) return config, input_ids, token_type_ids, attention_mask def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': attention_mask} return config, inputs_dict def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = True __lowercase = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = True SCREAMING_SNAKE_CASE : Dict = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): __lowercase = FlaxRobertaModelTester(self ) @slow def SCREAMING_SNAKE_CASE ( self : int ): for model_class_name in self.all_model_classes: __lowercase = model_class_name.from_pretrained('''roberta-base''' ,from_pt=lowercase__ ) __lowercase = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowercase__ )
701
'''simple docstring''' import argparse import json import os import torch from transformers.file_utils import has_file from diffusers import UNetaDConditionModel, UNetaDModel lowerCAmelCase__ = False lowerCAmelCase__ = True lowerCAmelCase__ = False if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--repo_path''', default=None, type=str, required=True, help='''The config json file corresponding to the architecture.''', ) parser.add_argument('''--dump_path''', default=None, type=str, required=True, help='''Path to the output model.''') lowerCAmelCase__ = parser.parse_args() lowerCAmelCase__ = { '''image_size''': '''sample_size''', '''num_res_blocks''': '''layers_per_block''', '''block_channels''': '''block_out_channels''', '''down_blocks''': '''down_block_types''', '''up_blocks''': '''up_block_types''', '''downscale_freq_shift''': '''freq_shift''', '''resnet_num_groups''': '''norm_num_groups''', '''resnet_act_fn''': '''act_fn''', '''resnet_eps''': '''norm_eps''', '''num_head_channels''': '''attention_head_dim''', } lowerCAmelCase__ = { '''time_steps''': '''time_proj''', '''mid''': '''mid_block''', '''downsample_blocks''': '''down_blocks''', '''upsample_blocks''': '''up_blocks''', } lowerCAmelCase__ = '''''' if has_file(args.repo_path, '''config.json''') else '''unet''' with open(os.path.join(args.repo_path, subfolder, '''config.json'''), '''r''', encoding='''utf-8''') as reader: lowerCAmelCase__ = reader.read() lowerCAmelCase__ = json.loads(text) if do_only_config: for key in config_parameters_to_change.keys(): config.pop(key, None) if has_file(args.repo_path, '''config.json'''): lowerCAmelCase__ = UNetaDModel(**config) else: lowerCAmelCase__ = UNetaDConditionModel if '''ldm-text2im-large-256''' in args.repo_path else UNetaDModel lowerCAmelCase__ = class_name(**config) if do_only_config: model.save_config(os.path.join(args.repo_path, subfolder)) lowerCAmelCase__ = dict(model.config) if do_only_renaming: for key, value in config_parameters_to_change.items(): if key in config: lowerCAmelCase__ = config[key] del config[key] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''down_block_types''']] lowerCAmelCase__ = [k.replace('''UNetRes''', '''''') for k in config['''up_block_types''']] if do_only_weights: lowerCAmelCase__ = torch.load(os.path.join(args.repo_path, subfolder, '''diffusion_pytorch_model.bin''')) lowerCAmelCase__ = {} for param_key, param_value in state_dict.items(): if param_key.endswith('''.op.bias''') or param_key.endswith('''.op.weight'''): continue lowerCAmelCase__ = False for key, new_key in key_parameters_to_change.items(): if not has_changed and param_key.split('''.''')[0] == key: lowerCAmelCase__ = param_value lowerCAmelCase__ = True if not has_changed: lowerCAmelCase__ = param_value model.load_state_dict(new_state_dict) model.save_pretrained(os.path.join(args.repo_path, subfolder))
624
0
'''simple docstring''' import unittest from transformers import is_torch_available from transformers.testing_utils import require_torch if is_torch_available(): import torch from transformers.generation import DisjunctiveConstraint @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): # For consistency across different places the DisjunctiveConstraint is called, # dc.token_ids is a list of integers. It is also initialized only by integers. __lowercase = [[1, 2, 4], [1, 2, 3, 4]] __lowercase = DisjunctiveConstraint(lowercase__ ) self.assertTrue(isinstance(dc.token_ids ,lowercase__ ) ) with self.assertRaises(lowercase__ ): DisjunctiveConstraint(torch.LongTensor([[1, 2, 4], [1, 2, 3]] ) ) with self.assertRaises(lowercase__ ): DisjunctiveConstraint([torch.LongTensor([1, 2, 4] ), torch.LongTensor([1, 2, 3, 4, 5] )] ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): # We can't have constraints that are complete subsets of another. This leads to a preverse # interpretation of "constraint fulfillment": does generating [1,2,3] fulfill the constraint? # It would mean that it generated [1,2] which fulfills it, but it's in the middle of potentially # fulfilling [1,2,3,4]. If we believe that [1,2,3] does fulfill the constraint, then the algorithm # will necessarily never reach [1,2,3,4], giving users a false sense of control (better to just not allow it). __lowercase = [[1, 2], [1, 2, 3, 4]] with self.assertRaises(lowercase__ ): DisjunctiveConstraint(lowercase__ ) # fails here def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = [[1, 2, 3], [1, 2, 4]] __lowercase = DisjunctiveConstraint(lowercase__ ) __lowercase , __lowercase , __lowercase = dc.update(1 ) __lowercase = stepped is True and completed is False and reset is False self.assertTrue(lowercase__ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowercase , __lowercase , __lowercase = dc.update(2 ) __lowercase = stepped is True and completed is False and reset is False self.assertTrue(lowercase__ ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowercase , __lowercase , __lowercase = dc.update(3 ) __lowercase = stepped is True and completed is True and reset is False self.assertTrue(lowercase__ ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 3] ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = [[1, 2, 3], [1, 2, 4, 5], [1, 2, 5]] __lowercase = DisjunctiveConstraint(lowercase__ ) __lowercase , __lowercase , __lowercase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1] ) __lowercase , __lowercase , __lowercase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2] ) __lowercase , __lowercase , __lowercase = dc.update(4 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.current_seq == [1, 2, 4] ) __lowercase , __lowercase , __lowercase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.current_seq == [1, 2, 4, 5] ) dc.reset() __lowercase , __lowercase , __lowercase = dc.update(1 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 3 ) self.assertTrue(dc.current_seq == [1] ) __lowercase , __lowercase , __lowercase = dc.update(2 ) self.assertTrue(not dc.completed ) self.assertTrue(dc.remaining() == 2 ) self.assertTrue(dc.current_seq == [1, 2] ) __lowercase , __lowercase , __lowercase = dc.update(5 ) self.assertTrue(dc.completed ) # Completed! self.assertTrue(dc.remaining() == 0 ) self.assertTrue(dc.current_seq == [1, 2, 5] )
702
'''simple docstring''' import inspect import unittest from math import floor from transformers import CvtConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import CvtForImageClassification, CvtModel from transformers.models.cvt.modeling_cvt import CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class lowercase_ (lowerCamelCase__ ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase__ ,'''embed_dim''' ) ) self.parent.assertTrue(hasattr(lowercase__ ,'''num_heads''' ) ) class lowercase_ : """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Optional[int]=1_3 ,lowercase__ : List[Any]=6_4 ,lowercase__ : Optional[int]=3 ,lowercase__ : Dict=[1_6, 4_8, 9_6] ,lowercase__ : Optional[Any]=[1, 3, 6] ,lowercase__ : Tuple=[1, 2, 1_0] ,lowercase__ : Optional[int]=[7, 3, 3] ,lowercase__ : str=[4, 2, 2] ,lowercase__ : Dict=[2, 1, 1] ,lowercase__ : Tuple=[2, 2, 2] ,lowercase__ : Tuple=[False, False, True] ,lowercase__ : int=[0.0, 0.0, 0.0] ,lowercase__ : str=0.0_2 ,lowercase__ : Union[str, Any]=1e-1_2 ,lowercase__ : Optional[int]=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Optional[Any]=2 ,): __lowercase = parent __lowercase = batch_size __lowercase = image_size __lowercase = patch_sizes __lowercase = patch_stride __lowercase = patch_padding __lowercase = is_training __lowercase = use_labels __lowercase = num_labels __lowercase = num_channels __lowercase = embed_dim __lowercase = num_heads __lowercase = stride_kv __lowercase = depth __lowercase = cls_token __lowercase = attention_drop_rate __lowercase = initializer_range __lowercase = layer_norm_eps def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.num_labels ) __lowercase = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE ( self : str ): return CvtConfig( image_size=self.image_size ,num_labels=self.num_labels ,num_channels=self.num_channels ,embed_dim=self.embed_dim ,num_heads=self.num_heads ,patch_sizes=self.patch_sizes ,patch_padding=self.patch_padding ,patch_stride=self.patch_stride ,stride_kv=self.stride_kv ,depth=self.depth ,cls_token=self.cls_token ,attention_drop_rate=self.attention_drop_rate ,initializer_range=self.initializer_range ,) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Any ,lowercase__ : List[str] ): __lowercase = CvtModel(config=lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ) __lowercase = (self.image_size, self.image_size) __lowercase , __lowercase = image_size[0], image_size[1] for i in range(len(self.depth ) ): __lowercase = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) __lowercase = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.embed_dim[-1], height, width) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Dict ): __lowercase = self.num_labels __lowercase = CvtForImageClassification(lowercase__ ) model.to(lowercase__ ) model.eval() __lowercase = model(lowercase__ ,labels=lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.prepare_config_and_inputs() __lowercase , __lowercase , __lowercase = config_and_inputs __lowercase = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Any = (CvtModel, CvtForImageClassification) if is_torch_available() else () SCREAMING_SNAKE_CASE : Optional[int] = ( {'feature-extraction': CvtModel, 'image-classification': CvtForImageClassification} if is_torch_available() else {} ) SCREAMING_SNAKE_CASE : Optional[int] = False SCREAMING_SNAKE_CASE : Union[str, Any] = False SCREAMING_SNAKE_CASE : int = False SCREAMING_SNAKE_CASE : Optional[Any] = False SCREAMING_SNAKE_CASE : str = False def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ,has_text_modality=lowercase__ ,hidden_size=3_7 ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def SCREAMING_SNAKE_CASE ( self : str ): return @unittest.skip(reason='''Cvt does not output attentions''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @unittest.skip(reason='''Cvt does not support input and output embeddings''' ) def SCREAMING_SNAKE_CASE ( self : str ): pass def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = model_class(lowercase__ ) __lowercase = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __lowercase = [*signature.parameters.keys()] __lowercase = ['''pixel_values'''] self.assertListEqual(arg_names[:1] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ): def check_hidden_states_output(lowercase__ : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : str ): __lowercase = model_class(lowercase__ ) model.to(lowercase__ ) model.eval() with torch.no_grad(): __lowercase = model(**self._prepare_for_class(lowercase__ ,lowercase__ ) ) __lowercase = outputs.hidden_states __lowercase = len(self.model_tester.depth ) self.assertEqual(len(lowercase__ ) ,lowercase__ ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) ,[ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] ,) __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __lowercase = True check_hidden_states_output(lowercase__ ,lowercase__ ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase__ ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE ( self : Any ): pass @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for model_name in CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase = CvtModel.from_pretrained(lowercase__ ) self.assertIsNotNone(lowercase__ ) def _A ( ): """simple docstring""" __lowercase = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class lowercase_ (unittest.TestCase ): """simple docstring""" @cached_property def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): return AutoImageProcessor.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = CvtForImageClassification.from_pretrained(CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(lowercase__ ) __lowercase = self.default_image_processor __lowercase = prepare_img() __lowercase = image_processor(images=lowercase__ ,return_tensors='''pt''' ).to(lowercase__ ) # forward pass with torch.no_grad(): __lowercase = model(**lowercase__ ) # verify the logits __lowercase = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape ,lowercase__ ) __lowercase = torch.tensor([0.9_2_8_5, 0.9_0_1_5, -0.3_1_5_0] ).to(lowercase__ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] ,lowercase__ ,atol=1e-4 ) )
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
703
'''simple docstring''' def _A ( ): """simple docstring""" for n in range(1 , 1000000 ): yield n * (n + 1) // 2 def _A ( A__ ): """simple docstring""" __lowercase = 1 __lowercase = 2 while i * i <= n: __lowercase = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _A ( ): """simple docstring""" return next(i for i in triangle_number_generator() if count_divisors(A__ ) > 500 ) if __name__ == "__main__": print(solution())
624
0
'''simple docstring''' import collections from typing import List, Optional, Union from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, add_end_docstrings, add_start_docstrings, logging from ..bert.tokenization_bert_fast import BertTokenizerFast from .tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer, DPRReaderTokenizer lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''vocab.txt''', '''tokenizer_file''': '''tokenizer.json'''} lowerCAmelCase__ = { '''vocab_file''': { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/vocab.txt''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/tokenizer.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase__ = { '''vocab_file''': { '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/vocab.txt''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/tokenizer.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase__ = { '''vocab_file''': { '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/vocab.txt''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/vocab.txt''' ), }, '''tokenizer_file''': { '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/tokenizer.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/tokenizer.json''' ), }, } lowerCAmelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': 512, '''facebook/dpr-ctx_encoder-multiset-base''': 512, } lowerCAmelCase__ = { '''facebook/dpr-question_encoder-single-nq-base''': 512, '''facebook/dpr-question_encoder-multiset-base''': 512, } lowerCAmelCase__ = { '''facebook/dpr-reader-single-nq-base''': 512, '''facebook/dpr-reader-multiset-base''': 512, } lowerCAmelCase__ = { '''facebook/dpr-ctx_encoder-single-nq-base''': {'''do_lower_case''': True}, '''facebook/dpr-ctx_encoder-multiset-base''': {'''do_lower_case''': True}, } lowerCAmelCase__ = { '''facebook/dpr-question_encoder-single-nq-base''': {'''do_lower_case''': True}, '''facebook/dpr-question_encoder-multiset-base''': {'''do_lower_case''': True}, } lowerCAmelCase__ = { '''facebook/dpr-reader-single-nq-base''': {'''do_lower_case''': True}, '''facebook/dpr-reader-multiset-base''': {'''do_lower_case''': True}, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Dict = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Dict = CONTEXT_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = CONTEXT_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : List[Any] = CONTEXT_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : str = DPRContextEncoderTokenizer class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : int = QUESTION_ENCODER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Optional[int] = QUESTION_ENCODER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Tuple = QUESTION_ENCODER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : Any = DPRQuestionEncoderTokenizer lowerCAmelCase__ = collections.namedtuple( '''DPRSpanPrediction''', ['''span_score''', '''relevance_score''', '''doc_id''', '''start_index''', '''end_index''', '''text'''] ) lowerCAmelCase__ = collections.namedtuple('''DPRReaderOutput''', ['''start_logits''', '''end_logits''', '''relevance_logits''']) lowerCAmelCase__ = R''' Return a dictionary with the token ids of the input strings and other information to give to `.decode_best_spans`. It converts the strings of a question and different passages (title and text) in a sequence of IDs (integers), using the tokenizer and vocabulary. The resulting `input_ids` is a matrix of size `(n_passages, sequence_length)` with the format: [CLS] <question token ids> [SEP] <titles ids> [SEP] <texts ids> Args: questions (`str` or `List[str]`): The questions to be encoded. You can specify one question for many passages. In this case, the question will be duplicated like `[questions] * n_passages`. Otherwise you have to specify as many questions as in `titles` or `texts`. titles (`str` or `List[str]`): The passages titles to be encoded. This can be a string or a list of strings if there are several passages. texts (`str` or `List[str]`): The passages texts to be encoded. This can be a string or a list of strings if there are several passages. padding (`bool`, `str` or [`~utils.PaddingStrategy`], *optional*, defaults to `False`): Activates and controls padding. Accepts the following values: - `True` or `\'longest\'`: Pad to the longest sequence in the batch (or no padding if only a single sequence if provided). - `\'max_length\'`: Pad to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. - `False` or `\'do_not_pad\'` (default): No padding (i.e., can output a batch with sequences of different lengths). truncation (`bool`, `str` or [`~tokenization_utils_base.TruncationStrategy`], *optional*, defaults to `False`): Activates and controls truncation. Accepts the following values: - `True` or `\'longest_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will truncate token by token, removing a token from the longest sequence in the pair if a pair of sequences (or a batch of pairs) is provided. - `\'only_first\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the first sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `\'only_second\'`: Truncate to a maximum length specified with the argument `max_length` or to the maximum acceptable input length for the model if that argument is not provided. This will only truncate the second sequence of a pair if a pair of sequences (or a batch of pairs) is provided. - `False` or `\'do_not_truncate\'` (default): No truncation (i.e., can output batch with sequence lengths greater than the model maximum admissible input size). max_length (`int`, *optional*): Controls the maximum length to use by one of the truncation/padding parameters. If left unset or set to `None`, this will use the predefined model maximum length if a maximum length is required by one of the truncation/padding parameters. If the model has no specific maximum input length (like XLNet) truncation/padding to a maximum length will be deactivated. return_tensors (`str` or [`~utils.TensorType`], *optional*): If set, will return tensors instead of list of python integers. Acceptable values are: - `\'tf\'`: Return TensorFlow `tf.constant` objects. - `\'pt\'`: Return PyTorch `torch.Tensor` objects. - `\'np\'`: Return Numpy `np.ndarray` objects. return_attention_mask (`bool`, *optional*): Whether or not to return the attention mask. If not set, will return the attention mask according to the specific tokenizer\'s default, defined by the `return_outputs` attribute. [What are attention masks?](../glossary#attention-mask) Return: `Dict[str, List[List[int]]]`: A dictionary with the following keys: - `input_ids`: List of token ids to be fed to a model. - `attention_mask`: List of indices specifying which tokens should be attended to by the model. ''' @add_start_docstrings(lowerCamelCase__ ) class lowercase_ : """simple docstring""" def __call__( self : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[str] = None ,lowercase__ : Optional[str] = None ,lowercase__ : Union[bool, str] = False ,lowercase__ : Union[bool, str] = False ,lowercase__ : Optional[int] = None ,lowercase__ : Optional[Union[str, TensorType]] = None ,lowercase__ : Optional[bool] = None ,**lowercase__ : str ,): if titles is None and texts is None: return super().__call__( lowercase__ ,padding=lowercase__ ,truncation=lowercase__ ,max_length=lowercase__ ,return_tensors=lowercase__ ,return_attention_mask=lowercase__ ,**lowercase__ ,) elif titles is None or texts is None: __lowercase = titles if texts is None else texts return super().__call__( lowercase__ ,lowercase__ ,padding=lowercase__ ,truncation=lowercase__ ,max_length=lowercase__ ,return_tensors=lowercase__ ,return_attention_mask=lowercase__ ,**lowercase__ ,) __lowercase = titles if not isinstance(lowercase__ ,lowercase__ ) else [titles] __lowercase = texts if not isinstance(lowercase__ ,lowercase__ ) else [texts] __lowercase = len(lowercase__ ) __lowercase = questions if not isinstance(lowercase__ ,lowercase__ ) else [questions] * n_passages assert len(lowercase__ ) == len( lowercase__ ), F"There should be as many titles than texts but got {len(lowercase__ )} titles and {len(lowercase__ )} texts." __lowercase = super().__call__(lowercase__ ,lowercase__ ,padding=lowercase__ ,truncation=lowercase__ )['''input_ids'''] __lowercase = super().__call__(lowercase__ ,add_special_tokens=lowercase__ ,padding=lowercase__ ,truncation=lowercase__ )['''input_ids'''] __lowercase = { '''input_ids''': [ (encoded_question_and_title + encoded_text)[:max_length] if max_length is not None and truncation else encoded_question_and_title + encoded_text for encoded_question_and_title, encoded_text in zip(lowercase__ ,lowercase__ ) ] } if return_attention_mask is not False: __lowercase = [] for input_ids in encoded_inputs["input_ids"]: attention_mask.append([int(input_id != self.pad_token_id ) for input_id in input_ids] ) __lowercase = attention_mask return self.pad(lowercase__ ,padding=lowercase__ ,max_length=lowercase__ ,return_tensors=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : BatchEncoding ,lowercase__ : DPRReaderOutput ,lowercase__ : int = 1_6 ,lowercase__ : int = 6_4 ,lowercase__ : int = 4 ,): __lowercase = reader_input['''input_ids'''] __lowercase , __lowercase , __lowercase = reader_output[:3] __lowercase = len(lowercase__ ) __lowercase = sorted(range(lowercase__ ) ,reverse=lowercase__ ,key=relevance_logits.__getitem__ ) __lowercase = [] for doc_id in sorted_docs: __lowercase = list(input_ids[doc_id] ) # assuming question & title information is at the beginning of the sequence __lowercase = sequence_ids.index(self.sep_token_id ,2 ) + 1 # second sep id if sequence_ids[-1] == self.pad_token_id: __lowercase = sequence_ids.index(self.pad_token_id ) else: __lowercase = len(lowercase__ ) __lowercase = self._get_best_spans( start_logits=start_logits[doc_id][passage_offset:sequence_len] ,end_logits=end_logits[doc_id][passage_offset:sequence_len] ,max_answer_length=lowercase__ ,top_spans=lowercase__ ,) for start_index, end_index in best_spans: start_index += passage_offset end_index += passage_offset nbest_spans_predictions.append( DPRSpanPrediction( span_score=start_logits[doc_id][start_index] + end_logits[doc_id][end_index] ,relevance_score=relevance_logits[doc_id] ,doc_id=lowercase__ ,start_index=lowercase__ ,end_index=lowercase__ ,text=self.decode(sequence_ids[start_index : end_index + 1] ) ,) ) if len(lowercase__ ) >= num_spans: break return nbest_spans_predictions[:num_spans] def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : List[int] ,lowercase__ : List[int] ,lowercase__ : int ,lowercase__ : int ,): __lowercase = [] for start_index, start_score in enumerate(lowercase__ ): for answer_length, end_score in enumerate(end_logits[start_index : start_index + max_answer_length] ): scores.append(((start_index, start_index + answer_length), start_score + end_score) ) __lowercase = sorted(lowercase__ ,key=lambda lowercase__ : x[1] ,reverse=lowercase__ ) __lowercase = [] for (start_index, end_index), score in scores: assert start_index <= end_index, F"Wrong span indices: [{start_index}:{end_index}]" __lowercase = end_index - start_index + 1 assert length <= max_answer_length, F"Span is too long: {length} > {max_answer_length}" if any( start_index <= prev_start_index <= prev_end_index <= end_index or prev_start_index <= start_index <= end_index <= prev_end_index for (prev_start_index, prev_end_index) in chosen_span_intervals ): continue chosen_span_intervals.append((start_index, end_index) ) if len(lowercase__ ) == top_spans: break return chosen_span_intervals @add_end_docstrings(lowerCamelCase__ ) class lowercase_ (lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[Any] = READER_PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = READER_PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES SCREAMING_SNAKE_CASE : Any = READER_PRETRAINED_INIT_CONFIGURATION SCREAMING_SNAKE_CASE : Tuple = ['input_ids', 'attention_mask'] SCREAMING_SNAKE_CASE : Dict = DPRReaderTokenizer
704
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
624
0
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging lowerCAmelCase__ = logging.get_logger(__name__) if is_vision_available(): import PIL class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[int] = ['pixel_values'] def __init__( self : List[str] ,lowercase__ : bool = True ,lowercase__ : Dict[str, int] = None ,lowercase__ : PILImageResampling = PILImageResampling.BICUBIC ,lowercase__ : bool = True ,lowercase__ : Dict[str, int] = None ,lowercase__ : bool = True ,lowercase__ : Union[int, float] = 1 / 2_5_5 ,lowercase__ : bool = True ,lowercase__ : Optional[Union[float, List[float]]] = None ,lowercase__ : Optional[Union[float, List[float]]] = None ,lowercase__ : bool = True ,**lowercase__ : str ,): super().__init__(**lowercase__ ) __lowercase = size if size is not None else {'''shortest_edge''': 2_2_4} __lowercase = get_size_dict(lowercase__ ,default_to_square=lowercase__ ) __lowercase = crop_size if crop_size is not None else {'''height''': 2_2_4, '''width''': 2_2_4} __lowercase = get_size_dict(lowercase__ ,default_to_square=lowercase__ ,param_name='''crop_size''' ) __lowercase = do_resize __lowercase = size __lowercase = resample __lowercase = do_center_crop __lowercase = crop_size __lowercase = do_rescale __lowercase = rescale_factor __lowercase = do_normalize __lowercase = image_mean if image_mean is not None else OPENAI_CLIP_MEAN __lowercase = image_std if image_std is not None else OPENAI_CLIP_STD __lowercase = do_convert_rgb def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : np.ndarray ,lowercase__ : Dict[str, int] ,lowercase__ : PILImageResampling = PILImageResampling.BICUBIC ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,**lowercase__ : Union[str, Any] ,): __lowercase = get_size_dict(lowercase__ ,default_to_square=lowercase__ ) if "shortest_edge" not in size: raise ValueError(F"The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}" ) __lowercase = get_resize_output_image_size(lowercase__ ,size=size['''shortest_edge'''] ,default_to_square=lowercase__ ) return resize(lowercase__ ,size=lowercase__ ,resample=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : np.ndarray ,lowercase__ : Dict[str, int] ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,**lowercase__ : List[str] ,): __lowercase = get_size_dict(lowercase__ ) if "height" not in size or "width" not in size: raise ValueError(F"The `size` parameter must contain the keys (height, width). Got {size.keys()}" ) return center_crop(lowercase__ ,size=(size['''height'''], size['''width''']) ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : np.ndarray ,lowercase__ : Union[int, float] ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,**lowercase__ : int ,): return rescale(lowercase__ ,scale=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ,lowercase__ : np.ndarray ,lowercase__ : Union[float, List[float]] ,lowercase__ : Union[float, List[float]] ,lowercase__ : Optional[Union[str, ChannelDimension]] = None ,**lowercase__ : str ,): return normalize(lowercase__ ,mean=lowercase__ ,std=lowercase__ ,data_format=lowercase__ ,**lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : ImageInput ,lowercase__ : bool = None ,lowercase__ : Dict[str, int] = None ,lowercase__ : PILImageResampling = None ,lowercase__ : bool = None ,lowercase__ : int = None ,lowercase__ : bool = None ,lowercase__ : float = None ,lowercase__ : bool = None ,lowercase__ : Optional[Union[float, List[float]]] = None ,lowercase__ : Optional[Union[float, List[float]]] = None ,lowercase__ : bool = None ,lowercase__ : Optional[Union[str, TensorType]] = None ,lowercase__ : Optional[ChannelDimension] = ChannelDimension.FIRST ,**lowercase__ : str ,): __lowercase = do_resize if do_resize is not None else self.do_resize __lowercase = size if size is not None else self.size __lowercase = get_size_dict(lowercase__ ,param_name='''size''' ,default_to_square=lowercase__ ) __lowercase = resample if resample is not None else self.resample __lowercase = do_center_crop if do_center_crop is not None else self.do_center_crop __lowercase = crop_size if crop_size is not None else self.crop_size __lowercase = get_size_dict(lowercase__ ,param_name='''crop_size''' ,default_to_square=lowercase__ ) __lowercase = do_rescale if do_rescale is not None else self.do_rescale __lowercase = rescale_factor if rescale_factor is not None else self.rescale_factor __lowercase = do_normalize if do_normalize is not None else self.do_normalize __lowercase = image_mean if image_mean is not None else self.image_mean __lowercase = image_std if image_std is not None else self.image_std __lowercase = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb __lowercase = make_list_of_images(lowercase__ ) if not valid_images(lowercase__ ): raise ValueError( '''Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ''' '''torch.Tensor, tf.Tensor or jax.ndarray.''' ) if do_resize and size is None: raise ValueError('''Size must be specified if do_resize is True.''' ) if do_center_crop and crop_size is None: raise ValueError('''Crop size must be specified if do_center_crop is True.''' ) if do_rescale and rescale_factor is None: raise ValueError('''Rescale factor must be specified if do_rescale is True.''' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('''Image mean and std must be specified if do_normalize is True.''' ) # PIL RGBA images are converted to RGB if do_convert_rgb: __lowercase = [convert_to_rgb(lowercase__ ) for image in images] # All transformations expect numpy arrays. __lowercase = [to_numpy_array(lowercase__ ) for image in images] if do_resize: __lowercase = [self.resize(image=lowercase__ ,size=lowercase__ ,resample=lowercase__ ) for image in images] if do_center_crop: __lowercase = [self.center_crop(image=lowercase__ ,size=lowercase__ ) for image in images] if do_rescale: __lowercase = [self.rescale(image=lowercase__ ,scale=lowercase__ ) for image in images] if do_normalize: __lowercase = [self.normalize(image=lowercase__ ,mean=lowercase__ ,std=lowercase__ ) for image in images] __lowercase = [to_channel_dimension_format(lowercase__ ,lowercase__ ) for image in images] __lowercase = {'''pixel_values''': images} return BatchFeature(data=lowercase__ ,tensor_type=lowercase__ )
705
'''simple docstring''' import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = TaConfig.from_json_file(A__ ) print(F"Building PyTorch model from configuration: {config}" ) __lowercase = TaForConditionalGeneration(A__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(A__ , A__ , A__ ) # Save pytorch-model print(F"Save PyTorch model to {pytorch_dump_path}" ) model.save_pretrained(A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
624
0
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
706
'''simple docstring''' from argparse import ArgumentParser from . import BaseTransformersCLICommand def _A ( A__ ): """simple docstring""" return DownloadCommand(args.model , args.cache_dir , args.force , args.trust_remote_code ) class lowercase_ (lowerCamelCase__ ): """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( lowercase__ : ArgumentParser ): __lowercase = parser.add_parser('''download''' ) download_parser.add_argument( '''--cache-dir''' ,type=lowercase__ ,default=lowercase__ ,help='''Path to location to store the models''' ) download_parser.add_argument( '''--force''' ,action='''store_true''' ,help='''Force the model to be download even if already in cache-dir''' ) download_parser.add_argument( '''--trust-remote-code''' ,action='''store_true''' ,help='''Whether or not to allow for custom models defined on the Hub in their own modeling files. Use only if you\'ve reviewed the code as it will execute on your local machine''' ,) download_parser.add_argument('''model''' ,type=lowercase__ ,help='''Name of the model to download''' ) download_parser.set_defaults(func=lowercase__ ) def __init__( self : str ,lowercase__ : str ,lowercase__ : str ,lowercase__ : bool ,lowercase__ : bool ): __lowercase = model __lowercase = cache __lowercase = force __lowercase = trust_remote_code def SCREAMING_SNAKE_CASE ( self : Any ): from ..models.auto import AutoModel, AutoTokenizer AutoModel.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code ) AutoTokenizer.from_pretrained( self._model ,cache_dir=self._cache ,force_download=self._force ,trust_remote_code=self._trust_remote_code )
624
0
'''simple docstring''' import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse('''0.8.3'''): raise Exception('''requires gluonnlp == 0.8.3''') if version.parse(mx.__version__) != version.parse('''1.5.0'''): raise Exception('''requires mxnet == 1.5.0''') logging.set_verbosity_info() lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = '''The Nymphenburg Palace is a beautiful palace in Munich!''' def _A ( A__ , A__ ): """simple docstring""" __lowercase = { '''attention_cell''': '''multi_head''', '''num_layers''': 4, '''units''': 1024, '''hidden_size''': 768, '''max_length''': 512, '''num_heads''': 8, '''scaled''': True, '''dropout''': 0.1, '''use_residual''': True, '''embed_size''': 1024, '''embed_dropout''': 0.1, '''word_embed''': None, '''layer_norm_eps''': 1e-5, '''token_type_vocab_size''': 2, } __lowercase = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py __lowercase = BERTEncoder( attention_cell=predefined_args['''attention_cell'''] , num_layers=predefined_args['''num_layers'''] , units=predefined_args['''units'''] , hidden_size=predefined_args['''hidden_size'''] , max_length=predefined_args['''max_length'''] , num_heads=predefined_args['''num_heads'''] , scaled=predefined_args['''scaled'''] , dropout=predefined_args['''dropout'''] , output_attention=A__ , output_all_encodings=A__ , use_residual=predefined_args['''use_residual'''] , activation=predefined_args.get('''activation''' , '''gelu''' ) , layer_norm_eps=predefined_args.get('''layer_norm_eps''' , A__ ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later __lowercase = '''openwebtext_ccnews_stories_books_cased''' # Specify download folder to Gluonnlp's vocab __lowercase = os.path.join(get_home_dir() , '''models''' ) __lowercase = _load_vocab(A__ , A__ , A__ , cls=A__ ) __lowercase = nlp.model.BERTModel( A__ , len(A__ ) , units=predefined_args['''units'''] , embed_size=predefined_args['''embed_size'''] , embed_dropout=predefined_args['''embed_dropout'''] , word_embed=predefined_args['''word_embed'''] , use_pooler=A__ , use_token_type_embed=A__ , token_type_vocab_size=predefined_args['''token_type_vocab_size'''] , use_classifier=A__ , use_decoder=A__ , ) original_bort.load_parameters(A__ , cast_dtype=A__ , ignore_extra=A__ ) __lowercase = original_bort._collect_params_with_prefix() # Build our config 🤗 __lowercase = { '''architectures''': ['''BertForMaskedLM'''], '''attention_probs_dropout_prob''': predefined_args['''dropout'''], '''hidden_act''': '''gelu''', '''hidden_dropout_prob''': predefined_args['''dropout'''], '''hidden_size''': predefined_args['''embed_size'''], '''initializer_range''': 0.0_2, '''intermediate_size''': predefined_args['''hidden_size'''], '''layer_norm_eps''': predefined_args['''layer_norm_eps'''], '''max_position_embeddings''': predefined_args['''max_length'''], '''model_type''': '''bort''', '''num_attention_heads''': predefined_args['''num_heads'''], '''num_hidden_layers''': predefined_args['''num_layers'''], '''pad_token_id''': 1, # 2 = BERT, 1 = RoBERTa '''type_vocab_size''': 1, # 2 = BERT, 1 = RoBERTa '''vocab_size''': len(A__ ), } __lowercase = BertConfig.from_dict(A__ ) __lowercase = BertForMaskedLM(A__ ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(A__ ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(A__ , A__ ): __lowercase = hf_param.shape __lowercase = to_torch(params[gluon_param] ) __lowercase = gluon_param.shape assert ( shape_hf == shape_gluon ), F"The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers" return gluon_param __lowercase = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , '''word_embed.0.weight''' ) __lowercase = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , '''encoder.position_weight''' ) __lowercase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , '''encoder.layer_norm.beta''' ) __lowercase = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , '''encoder.layer_norm.gamma''' ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) __lowercase = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): __lowercase = hf_bort_model.bert.encoder.layer[i] # self attention __lowercase = layer.attention.self __lowercase = check_and_map_params( self_attn.key.bias.data , F"encoder.transformer_cells.{i}.attention_cell.proj_key.bias" ) __lowercase = check_and_map_params( self_attn.key.weight.data , F"encoder.transformer_cells.{i}.attention_cell.proj_key.weight" ) __lowercase = check_and_map_params( self_attn.query.bias.data , F"encoder.transformer_cells.{i}.attention_cell.proj_query.bias" ) __lowercase = check_and_map_params( self_attn.query.weight.data , F"encoder.transformer_cells.{i}.attention_cell.proj_query.weight" ) __lowercase = check_and_map_params( self_attn.value.bias.data , F"encoder.transformer_cells.{i}.attention_cell.proj_value.bias" ) __lowercase = check_and_map_params( self_attn.value.weight.data , F"encoder.transformer_cells.{i}.attention_cell.proj_value.weight" ) # self attention output __lowercase = layer.attention.output __lowercase = check_and_map_params( self_output.dense.bias , F"encoder.transformer_cells.{i}.proj.bias" ) __lowercase = check_and_map_params( self_output.dense.weight , F"encoder.transformer_cells.{i}.proj.weight" ) __lowercase = check_and_map_params( self_output.LayerNorm.bias , F"encoder.transformer_cells.{i}.layer_norm.beta" ) __lowercase = check_and_map_params( self_output.LayerNorm.weight , F"encoder.transformer_cells.{i}.layer_norm.gamma" ) # intermediate __lowercase = layer.intermediate __lowercase = check_and_map_params( intermediate.dense.bias , F"encoder.transformer_cells.{i}.ffn.ffn_1.bias" ) __lowercase = check_and_map_params( intermediate.dense.weight , F"encoder.transformer_cells.{i}.ffn.ffn_1.weight" ) # output __lowercase = layer.output __lowercase = check_and_map_params( bert_output.dense.bias , F"encoder.transformer_cells.{i}.ffn.ffn_2.bias" ) __lowercase = check_and_map_params( bert_output.dense.weight , F"encoder.transformer_cells.{i}.ffn.ffn_2.weight" ) __lowercase = check_and_map_params( bert_output.LayerNorm.bias , F"encoder.transformer_cells.{i}.ffn.layer_norm.beta" ) __lowercase = check_and_map_params( bert_output.LayerNorm.weight , F"encoder.transformer_cells.{i}.ffn.layer_norm.gamma" ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models __lowercase = RobertaTokenizer.from_pretrained('''roberta-base''' ) __lowercase = tokenizer.encode_plus(A__ )['''input_ids'''] # Get gluon output __lowercase = mx.nd.array([input_ids] ) __lowercase = original_bort(inputs=A__ , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(A__ ) __lowercase = BertModel.from_pretrained(A__ ) hf_bort_model.eval() __lowercase = tokenizer.encode_plus(A__ , return_tensors='''pt''' ) __lowercase = hf_bort_model(**A__ )[0] __lowercase = output_gluon[0].asnumpy() __lowercase = output_hf[0].detach().numpy() __lowercase = np.max(np.abs(hf_layer - gluon_layer ) ).item() __lowercase = np.allclose(A__ , A__ , atol=1e-3 ) if success: print('''✔️ Both model do output the same tensors''' ) else: print('''❌ Both model do **NOT** output the same tensors''' ) print('''Absolute difference is:''' , A__ ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--bort_checkpoint_path''', default=None, type=str, required=True, help='''Path the official Bort params file.''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
707
'''simple docstring''' import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFGPTaLMHeadModel, is_keras_nlp_available, is_tf_available from transformers.models.gpta.tokenization_gpta import GPTaTokenizer from transformers.testing_utils import require_keras_nlp, require_tf, slow if is_tf_available(): import tensorflow as tf if is_keras_nlp_available(): from transformers.models.gpta import TFGPTaTokenizer lowerCAmelCase__ = ['''gpt2'''] lowerCAmelCase__ = '''gpt2''' if is_tf_available(): class lowercase_ (tf.Module ): """simple docstring""" def __init__( self : List[str] ,lowercase__ : Tuple ): super().__init__() __lowercase = tokenizer __lowercase = AutoConfig.from_pretrained(lowercase__ ) __lowercase = TFGPTaLMHeadModel.from_config(lowercase__ ) @tf.function(input_signature=(tf.TensorSpec((None,) ,tf.string ,name='''text''' ),) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Dict ): __lowercase = self.tokenizer(lowercase__ ) __lowercase = tokenized['''input_ids'''].to_tensor() __lowercase = tf.cast(input_ids_dense > 0 ,tf.intaa ) # input_mask = tf.reshape(input_mask, [-1, MAX_SEQ_LEN]) __lowercase = self.model(input_ids=lowercase__ ,attention_mask=lowercase__ )['''logits'''] return outputs @require_tf @require_keras_nlp class lowercase_ (unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE ( self : List[Any] ): super().setUp() __lowercase = [GPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in (TOKENIZER_CHECKPOINTS)] __lowercase = [TFGPTaTokenizer.from_pretrained(lowercase__ ) for checkpoint in TOKENIZER_CHECKPOINTS] assert len(self.tokenizers ) == len(self.tf_tokenizers ) __lowercase = [ '''This is a straightforward English test sentence.''', '''This one has some weird characters\rto\nsee\r\nif those\u00E9break things.''', '''Now we\'re going to add some Chinese: 一 二 三 一二三''', '''And some much more rare Chinese: 齉 堃 齉堃''', '''Je vais aussi écrire en français pour tester les accents''', '''Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ''', ] __lowercase = list(zip(self.test_sentences ,self.test_sentences[::-1] ) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tokenizer, tf_tokenizer in zip(self.tokenizers ,self.tf_tokenizers ): for test_inputs in self.test_sentences: __lowercase = tokenizer([test_inputs] ,return_tensors='''tf''' ) __lowercase = tf_tokenizer([test_inputs] ) for key in python_outputs.keys(): # convert them to numpy to avoid messing with ragged tensors __lowercase = python_outputs[key].numpy() __lowercase = tf_outputs[key].numpy() self.assertTrue(tf.reduce_all(python_outputs_values.shape == tf_outputs_values.shape ) ) self.assertTrue(tf.reduce_all(tf.cast(lowercase__ ,tf.intaa ) == tf_outputs_values ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.function(lowercase__ ) for test_inputs in self.test_sentences: __lowercase = tf.constant(lowercase__ ) __lowercase = compiled_tokenizer(lowercase__ ) __lowercase = tf_tokenizer(lowercase__ ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : str ): for tf_tokenizer in self.tf_tokenizers: __lowercase = ModelToSave(tokenizer=lowercase__ ) __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = model.serving(lowercase__ ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: __lowercase = Path(lowercase__ ) / '''saved.model''' tf.saved_model.save(lowercase__ ,lowercase__ ,signatures={'''serving_default''': model.serving} ) __lowercase = tf.saved_model.load(lowercase__ ) __lowercase = loaded_model.signatures['''serving_default'''](lowercase__ )['''output_0'''] # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertTrue(tf.reduce_all(out == loaded_output ) ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ): for tf_tokenizer in self.tf_tokenizers: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ) # Build model with some sample inputs __lowercase = tf_tokenizer.get_config() __lowercase = TFGPTaTokenizer.from_config(lowercase__ ) __lowercase = model_from_config(lowercase__ ) for key in from_config_output.keys(): self.assertTrue(tf.reduce_all(from_config_output[key] == out[key] ) ) @slow def SCREAMING_SNAKE_CASE ( self : Optional[int] ): for tf_tokenizer in self.tf_tokenizers: # for the test to run __lowercase = 1_2_3_1_2_3 for max_length in [3, 5, 1_0_2_4]: __lowercase = tf.convert_to_tensor([self.test_sentences[0]] ) __lowercase = tf_tokenizer(lowercase__ ,max_length=lowercase__ ) __lowercase = out['''input_ids'''].numpy().shape[1] assert out_length == max_length
624
0
import argparse import datetime import io import itertools import json import math import os import platform import re import shlex import subprocess import sys from pathlib import Path from statistics import fmean import pandas as pd import torch from tqdm import tqdm import transformers lowerCAmelCase__ = float('''nan''') class lowercase_ : """simple docstring""" def __init__( self : int ,lowercase__ : List[Any] ): __lowercase = sys.stdout __lowercase = open(lowercase__ ,'''a''' ) def __getattr__( self : str ,lowercase__ : List[Any] ): return getattr(self.stdout ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[Any] ): self.stdout.write(lowercase__ ) # strip tqdm codes self.file.write(re.sub(r'''^.*\r''' ,'''''' ,lowercase__ ,0 ,re.M ) ) def _A ( A__=80 , A__=False ): """simple docstring""" __lowercase = [] # deal with critical env vars __lowercase = ['''CUDA_VISIBLE_DEVICES'''] for key in env_keys: __lowercase = os.environ.get(A__ , A__ ) if val is not None: cmd.append(F"{key}={val}" ) # python executable (not always needed if the script is executable) __lowercase = sys.executable if full_python_path else sys.executable.split('''/''' )[-1] cmd.append(A__ ) # now the normal args cmd += list(map(shlex.quote , sys.argv ) ) # split up into up to MAX_WIDTH lines with shell multi-line escapes __lowercase = [] __lowercase = '''''' while len(A__ ) > 0: current_line += F"{cmd.pop(0 )} " if len(A__ ) == 0 or len(A__ ) + len(cmd[0] ) + 1 > max_width - 1: lines.append(A__ ) __lowercase = '''''' return "\\\n".join(A__ ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = re.sub(R'''[\\\n]+''' , ''' ''' , args.base_cmd ) # remove --output_dir if any and set our own __lowercase = re.sub('''--output_dir\s+[^\s]+''' , '''''' , args.base_cmd ) args.base_cmd += F" --output_dir {output_dir}" # ensure we have --overwrite_output_dir __lowercase = re.sub('''--overwrite_output_dir\s+''' , '''''' , args.base_cmd ) args.base_cmd += " --overwrite_output_dir" return [sys.executable] + shlex.split(args.base_cmd ) def _A ( A__ , A__ , A__ , A__ , A__ , A__ , A__ ): """simple docstring""" if 0: import random from time import sleep sleep(0 ) return dict( {k: random.uniform(0 , 100 ) for k in metric_keys} , **{target_metric_key: random.choice([nan, 10.31, 100.2, 55.6666, 222.22222222] )} , ) __lowercase = subprocess.run(A__ , capture_output=A__ , text=A__ ) if verbose: print('''STDOUT''' , result.stdout ) print('''STDERR''' , result.stderr ) # save the streams __lowercase = variation.replace(''' ''' , '''-''' ) with open(Path(A__ ) / F"log.{prefix}.stdout.txt" , '''w''' ) as f: f.write(result.stdout ) with open(Path(A__ ) / F"log.{prefix}.stderr.txt" , '''w''' ) as f: f.write(result.stderr ) if result.returncode != 0: if verbose: print('''failed''' ) return {target_metric_key: nan} with io.open(F"{output_dir}/all_results.json" , '''r''' , encoding='''utf-8''' ) as f: __lowercase = json.load(A__ ) # filter out just the keys we want return {k: v for k, v in metrics.items() if k in metric_keys} def _A ( A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , A__ , ): """simple docstring""" __lowercase = [] __lowercase = [] __lowercase = F"{id}: {variation:<{longest_variation_len}}" __lowercase = F"{preamble}: " __lowercase = set(report_metric_keys + [target_metric_key] ) for i in tqdm(range(A__ ) , desc=A__ , leave=A__ ): __lowercase = process_run_single( A__ , A__ , A__ , A__ , A__ , A__ , A__ ) __lowercase = single_run_metrics[target_metric_key] if not math.isnan(A__ ): metrics.append(A__ ) results.append(A__ ) outcome += "✓" else: outcome += "✘" __lowercase = F"\33[2K\r{outcome}" if len(A__ ) > 0: __lowercase = {k: fmean([x[k] for x in metrics] ) for k in metrics[0].keys()} __lowercase = round(mean_metrics[target_metric_key] , 2 ) __lowercase = F"{outcome} {mean_target}" if len(A__ ) > 1: results_str += F" {tuple(round(A__ , 2 ) for x in results )}" print(A__ ) __lowercase = variation return mean_metrics else: print(A__ ) return {variation_key: variation, target_metric_key: nan} def _A ( ): """simple docstring""" __lowercase = torch.cuda.get_device_properties(torch.device('''cuda''' ) ) return F"\nDatetime : {datetime.datetime.now().strftime('%Y-%m-%d %H:%M:%S' )}\n\nSoftware:\ntransformers: {transformers.__version__}\ntorch : {torch.__version__}\ncuda : {torch.version.cuda}\npython : {platform.python_version()}\n\nHardware:\n{torch.cuda.device_count()} GPUs : {properties.name}, {properties.total_memory/2**30:0.2f}GB\n" def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = pd.DataFrame(A__ ) __lowercase = '''variation''' __lowercase = '''diff_%''' __lowercase = nan if base_variation is not None and len(df[df[variation_key] == base_variation] ): # this may still return nan __lowercase = df.loc[df[variation_key] == base_variation][target_metric_key].item() if math.isnan(A__ ): # as a fallback, use the minimal value as the sentinel __lowercase = df.loc[df[target_metric_key] != nan][target_metric_key].min() # create diff column if possible if not math.isnan(A__ ): __lowercase = df.apply( lambda A__ : round(100 * (r[target_metric_key] - sentinel_value) / sentinel_value ) if not math.isnan(r[target_metric_key] ) else 0 , axis='''columns''' , ) # re-order columns __lowercase = [variation_key, target_metric_key, diff_key, *report_metric_keys] __lowercase = df.reindex(A__ , axis='''columns''' ) # reorder cols # capitalize __lowercase = df.rename(str.capitalize , axis='''columns''' ) # make the cols as narrow as possible __lowercase = df.rename(lambda A__ : c.replace('''_''' , '''<br>''' ) , axis='''columns''' ) __lowercase = df.rename(lambda A__ : c.replace('''_''' , '''\n''' ) , axis='''columns''' ) __lowercase = ['''''', '''Copy between the cut-here-lines and paste as is to github or a forum'''] report += ["----------8<-----------------8<--------"] report += ["*** Results:", df_github.to_markdown(index=A__ , floatfmt='''.2f''' )] report += ["```"] report += ["*** Setup:", get_versions()] report += ["*** The benchmark command line was:", get_original_command()] report += ["```"] report += ["----------8<-----------------8<--------"] report += ["*** Results (console):", df_console.to_markdown(index=A__ , floatfmt='''.2f''' )] print('''\n\n'''.join(A__ ) ) def _A ( ): """simple docstring""" __lowercase = argparse.ArgumentParser() parser.add_argument( '''--base-cmd''' , default=A__ , type=A__ , required=A__ , help='''Base cmd''' , ) parser.add_argument( '''--variations''' , default=A__ , type=A__ , nargs='''+''' , required=A__ , help='''Multi-dimensional variations, example: \'|--fp16|--bf16\' \'|--tf32\'''' , ) parser.add_argument( '''--base-variation''' , default=A__ , type=A__ , help='''Baseline variation to compare to. if None the minimal target value will be used to compare against''' , ) parser.add_argument( '''--target-metric-key''' , default=A__ , type=A__ , required=A__ , help='''Target metric key in output_dir/all_results.json, e.g., train_samples_per_second''' , ) parser.add_argument( '''--report-metric-keys''' , default='''''' , type=A__ , help='''Report metric keys - other metric keys from output_dir/all_results.json to report, e.g., train_loss. Use a single argument e.g., \'train_loss train_samples''' , ) parser.add_argument( '''--repeat-times''' , default=1 , type=A__ , help='''How many times to re-run each variation - an average will be reported''' , ) parser.add_argument( '''--output_dir''' , default='''output_benchmark''' , type=A__ , help='''The output directory where all the benchmark reports will go to and additionally this directory will be used to override --output_dir in the script that is being benchmarked''' , ) parser.add_argument( '''--verbose''' , default=A__ , action='''store_true''' , help='''Whether to show the outputs of each run or just the benchmark progress''' , ) __lowercase = parser.parse_args() __lowercase = args.output_dir Path(A__ ).mkdir(exist_ok=A__ ) __lowercase = get_base_command(A__ , A__ ) # split each dimension into its --foo variations __lowercase = [list(map(str.strip , re.split(R'''\|''' , A__ ) ) ) for x in args.variations] # build a cartesian product of dimensions and convert those back into cmd-line arg strings, # while stripping white space for inputs that were empty __lowercase = list(map(str.strip , map(''' '''.join , itertools.product(*A__ ) ) ) ) __lowercase = max(len(A__ ) for x in variations ) # split wanted keys __lowercase = args.report_metric_keys.split() # capture prints into a log file for convenience __lowercase = F"benchmark-report-{datetime.datetime.now().strftime('%Y-%m-%d-%H-%M-%S' )}.txt" print(F"\nNote: each run's output is also logged under {output_dir}/log.*.std*.txt" ) print(F"and this script's output is also piped into {report_fn}" ) __lowercase = Tee(A__ ) print(F"\n*** Running {len(A__ )} benchmarks:" ) print(F"Base command: {' '.join(A__ )}" ) __lowercase = '''variation''' __lowercase = [] for id, variation in enumerate(tqdm(A__ , desc='''Total completion: ''' , leave=A__ ) ): __lowercase = base_cmd + variation.split() results.append( process_run( id + 1 , A__ , A__ , A__ , A__ , args.target_metric_key , A__ , args.repeat_times , A__ , args.verbose , ) ) process_results(A__ , args.target_metric_key , A__ , args.base_variation , A__ ) if __name__ == "__main__": main()
708
'''simple docstring''' from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCAmelCase__ = numpy.array([0, 0]) lowerCAmelCase__ = numpy.array([0.5, 0.8_660_254]) lowerCAmelCase__ = numpy.array([1, 0]) lowerCAmelCase__ = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def _A ( A__ , A__ ): """simple docstring""" __lowercase = initial_vectors for _ in range(A__ ): __lowercase = iteration_step(A__ ) return vectors def _A ( A__ ): """simple docstring""" __lowercase = [] for i, start_vector in enumerate(vectors[:-1] ): __lowercase = vectors[i + 1] new_vectors.append(A__ ) __lowercase = end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def _A ( A__ , A__ ): """simple docstring""" __lowercase = numpy.radians(A__ ) __lowercase , __lowercase = numpy.cos(A__ ), numpy.sin(A__ ) __lowercase = numpy.array(((c, -s), (s, c)) ) return numpy.dot(A__ , A__ ) def _A ( A__ ): """simple docstring""" __lowercase = plt.gca() axes.set_aspect('''equal''' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() __lowercase , __lowercase = zip(*A__ ) plt.plot(A__ , A__ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCAmelCase__ = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
624
0
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
709
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
624
0
from math import loga def _A ( A__ ): """simple docstring""" if a < 0: raise ValueError('''Input value must be a positive integer''' ) elif isinstance(A__ , A__ ): raise TypeError('''Input value must be a \'int\' type''' ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
710
'''simple docstring''' import glob import os import random from string import ascii_lowercase, digits import cva import numpy as np # Parrameters lowerCAmelCase__ = (720, 1280) # Height, Width lowerCAmelCase__ = (0.4, 0.6) # if height or width lower than this scale, drop it. lowerCAmelCase__ = 1 / 100 lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = '''''' lowerCAmelCase__ = 250 def _A ( ): """simple docstring""" __lowercase , __lowercase = get_dataset(A__ , A__ ) for index in range(A__ ): __lowercase = random.sample(range(len(A__ ) ) , 4 ) __lowercase , __lowercase , __lowercase = update_image_and_anno( A__ , A__ , A__ , A__ , A__ , filter_scale=A__ , ) # Get random string code: '7b7ad245cdff75241935e4dd860f3bad' __lowercase = random_chars(32 ) __lowercase = path.split(os.sep )[-1].rsplit('''.''' , 1 )[0] __lowercase = F"{OUTPUT_DIR}/{file_name}_MOSAIC_{letter_code}" cva.imwrite(F"{file_root}.jpg" , A__ , [cva.IMWRITE_JPEG_QUALITY, 85] ) print(F"Succeeded {index+1}/{NUMBER_IMAGES} with {file_name}" ) __lowercase = [] for anno in new_annos: __lowercase = anno[3] - anno[1] __lowercase = anno[4] - anno[2] __lowercase = anno[1] + width / 2 __lowercase = anno[2] + height / 2 __lowercase = F"{anno[0]} {x_center} {y_center} {width} {height}" annos_list.append(A__ ) with open(F"{file_root}.txt" , '''w''' ) as outfile: outfile.write('''\n'''.join(line for line in annos_list ) ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = [] __lowercase = [] for label_file in glob.glob(os.path.join(A__ , '''*.txt''' ) ): __lowercase = label_file.split(os.sep )[-1].rsplit('''.''' , 1 )[0] with open(A__ ) as in_file: __lowercase = in_file.readlines() __lowercase = os.path.join(A__ , F"{label_name}.jpg" ) __lowercase = [] for obj_list in obj_lists: __lowercase = obj_list.rstrip('''\n''' ).split(''' ''' ) __lowercase = float(obj[1] ) - float(obj[3] ) / 2 __lowercase = float(obj[2] ) - float(obj[4] ) / 2 __lowercase = float(obj[1] ) + float(obj[3] ) / 2 __lowercase = float(obj[2] ) + float(obj[4] ) / 2 boxes.append([int(obj[0] ), xmin, ymin, xmax, ymax] ) if not boxes: continue img_paths.append(A__ ) labels.append(A__ ) return img_paths, labels def _A ( A__ , A__ , A__ , A__ , A__ , A__ = 0.0 , ): """simple docstring""" __lowercase = np.zeros([output_size[0], output_size[1], 3] , dtype=np.uinta ) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = scale_range[0] + random.random() * (scale_range[1] - scale_range[0]) __lowercase = int(scale_x * output_size[1] ) __lowercase = int(scale_y * output_size[0] ) __lowercase = [] __lowercase = [] for i, index in enumerate(A__ ): __lowercase = all_img_list[index] path_list.append(A__ ) __lowercase = all_annos[index] __lowercase = cva.imread(A__ ) if i == 0: # top-left __lowercase = cva.resize(A__ , (divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = bbox[2] * scale_y __lowercase = bbox[3] * scale_x __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 1: # top-right __lowercase = cva.resize(A__ , (output_size[1] - divid_point_x, divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = bbox[2] * scale_y __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = bbox[4] * scale_y new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) elif i == 2: # bottom-left __lowercase = cva.resize(A__ , (divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = bbox[1] * scale_x __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = bbox[3] * scale_x __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) else: # bottom-right __lowercase = cva.resize( A__ , (output_size[1] - divid_point_x, output_size[0] - divid_point_y) ) __lowercase = img for bbox in img_annos: __lowercase = scale_x + bbox[1] * (1 - scale_x) __lowercase = scale_y + bbox[2] * (1 - scale_y) __lowercase = scale_x + bbox[3] * (1 - scale_x) __lowercase = scale_y + bbox[4] * (1 - scale_y) new_anno.append([bbox[0], xmin, ymin, xmax, ymax] ) # Remove bounding box small than scale of filter if filter_scale > 0: __lowercase = [ anno for anno in new_anno if filter_scale < (anno[3] - anno[1]) and filter_scale < (anno[4] - anno[2]) ] return output_img, new_anno, path_list[0] def _A ( A__ ): """simple docstring""" assert number_char > 1, "The number of character should greater than 1" __lowercase = ascii_lowercase + digits return "".join(random.choice(A__ ) for _ in range(A__ ) ) if __name__ == "__main__": main() print('''DONE ✅''')
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" if p < 2: raise ValueError('''p should not be less than 2!''' ) elif p == 2: return True __lowercase = 4 __lowercase = (1 << p) - 1 for _ in range(p - 2 ): __lowercase = ((s * s) - 2) % m return s == 0 if __name__ == "__main__": print(lucas_lehmer_test(7)) print(lucas_lehmer_test(11))
711
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging lowerCAmelCase__ = logging.get_logger(__name__) lowerCAmelCase__ = {'''vocab_file''': '''sentencepiece.model'''} lowerCAmelCase__ = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } lowerCAmelCase__ = { '''google/rembert''': 256, } class lowercase_ (lowerCamelCase__ ): """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = VOCAB_FILES_NAMES SCREAMING_SNAKE_CASE : Optional[int] = PRETRAINED_VOCAB_FILES_MAP SCREAMING_SNAKE_CASE : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : str ,lowercase__ : Optional[Any] ,lowercase__ : List[str]=False ,lowercase__ : Dict=True ,lowercase__ : List[str]=True ,lowercase__ : Dict="[CLS]" ,lowercase__ : Union[str, Any]="[SEP]" ,lowercase__ : List[str]="[UNK]" ,lowercase__ : int="[SEP]" ,lowercase__ : List[str]="[PAD]" ,lowercase__ : Optional[int]="[CLS]" ,lowercase__ : List[Any]="[MASK]" ,**lowercase__ : int ,): super().__init__( do_lower_case=lowercase__ ,remove_space=lowercase__ ,keep_accents=lowercase__ ,bos_token=lowercase__ ,eos_token=lowercase__ ,unk_token=lowercase__ ,sep_token=lowercase__ ,pad_token=lowercase__ ,cls_token=lowercase__ ,mask_token=lowercase__ ,**lowercase__ ,) __lowercase = do_lower_case __lowercase = remove_space __lowercase = keep_accents __lowercase = vocab_file __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(lowercase__ ) @property def SCREAMING_SNAKE_CASE ( self : str ): return len(self.sp_model ) def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = {self.convert_ids_to_tokens(lowercase__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : List[Any] ): __lowercase = self.__dict__.copy() __lowercase = None return state def __setstate__( self : str ,lowercase__ : Optional[int] ): __lowercase = d __lowercase = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def SCREAMING_SNAKE_CASE ( self : List[str] ,lowercase__ : List[str] ,lowercase__ : List[Any]=False ): __lowercase = self.sp_model.EncodeAsPieces(lowercase__ ) return pieces def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : List[Any] ): return self.sp_model.PieceToId(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : str ): return self.sp_model.IdToPiece(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Tuple ): __lowercase = self.sp_model.decode_pieces(lowercase__ ) return out_string def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ,lowercase__ : bool = False ): if already_has_special_tokens: if token_ids_a is not None: raise ValueError( '''You should not supply a second sequence if the provided sequence of ''' '''ids is already formatted with special tokens for the model.''' ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(lowercase__ )) + [1] + ([0] * len(lowercase__ )) + [1] return [1] + ([0] * len(lowercase__ )) + [1] def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[int] ,lowercase__ : Optional[List[int]] = None ): __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : str ,lowercase__ : Optional[str] = None ): if not os.path.isdir(lowercase__ ): logger.error('''Vocabulary path ({}) should be a directory'''.format(lowercase__ ) ) return __lowercase = os.path.join( lowercase__ ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowercase__ ): copyfile(self.vocab_file ,lowercase__ ) return (out_vocab_file,)
624
0
'''simple docstring''' def _A ( A__ ): # noqa: E741 __lowercase = len(A__ ) __lowercase = 0 __lowercase = [0] * n __lowercase = [False] * n __lowercase = [False] * n def dfs(A__ , A__ , A__ , A__ ): if parent == root: out_edge_count += 1 __lowercase = True __lowercase = at for to in l[at]: if to == parent: pass elif not visited[to]: __lowercase = dfs(A__ , A__ , A__ , A__ ) __lowercase = min(low[at] , low[to] ) # AP found via bridge if at < low[to]: __lowercase = True # AP found via cycle if at == low[to]: __lowercase = True else: __lowercase = min(low[at] , A__ ) return out_edge_count for i in range(A__ ): if not visited[i]: __lowercase = 0 __lowercase = dfs(A__ , A__ , -1 , A__ ) __lowercase = out_edge_count > 1 for x in range(len(A__ ) ): if is_art[x] is True: print(A__ ) # Adjacency list of graph lowerCAmelCase__ = { 0: [1, 2], 1: [0, 2], 2: [0, 1, 3, 5], 3: [2, 4], 4: [3], 5: [2, 6, 8], 6: [5, 7], 7: [6, 8], 8: [5, 7], } compute_ap(data)
712
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' def _A ( A__ = 1000000 ): """simple docstring""" __lowercase = set(range(3 , A__ , 2 ) ) primes.add(2 ) for p in range(3 , A__ , 2 ): if p not in primes: continue primes.difference_update(set(range(p * p , A__ , A__ ) ) ) __lowercase = [float(A__ ) for n in range(limit + 1 )] for p in primes: for n in range(A__ , limit + 1 , A__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(f'{solution() = }')
713
'''simple docstring''' import copy import json import os import tempfile from transformers import is_torch_available from .test_configuration_utils import config_common_kwargs class lowercase_ (lowerCamelCase__ ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase__ : int ,lowercase__ : List[str]=None ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=None ,**lowercase__ : Dict ): __lowercase = parent __lowercase = config_class __lowercase = has_text_modality __lowercase = kwargs __lowercase = common_properties def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = ( ['''hidden_size''', '''num_attention_heads''', '''num_hidden_layers'''] if self.common_properties is None else self.common_properties ) # Add common fields for text models if self.has_text_modality: common_properties.extend(['''vocab_size'''] ) # Test that config has the common properties as getters for prop in common_properties: self.parent.assertTrue(hasattr(lowercase__ ,lowercase__ ) ,msg=F"`{prop}` does not exist" ) # Test that config has the common properties as setter for idx, name in enumerate(lowercase__ ): try: setattr(lowercase__ ,lowercase__ ,lowercase__ ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass # Test if config class can be called with Config(prop_name=..) for idx, name in enumerate(lowercase__ ): try: __lowercase = self.config_class(**{name: idx} ) self.parent.assertEqual( getattr(lowercase__ ,lowercase__ ) ,lowercase__ ,msg=F"`{name} value {idx} expected, but was {getattr(lowercase__ ,lowercase__ )}" ) except NotImplementedError: # Some models might not be able to implement setters for common_properties # In that case, a NotImplementedError is raised pass def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = json.loads(config.to_json_string() ) for key, value in self.inputs_dict.items(): self.parent.assertEqual(obj[key] ,lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,'''config.json''' ) config_first.to_json_file(lowercase__ ) __lowercase = self.config_class.from_json_file(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ) with tempfile.TemporaryDirectory() as tmpdirname: config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.config_class(**self.inputs_dict ) __lowercase = '''test''' with tempfile.TemporaryDirectory() as tmpdirname: __lowercase = os.path.join(lowercase__ ,lowercase__ ) config_first.save_pretrained(lowercase__ ) __lowercase = self.config_class.from_pretrained(lowercase__ ,subfolder=lowercase__ ) self.parent.assertEqual(config_second.to_dict() ,config_first.to_dict() ) def SCREAMING_SNAKE_CASE ( self : Optional[Any] ): __lowercase = self.config_class(**self.inputs_dict ,num_labels=5 ) self.parent.assertEqual(len(config.idalabel ) ,5 ) self.parent.assertEqual(len(config.labelaid ) ,5 ) __lowercase = 3 self.parent.assertEqual(len(config.idalabel ) ,3 ) self.parent.assertEqual(len(config.labelaid ) ,3 ) def SCREAMING_SNAKE_CASE ( self : Tuple ): if self.config_class.is_composition: return __lowercase = self.config_class() self.parent.assertIsNotNone(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = copy.deepcopy(lowercase__ ) __lowercase = self.config_class(**lowercase__ ) __lowercase = [] for key, value in config_common_kwargs.items(): if key == "torch_dtype": if not is_torch_available(): continue else: import torch if config.torch_dtype != torch.floataa: wrong_values.append(('''torch_dtype''', config.torch_dtype, torch.floataa) ) elif getattr(lowercase__ ,lowercase__ ) != value: wrong_values.append((key, getattr(lowercase__ ,lowercase__ ), value) ) if len(lowercase__ ) > 0: __lowercase = '''\n'''.join([F"- {v[0]}: got {v[1]} instead of {v[2]}" for v in wrong_values] ) raise ValueError(F"The following keys were not properly set in the config:\n{errors}" ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ): self.create_and_test_config_common_properties() self.create_and_test_config_to_json_string() self.create_and_test_config_to_json_file() self.create_and_test_config_from_and_save_pretrained() self.create_and_test_config_from_and_save_pretrained_subfolder() self.create_and_test_config_with_num_labels() self.check_config_can_be_init_without_params() self.check_config_arguments_init()
624
0
'''simple docstring''' def _A ( A__ ): """simple docstring""" __lowercase = abs(A__ ) __lowercase = 0 while n > 0: res += n % 10 n //= 10 return res def _A ( A__ ): """simple docstring""" __lowercase = abs(A__ ) return n if n < 10 else n % 10 + sum_of_digits(n // 10 ) def _A ( A__ ): """simple docstring""" return sum(int(A__ ) for c in str(abs(A__ ) ) ) def _A ( ): """simple docstring""" from collections.abc import Callable from timeit import timeit def benchmark_a_function(A__ , A__ ) -> None: __lowercase = F"{func.__name__}({value})" __lowercase = timeit(F"__main__.{call}" , setup='''import __main__''' ) print(F"{call:56} = {func(A__ )} -- {timing:.4f} seconds" ) for value in (262144, 1125899906842624, 1267650600228229401496703205376): for func in (sum_of_digits, sum_of_digits_recursion, sum_of_digits_compact): benchmark_a_function(A__ , A__ ) print() if __name__ == "__main__": import doctest doctest.testmod() benchmark()
714
'''simple docstring''' import re def _A ( A__ ): """simple docstring""" __lowercase = re.compile( R'''^(?:0|94|\+94|0{2}94)''' R'''7(0|1|2|4|5|6|7|8)''' R'''(-| |)''' R'''\d{7}$''' ) return bool(re.search(A__ , A__ ) ) if __name__ == "__main__": lowerCAmelCase__ = '''0094702343221''' print(is_sri_lankan_phone_number(phone))
624
0
'''simple docstring''' import darl # noqa import gym import tqdm from diffusers.experimental import ValueGuidedRLPipeline lowerCAmelCase__ = { '''n_samples''': 64, '''horizon''': 32, '''num_inference_steps''': 20, '''n_guide_steps''': 2, # can set to 0 for faster sampling, does not use value network '''scale_grad_by_std''': True, '''scale''': 0.1, '''eta''': 0.0, '''t_grad_cutoff''': 2, '''device''': '''cpu''', } if __name__ == "__main__": lowerCAmelCase__ = '''hopper-medium-v2''' lowerCAmelCase__ = gym.make(env_name) lowerCAmelCase__ = ValueGuidedRLPipeline.from_pretrained( '''bglick13/hopper-medium-v2-value-function-hor32''', env=env, ) env.seed(0) lowerCAmelCase__ = env.reset() lowerCAmelCase__ = 0 lowerCAmelCase__ = 0 lowerCAmelCase__ = 1000 lowerCAmelCase__ = [obs.copy()] try: for t in tqdm.tqdm(range(T)): # call the policy lowerCAmelCase__ = pipeline(obs, planning_horizon=32) # execute action in environment lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ = env.step(denorm_actions) lowerCAmelCase__ = env.get_normalized_score(total_reward) # update return total_reward += reward total_score += score print( f'Step: {t}, Reward: {reward}, Total Reward: {total_reward}, Score: {score}, Total Score:' f' {total_score}' ) # save observations for rendering rollout.append(next_observation.copy()) lowerCAmelCase__ = next_observation except KeyboardInterrupt: pass print(f'Total reward: {total_reward}')
715
'''simple docstring''' from __future__ import annotations from typing import Any class lowercase_ : """simple docstring""" def __init__( self : Any ,lowercase__ : int ,lowercase__ : int ,lowercase__ : float = 0 ): __lowercase , __lowercase = row, column __lowercase = [[default_value for c in range(lowercase__ )] for r in range(lowercase__ )] def __str__( self : List[str] ): __lowercase = F"Matrix consist of {self.row} rows and {self.column} columns\n" # Make string identifier __lowercase = 0 for row_vector in self.array: for obj in row_vector: __lowercase = max(lowercase__ ,len(str(lowercase__ ) ) ) __lowercase = F"%{max_element_length}s" # Make string and return def single_line(lowercase__ : list[float] ) -> str: nonlocal string_format_identifier __lowercase = '''[''' line += ", ".join(string_format_identifier % (obj,) for obj in row_vector ) line += "]" return line s += "\n".join(single_line(lowercase__ ) for row_vector in self.array ) return s def __repr__( self : List[str] ): return str(self ) def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : tuple[int, int] ): if not (isinstance(lowercase__ ,(list, tuple) ) and len(lowercase__ ) == 2): return False elif not (0 <= loc[0] < self.row and 0 <= loc[1] < self.column): return False else: return True def __getitem__( self : Tuple ,lowercase__ : tuple[int, int] ): assert self.validate_indicies(lowercase__ ) return self.array[loc[0]][loc[1]] def __setitem__( self : Tuple ,lowercase__ : tuple[int, int] ,lowercase__ : float ): assert self.validate_indicies(lowercase__ ) __lowercase = value def __add__( self : List[Any] ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) assert self.row == another.row and self.column == another.column # Add __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] + another[r, c] return result def __neg__( self : List[str] ): __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = -self[r, c] return result def __sub__( self : str ,lowercase__ : Matrix ): return self + (-another) def __mul__( self : Dict ,lowercase__ : int | float | Matrix ): if isinstance(lowercase__ ,(int, float) ): # Scalar multiplication __lowercase = Matrix(self.row ,self.column ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] * another return result elif isinstance(lowercase__ ,lowercase__ ): # Matrix multiplication assert self.column == another.row __lowercase = Matrix(self.row ,another.column ) for r in range(self.row ): for c in range(another.column ): for i in range(self.column ): result[r, c] += self[r, i] * another[i, c] return result else: __lowercase = F"Unsupported type given for another ({type(lowercase__ )})" raise TypeError(lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = Matrix(self.column ,self.row ) for r in range(self.row ): for c in range(self.column ): __lowercase = self[r, c] return result def SCREAMING_SNAKE_CASE ( self : Any ,lowercase__ : Matrix ,lowercase__ : Matrix ): assert isinstance(lowercase__ ,lowercase__ ) and isinstance(lowercase__ ,lowercase__ ) assert self.row == self.column == u.row == v.row # u, v should be column vector assert u.column == v.column == 1 # u, v should be column vector # Calculate __lowercase = v.transpose() __lowercase = (v_t * self * u)[0, 0] + 1 if numerator_factor == 0: return None # It's not invertable return self - ((self * u) * (v_t * self) * (1.0 / numerator_factor)) # Testing if __name__ == "__main__": def _A ( ): """simple docstring""" __lowercase = Matrix(3 , 3 , 0 ) for i in range(3 ): __lowercase = 1 print(F"a^(-1) is {ainv}" ) # u, v __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 1, 2, -3 __lowercase = Matrix(3 , 1 , 0 ) __lowercase , __lowercase , __lowercase = 4, -2, 5 print(F"u is {u}" ) print(F"v is {v}" ) print(F"uv^T is {u * v.transpose()}" ) # Sherman Morrison print(F"(a + uv^T)^(-1) is {ainv.sherman_morrison(A__ , A__ )}" ) def _A ( ): """simple docstring""" import doctest doctest.testmod() testa()
624
0
'''simple docstring''' from __future__ import annotations import unittest from transformers import FunnelConfig, is_tf_available from transformers.testing_utils import require_tf from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFFunnelBaseModel, TFFunnelForMaskedLM, TFFunnelForMultipleChoice, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForSequenceClassification, TFFunnelForTokenClassification, TFFunnelModel, ) class lowercase_ : """simple docstring""" def __init__( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : List[Any]=1_3 ,lowercase__ : Union[str, Any]=7 ,lowercase__ : List[Any]=True ,lowercase__ : str=True ,lowercase__ : Any=True ,lowercase__ : Union[str, Any]=True ,lowercase__ : Tuple=9_9 ,lowercase__ : List[str]=[1, 1, 2] ,lowercase__ : List[Any]=1 ,lowercase__ : List[Any]=3_2 ,lowercase__ : int=4 ,lowercase__ : Tuple=8 ,lowercase__ : Tuple=3_7 ,lowercase__ : str="gelu_new" ,lowercase__ : Dict=0.1 ,lowercase__ : Optional[int]=0.1 ,lowercase__ : Optional[Any]=0.0 ,lowercase__ : Tuple=5_1_2 ,lowercase__ : Any=3 ,lowercase__ : List[str]=0.0_2 ,lowercase__ : List[Any]=3 ,lowercase__ : List[Any]=4 ,lowercase__ : Optional[Any]=None ,lowercase__ : Tuple=False ,): __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_input_mask __lowercase = use_token_type_ids __lowercase = use_labels __lowercase = vocab_size __lowercase = block_sizes __lowercase = num_decoder_layers __lowercase = d_model __lowercase = n_head __lowercase = d_head __lowercase = d_inner __lowercase = hidden_act __lowercase = hidden_dropout __lowercase = attention_dropout __lowercase = activation_dropout __lowercase = max_position_embeddings __lowercase = type_vocab_size __lowercase = 2 __lowercase = num_labels __lowercase = num_choices __lowercase = scope __lowercase = initializer_std # Used in the tests to check the size of the first attention layer __lowercase = n_head # Used in the tests to check the size of the first hidden state __lowercase = self.d_model # Used in the tests to check the number of output hidden states/attentions __lowercase = sum(self.block_sizes ) + (0 if base else self.num_decoder_layers) # FunnelModel adds two hidden layers: input embeddings and the sum of the upsampled encoder hidden state with # the last hidden state of the first block (which is the first hidden state of the decoder). if not base: __lowercase = self.num_hidden_layers + 2 def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) __lowercase = None if self.use_input_mask: __lowercase = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase = None if self.use_token_type_ids: __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.type_vocab_size ) __lowercase = None __lowercase = None __lowercase = None if self.use_labels: __lowercase = ids_tensor([self.batch_size] ,self.type_sequence_label_size ) __lowercase = ids_tensor([self.batch_size, self.seq_length] ,self.num_labels ) __lowercase = ids_tensor([self.batch_size] ,self.num_choices ) __lowercase = FunnelConfig( vocab_size=self.vocab_size ,block_sizes=self.block_sizes ,num_decoder_layers=self.num_decoder_layers ,d_model=self.d_model ,n_head=self.n_head ,d_head=self.d_head ,d_inner=self.d_inner ,hidden_act=self.hidden_act ,hidden_dropout=self.hidden_dropout ,attention_dropout=self.attention_dropout ,activation_dropout=self.activation_dropout ,max_position_embeddings=self.max_position_embeddings ,type_vocab_size=self.type_vocab_size ,initializer_std=self.initializer_std ,) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[str] ,lowercase__ : Union[str, Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : Optional[int] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,): __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) __lowercase = False __lowercase = TFFunnelModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, self.seq_length, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : int ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Optional[Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Dict ,): __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) __lowercase = [input_ids, input_mask] __lowercase = model(lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 3, self.d_model) ) __lowercase = False __lowercase = TFFunnelBaseModel(config=lowercase__ ) __lowercase = model(lowercase__ ) self.parent.assertEqual(result.last_hidden_state.shape ,(self.batch_size, 2, self.d_model) ) def SCREAMING_SNAKE_CASE ( self : Tuple ,lowercase__ : Any ,lowercase__ : Any ,lowercase__ : List[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : int ,lowercase__ : List[str] ,): __lowercase = TFFunnelForPreTraining(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ,lowercase__ : Union[str, Any] ,lowercase__ : Optional[Any] ,lowercase__ : Dict ,lowercase__ : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : Tuple ,): __lowercase = TFFunnelForMaskedLM(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.vocab_size) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Tuple ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[Any] ,lowercase__ : str ,lowercase__ : Union[str, Any] ,lowercase__ : Tuple ,): __lowercase = self.num_labels __lowercase = TFFunnelForSequenceClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : Dict ,lowercase__ : List[str] ,lowercase__ : Optional[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,): __lowercase = self.num_choices __lowercase = TFFunnelForMultipleChoice(config=lowercase__ ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = tf.tile(tf.expand_dims(lowercase__ ,1 ) ,(1, self.num_choices, 1) ) __lowercase = { '''input_ids''': multiple_choice_inputs_ids, '''attention_mask''': multiple_choice_input_mask, '''token_type_ids''': multiple_choice_token_type_ids, } __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.num_choices) ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ,lowercase__ : Optional[Any] ,lowercase__ : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[Any] ,lowercase__ : Dict ,lowercase__ : Any ,lowercase__ : Union[str, Any] ,): __lowercase = self.num_labels __lowercase = TFFunnelForTokenClassification(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.logits.shape ,(self.batch_size, self.seq_length, self.num_labels) ) def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : Optional[Any] ,lowercase__ : Optional[int] ,lowercase__ : Dict ,lowercase__ : List[str] ,lowercase__ : str ,lowercase__ : Tuple ,lowercase__ : Any ,): __lowercase = TFFunnelForQuestionAnswering(config=lowercase__ ) __lowercase = {'''input_ids''': input_ids, '''attention_mask''': input_mask, '''token_type_ids''': token_type_ids} __lowercase = model(lowercase__ ) self.parent.assertEqual(result.start_logits.shape ,(self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape ,(self.batch_size, self.seq_length) ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.prepare_config_and_inputs() ( ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ( __lowercase ) , ) = config_and_inputs __lowercase = {'''input_ids''': input_ids, '''token_type_ids''': token_type_ids, '''attention_mask''': input_mask} return config, inputs_dict @require_tf class lowercase_ (lowerCamelCase__ , lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : int = ( ( TFFunnelModel, TFFunnelForMaskedLM, TFFunnelForPreTraining, TFFunnelForQuestionAnswering, TFFunnelForTokenClassification, ) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Union[str, Any] = ( { 'feature-extraction': (TFFunnelBaseModel, TFFunnelModel), 'fill-mask': TFFunnelForMaskedLM, 'question-answering': TFFunnelForQuestionAnswering, 'text-classification': TFFunnelForSequenceClassification, 'token-classification': TFFunnelForTokenClassification, 'zero-shot': TFFunnelForSequenceClassification, } if is_tf_available() else {} ) SCREAMING_SNAKE_CASE : Tuple = False SCREAMING_SNAKE_CASE : Any = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : str ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : Optional[int] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : List[Any] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : int ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*lowercase__ ) @require_tf class lowercase_ (lowerCamelCase__ , unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = ( (TFFunnelBaseModel, TFFunnelForMultipleChoice, TFFunnelForSequenceClassification) if is_tf_available() else () ) SCREAMING_SNAKE_CASE : Dict = False SCREAMING_SNAKE_CASE : List[str] = False def SCREAMING_SNAKE_CASE ( self : str ): __lowercase = TFFunnelModelTester(self ,base=lowercase__ ) __lowercase = ConfigTester(self ,config_class=lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): self.config_tester.run_common_tests() def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_base_model(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Dict ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*lowercase__ ) def SCREAMING_SNAKE_CASE ( self : Any ): __lowercase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*lowercase__ )
716
'''simple docstring''' def _A ( A__ = 50 ): """simple docstring""" __lowercase = [1] * (length + 1) for row_length in range(3 , length + 1 ): for block_length in range(3 , row_length + 1 ): for block_start in range(row_length - block_length ): ways_number[row_length] += ways_number[ row_length - block_start - block_length - 1 ] ways_number[row_length] += 1 return ways_number[length] if __name__ == "__main__": print(f'{solution() = }')
624
0
'''simple docstring''' import argparse import torch from transformers import YosoConfig, YosoForMaskedLM def _A ( A__ ): """simple docstring""" if "model" in orig_key: __lowercase = orig_key.replace('''model.''' , '''''' ) if "norm1" in orig_key: __lowercase = orig_key.replace('''norm1''' , '''attention.output.LayerNorm''' ) if "norm2" in orig_key: __lowercase = orig_key.replace('''norm2''' , '''output.LayerNorm''' ) if "norm" in orig_key: __lowercase = orig_key.replace('''norm''' , '''LayerNorm''' ) if "transformer" in orig_key: __lowercase = orig_key.split('''.''' )[0].split('''_''' )[-1] __lowercase = orig_key.replace(F"transformer_{layer_num}" , F"encoder.layer.{layer_num}" ) if "mha.attn" in orig_key: __lowercase = orig_key.replace('''mha.attn''' , '''attention.self''' ) if "mha" in orig_key: __lowercase = orig_key.replace('''mha''' , '''attention''' ) if "W_q" in orig_key: __lowercase = orig_key.replace('''W_q''' , '''self.query''' ) if "W_k" in orig_key: __lowercase = orig_key.replace('''W_k''' , '''self.key''' ) if "W_v" in orig_key: __lowercase = orig_key.replace('''W_v''' , '''self.value''' ) if "ff1" in orig_key: __lowercase = orig_key.replace('''ff1''' , '''intermediate.dense''' ) if "ff2" in orig_key: __lowercase = orig_key.replace('''ff2''' , '''output.dense''' ) if "ff" in orig_key: __lowercase = orig_key.replace('''ff''' , '''output.dense''' ) if "mlm_class" in orig_key: __lowercase = orig_key.replace('''mlm.mlm_class''' , '''cls.predictions.decoder''' ) if "mlm" in orig_key: __lowercase = orig_key.replace('''mlm''' , '''cls.predictions.transform''' ) if "cls" not in orig_key: __lowercase = '''yoso.''' + orig_key return orig_key def _A ( A__ , A__ ): """simple docstring""" for key in orig_state_dict.copy().keys(): __lowercase = orig_state_dict.pop(A__ ) if ("pooler" in key) or ("sen_class" in key): continue else: __lowercase = val __lowercase = orig_state_dict['''cls.predictions.decoder.bias'''] __lowercase = torch.arange(A__ ).expand((1, -1) ) + 2 return orig_state_dict def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = torch.load(A__ , map_location='''cpu''' )['''model_state_dict'''] __lowercase = YosoConfig.from_json_file(A__ ) __lowercase = YosoForMaskedLM(A__ ) __lowercase = convert_checkpoint_helper(config.max_position_embeddings , A__ ) print(model.load_state_dict(A__ ) ) model.eval() model.save_pretrained(A__ ) print(F"Checkpoint successfuly converted. Model saved at {pytorch_dump_path}" ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--pytorch_model_path''', default=None, type=str, required=True, help='''Path to YOSO pytorch checkpoint.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help='''The json file for YOSO model config.''', ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) lowerCAmelCase__ = parser.parse_args() convert_yoso_checkpoint(args.pytorch_model_path, args.config_file, args.pytorch_dump_path)
717
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) lowerCAmelCase__ = logging.getLogger(__name__) lowerCAmelCase__ = '''Hello world! cécé herlolip''' lowerCAmelCase__ = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = BertAbsConfig( temp_dir='''.''' , finetune_bert=A__ , large=A__ , share_emb=A__ , use_bert_emb=A__ , encoder='''bert''' , max_pos=512 , enc_layers=6 , enc_hidden_size=512 , enc_heads=8 , enc_ff_size=512 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=768 , dec_heads=8 , dec_ff_size=2048 , dec_dropout=0.2 , ) __lowercase = torch.load(A__ , lambda A__ , A__ : storage ) __lowercase = AbsSummarizer(A__ , torch.device('''cpu''' ) , A__ ) original.eval() __lowercase = BertAbsSummarizer(A__ , torch.device('''cpu''' ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info('''convert the model''' ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info('''Make sure that the models\' outputs are identical''' ) __lowercase = BertTokenizer.from_pretrained('''bert-base-uncased''' ) # prepare the model inputs __lowercase = tokenizer.encode('''This is sample éàalj\'-.''' ) encoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) __lowercase = tokenizer.encode('''This is sample 3 éàalj\'-.''' ) decoder_input_ids.extend([tokenizer.pad_token_id] * (512 - len(A__ )) ) __lowercase = torch.tensor(A__ ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __lowercase = encoder_input_ids __lowercase = decoder_input_ids __lowercase = __lowercase = None __lowercase = None __lowercase = __lowercase = None __lowercase = __lowercase = None __lowercase = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __lowercase = original(A__ , A__ , A__ , A__ , A__ , A__ , A__ )[0] __lowercase = original.generator(A__ ) __lowercase = new_model( A__ , A__ , A__ , A__ , A__ )[0] __lowercase = new_model.generator(A__ ) __lowercase = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print('''Maximum absolute difference beween weights: {:.2f}'''.format(A__ ) ) __lowercase = torch.allclose(A__ , A__ , atol=1e-3 ) if are_identical: logging.info('''all weights are equal up to 1e-3''' ) else: raise ValueError('''the weights are different. The new model is likely different from the original one.''' ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info('''saving the model\'s state dictionary''' ) torch.save( new_model.state_dict() , '''./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin''' ) if __name__ == "__main__": lowerCAmelCase__ = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) lowerCAmelCase__ = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
624
0
'''simple docstring''' import os import pytest from datasets import ( get_dataset_config_info, get_dataset_config_names, get_dataset_infos, get_dataset_split_names, inspect_dataset, inspect_metric, ) lowerCAmelCase__ = pytest.mark.integration @pytest.mark.parametrize('''path''' , ['''paws''', '''csv'''] ) def _A ( A__ , A__ ): """simple docstring""" inspect_dataset(A__ , A__ ) __lowercase = path + '''.py''' assert script_name in os.listdir(A__ ) assert "__pycache__" not in os.listdir(A__ ) @pytest.mark.filterwarnings('''ignore:inspect_metric is deprecated:FutureWarning''' ) @pytest.mark.filterwarnings('''ignore:metric_module_factory is deprecated:FutureWarning''' ) @pytest.mark.parametrize('''path''' , ['''accuracy'''] ) def _A ( A__ , A__ ): """simple docstring""" inspect_metric(A__ , A__ ) __lowercase = path + '''.py''' assert script_name in os.listdir(A__ ) assert "__pycache__" not in os.listdir(A__ ) @pytest.mark.parametrize( '''path, config_name, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = get_dataset_config_info(A__ , config_name=A__ ) assert info.config_name == config_name assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _A ( A__ , A__ , A__ ): """simple docstring""" with pytest.raises(A__ ): get_dataset_config_info(A__ , config_name=A__ ) @pytest.mark.parametrize( '''path, expected''' , [ ('''squad''', '''plain_text'''), ('''acronym_identification''', '''default'''), ('''lhoestq/squad''', '''plain_text'''), ('''lhoestq/test''', '''default'''), ('''lhoestq/demo1''', '''lhoestq--demo1'''), ('''dalle-mini/wit''', '''dalle-mini--wit'''), ] , ) def _A ( A__ , A__ ): """simple docstring""" __lowercase = get_dataset_config_names(A__ ) assert expected in config_names @pytest.mark.parametrize( '''path, expected_configs, expected_splits_in_first_config''' , [ ('''squad''', ['''plain_text'''], ['''train''', '''validation''']), ('''dalle-mini/wit''', ['''dalle-mini--wit'''], ['''train''']), ('''paws''', ['''labeled_final''', '''labeled_swap''', '''unlabeled_final'''], ['''train''', '''test''', '''validation''']), ] , ) def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = get_dataset_infos(A__ ) assert list(infos.keys() ) == expected_configs __lowercase = expected_configs[0] assert expected_config in infos __lowercase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits_in_first_config @pytest.mark.parametrize( '''path, expected_config, expected_splits''' , [ ('''squad''', '''plain_text''', ['''train''', '''validation''']), ('''dalle-mini/wit''', '''dalle-mini--wit''', ['''train''']), ('''paws''', '''labeled_final''', ['''train''', '''test''', '''validation''']), ] , ) def _A ( A__ , A__ , A__ ): """simple docstring""" __lowercase = get_dataset_infos(A__ ) assert expected_config in infos __lowercase = infos[expected_config] assert info.config_name == expected_config assert list(info.splits.keys() ) == expected_splits @pytest.mark.parametrize( '''path, config_name, expected_exception''' , [ ('''paws''', None, ValueError), ] , ) def _A ( A__ , A__ , A__ ): """simple docstring""" with pytest.raises(A__ ): get_dataset_split_names(A__ , config_name=A__ )
718
'''simple docstring''' import hashlib import unittest from transformers import MODEL_FOR_DEPTH_ESTIMATION_MAPPING, is_torch_available, is_vision_available from transformers.pipelines import DepthEstimationPipeline, pipeline from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_tf, require_timm, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_torch_available(): import torch if is_vision_available(): from PIL import Image else: class lowercase_ : """simple docstring""" @staticmethod def SCREAMING_SNAKE_CASE ( *lowercase__ : Union[str, Any] ,**lowercase__ : Tuple ): pass def _A ( A__ ): """simple docstring""" __lowercase = hashlib.mda(image.tobytes() ) return m.hexdigest() @is_pipeline_test @require_vision @require_timm @require_torch class lowercase_ (unittest.TestCase ): """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = MODEL_FOR_DEPTH_ESTIMATION_MAPPING def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : int ,lowercase__ : List[str] ,lowercase__ : int ): __lowercase = DepthEstimationPipeline(model=lowercase__ ,image_processor=lowercase__ ) return depth_estimator, [ "./tests/fixtures/tests_samples/COCO/000000039769.png", "./tests/fixtures/tests_samples/COCO/000000039769.png", ] def SCREAMING_SNAKE_CASE ( self : List[Any] ,lowercase__ : List[str] ,lowercase__ : List[str] ): __lowercase = depth_estimator('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) self.assertEqual({'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )} ,lowercase__ ) import datasets __lowercase = datasets.load_dataset('''hf-internal-testing/fixtures_image_utils''' ,'''image''' ,split='''test''' ) __lowercase = depth_estimator( [ Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ), '''http://images.cocodataset.org/val2017/000000039769.jpg''', # RGBA dataset[0]['''file'''], # LA dataset[1]['''file'''], # L dataset[2]['''file'''], ] ) self.assertEqual( [ {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, {'''predicted_depth''': ANY(torch.Tensor ), '''depth''': ANY(Image.Image )}, ] ,lowercase__ ,) @require_tf @unittest.skip('''Depth estimation is not implemented in TF''' ) def SCREAMING_SNAKE_CASE ( self : Dict ): pass @slow @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ): __lowercase = '''Intel/dpt-large''' __lowercase = pipeline('''depth-estimation''' ,model=lowercase__ ) __lowercase = depth_estimator('''http://images.cocodataset.org/val2017/000000039769.jpg''' ) __lowercase = hashimage(outputs['''depth'''] ) # This seems flaky. # self.assertEqual(outputs["depth"], "1a39394e282e9f3b0741a90b9f108977") self.assertEqual(nested_simplify(outputs['''predicted_depth'''].max().item() ) ,2_9.3_0_4 ) self.assertEqual(nested_simplify(outputs['''predicted_depth'''].min().item() ) ,2.6_6_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[Any] ): # This is highly irregular to have no small tests. self.skipTest('''There is not hf-internal-testing tiny model for either GLPN nor DPT''' )
624
0
import collections import os import re from pathlib import Path lowerCAmelCase__ = '''src/transformers''' # Matches is_xxx_available() lowerCAmelCase__ = re.compile(R'''is\_([a-z_]*)_available()''') # Catches a one-line _import_struct = {xxx} lowerCAmelCase__ = re.compile(R'''^_import_structure\s+=\s+\{([^\}]+)\}''') # Catches a line with a key-values pattern: "bla": ["foo", "bar"] lowerCAmelCase__ = re.compile(R'''\s+"\S*":\s+\[([^\]]*)\]''') # Catches a line if not is_foo_available lowerCAmelCase__ = re.compile(R'''^\s*if\s+not\s+is\_[a-z_]*\_available\(\)''') # Catches a line _import_struct["bla"].append("foo") lowerCAmelCase__ = re.compile(R'''^\s*_import_structure\["\S*"\]\.append\("(\S*)"\)''') # Catches a line _import_struct["bla"].extend(["foo", "bar"]) or _import_struct["bla"] = ["foo", "bar"] lowerCAmelCase__ = re.compile(R'''^\s*_import_structure\[\S*\](?:\.extend\(|\s*=\s+)\[([^\]]*)\]''') # Catches a line with an object between quotes and a comma: "MyModel", lowerCAmelCase__ = re.compile(R'''^\s+"([^"]+)",''') # Catches a line with objects between brackets only: ["foo", "bar"], lowerCAmelCase__ = re.compile(R'''^\s+\[([^\]]+)\]''') # Catches a line with from foo import bar, bla, boo lowerCAmelCase__ = re.compile(R'''\s+from\s+\S*\s+import\s+([^\(\s].*)\n''') # Catches a line with try: lowerCAmelCase__ = re.compile(R'''^\s*try:''') # Catches a line with else: lowerCAmelCase__ = re.compile(R'''^\s*else:''') def _A ( A__ ): """simple docstring""" if _re_test_backend.search(A__ ) is None: return None __lowercase = [b[0] for b in _re_backend.findall(A__ )] backends.sort() return "_and_".join(A__ ) def _A ( A__ ): """simple docstring""" with open(A__ , '''r''' , encoding='''utf-8''' , newline='''\n''' ) as f: __lowercase = f.readlines() __lowercase = 0 while line_index < len(A__ ) and not lines[line_index].startswith('''_import_structure = {''' ): line_index += 1 # If this is a traditional init, just return. if line_index >= len(A__ ): return None # First grab the objects without a specific backend in _import_structure __lowercase = [] while not lines[line_index].startswith('''if TYPE_CHECKING''' ) and find_backend(lines[line_index] ) is None: __lowercase = lines[line_index] # If we have everything on a single line, let's deal with it. if _re_one_line_import_struct.search(A__ ): __lowercase = _re_one_line_import_struct.search(A__ ).groups()[0] __lowercase = re.findall(R'''\[([^\]]+)\]''' , A__ ) for imp in imports: objects.extend([obj[1:-1] for obj in imp.split(''', ''' )] ) line_index += 1 continue __lowercase = _re_import_struct_key_value.search(A__ ) if single_line_import_search is not None: __lowercase = [obj[1:-1] for obj in single_line_import_search.groups()[0].split(''', ''' ) if len(A__ ) > 0] objects.extend(A__ ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) line_index += 1 __lowercase = {'''none''': objects} # Let's continue with backend-specific objects in _import_structure while not lines[line_index].startswith('''if TYPE_CHECKING''' ): # If the line is an if not is_backend_available, we grab all objects associated. __lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 4 ): __lowercase = lines[line_index] if _re_import_struct_add_one.search(A__ ) is not None: objects.append(_re_import_struct_add_one.search(A__ ).groups()[0] ) elif _re_import_struct_add_many.search(A__ ) is not None: __lowercase = _re_import_struct_add_many.search(A__ ).groups()[0].split(''', ''' ) __lowercase = [obj[1:-1] for obj in imports if len(A__ ) > 0] objects.extend(A__ ) elif _re_between_brackets.search(A__ ) is not None: __lowercase = _re_between_brackets.search(A__ ).groups()[0].split(''', ''' ) __lowercase = [obj[1:-1] for obj in imports if len(A__ ) > 0] objects.extend(A__ ) elif _re_quote_object.search(A__ ) is not None: objects.append(_re_quote_object.search(A__ ).groups()[0] ) elif line.startswith(''' ''' * 8 + '''"''' ): objects.append(line[9:-3] ) elif line.startswith(''' ''' * 12 + '''"''' ): objects.append(line[13:-3] ) line_index += 1 __lowercase = objects else: line_index += 1 # At this stage we are in the TYPE_CHECKING part, first grab the objects without a specific backend __lowercase = [] while ( line_index < len(A__ ) and find_backend(lines[line_index] ) is None and not lines[line_index].startswith('''else''' ) ): __lowercase = lines[line_index] __lowercase = _re_import.search(A__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 8 ): objects.append(line[8:-2] ) line_index += 1 __lowercase = {'''none''': objects} # Let's continue with backend-specific objects while line_index < len(A__ ): # If the line is an if is_backend_available, we grab all objects associated. __lowercase = find_backend(lines[line_index] ) # Check if the backend declaration is inside a try block: if _re_try.search(lines[line_index - 1] ) is None: __lowercase = None if backend is not None: line_index += 1 # Scroll until we hit the else block of try-except-else while _re_else.search(lines[line_index] ) is None: line_index += 1 line_index += 1 __lowercase = [] # Until we unindent, add backend objects to the list while len(lines[line_index] ) <= 1 or lines[line_index].startswith(''' ''' * 8 ): __lowercase = lines[line_index] __lowercase = _re_import.search(A__ ) if single_line_import_search is not None: objects.extend(single_line_import_search.groups()[0].split(''', ''' ) ) elif line.startswith(''' ''' * 12 ): objects.append(line[12:-2] ) line_index += 1 __lowercase = objects else: line_index += 1 return import_dict_objects, type_hint_objects def _A ( A__ , A__ ): """simple docstring""" def find_duplicates(A__ ): return [k for k, v in collections.Counter(A__ ).items() if v > 1] if list(import_dict_objects.keys() ) != list(type_hint_objects.keys() ): return ["Both sides of the init do not have the same backends!"] __lowercase = [] for key in import_dict_objects.keys(): __lowercase = find_duplicates(import_dict_objects[key] ) if duplicate_imports: errors.append(F"Duplicate _import_structure definitions for: {duplicate_imports}" ) __lowercase = find_duplicates(type_hint_objects[key] ) if duplicate_type_hints: errors.append(F"Duplicate TYPE_CHECKING objects for: {duplicate_type_hints}" ) if sorted(set(import_dict_objects[key] ) ) != sorted(set(type_hint_objects[key] ) ): __lowercase = '''base imports''' if key == '''none''' else F"{key} backend" errors.append(F"Differences for {name}:" ) for a in type_hint_objects[key]: if a not in import_dict_objects[key]: errors.append(F" {a} in TYPE_HINT but not in _import_structure." ) for a in import_dict_objects[key]: if a not in type_hint_objects[key]: errors.append(F" {a} in _import_structure but not in TYPE_HINT." ) return errors def _A ( ): """simple docstring""" __lowercase = [] for root, _, files in os.walk(A__ ): if "__init__.py" in files: __lowercase = os.path.join(A__ , '''__init__.py''' ) __lowercase = parse_init(A__ ) if objects is not None: __lowercase = analyze_results(*A__ ) if len(A__ ) > 0: __lowercase = F"Problem in {fname}, both halves do not define the same objects.\n{errors[0]}" failures.append('''\n'''.join(A__ ) ) if len(A__ ) > 0: raise ValueError('''\n\n'''.join(A__ ) ) def _A ( ): """simple docstring""" __lowercase = [] for path, directories, files in os.walk(A__ ): for folder in directories: # Ignore private modules if folder.startswith('''_''' ): directories.remove(A__ ) continue # Ignore leftovers from branches (empty folders apart from pycache) if len(list((Path(A__ ) / folder).glob('''*.py''' ) ) ) == 0: continue __lowercase = str((Path(A__ ) / folder).relative_to(A__ ) ) __lowercase = short_path.replace(os.path.sep , '''.''' ) submodules.append(A__ ) for fname in files: if fname == "__init__.py": continue __lowercase = str((Path(A__ ) / fname).relative_to(A__ ) ) __lowercase = short_path.replace('''.py''' , '''''' ).replace(os.path.sep , '''.''' ) if len(submodule.split('''.''' ) ) == 1: submodules.append(A__ ) return submodules lowerCAmelCase__ = [ '''convert_pytorch_checkpoint_to_tf2''', '''modeling_flax_pytorch_utils''', '''models.esm.openfold_utils''', ] def _A ( ): """simple docstring""" from transformers.utils import direct_transformers_import __lowercase = direct_transformers_import(A__ ) __lowercase = set(transformers._import_structure.keys() ) # This contains all the base keys of the _import_structure object defined in the init, but if the user is missing # some optional dependencies, they may not have all of them. Thus we read the init to read all additions and # (potentiall re-) add them. with open(os.path.join(A__ , '''__init__.py''' ) , '''r''' ) as f: __lowercase = f.read() import_structure_keys.update(set(re.findall(R'''import_structure\[\"([^\"]*)\"\]''' , A__ ) ) ) __lowercase = [ module for module in get_transformers_submodules() if module not in IGNORE_SUBMODULES and module not in import_structure_keys ] if len(A__ ) > 0: __lowercase = '''\n'''.join(F"- {module}" for module in module_not_registered ) raise ValueError( '''The following submodules are not properly registed in the main init of Transformers:\n''' F"{list_of_modules}\n" '''Make sure they appear somewhere in the keys of `_import_structure` with an empty list as value.''' ) if __name__ == "__main__": check_all_inits() check_submodules()
719
'''simple docstring''' from collections.abc import Callable import numpy as np def _A ( A__ , A__ , A__ , A__ , A__ ): """simple docstring""" __lowercase = int(np.ceil((x_end - xa) / step_size ) ) __lowercase = np.zeros((n + 1,) ) __lowercase = ya __lowercase = xa for k in range(A__ ): __lowercase = y[k] + step_size * ode_func(A__ , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
624
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) lowerCAmelCase__ = {'''configuration_vit_mae''': ['''VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTMAEConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTMAEForPreTraining''', '''ViTMAELayer''', '''ViTMAEModel''', '''ViTMAEPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ = [ '''TFViTMAEForPreTraining''', '''TFViTMAEModel''', '''TFViTMAEPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit_mae import VIT_MAE_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTMAEConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit_mae import ( VIT_MAE_PRETRAINED_MODEL_ARCHIVE_LIST, ViTMAEForPreTraining, ViTMAELayer, ViTMAEModel, ViTMAEPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit_mae import TFViTMAEForPreTraining, TFViTMAEModel, TFViTMAEPreTrainedModel else: import sys lowerCAmelCase__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
720
'''simple docstring''' def _A ( A__ ): """simple docstring""" if not nums: # Makes sure that the list is not empty raise ValueError('''List is empty''' ) __lowercase = sum(A__ ) / len(A__ ) # Calculate the average return sum(abs(x - average ) for x in nums ) / len(A__ ) if __name__ == "__main__": import doctest doctest.testmod()
624
0