code
stringlengths
82
54.1k
code_codestyle
int64
0
699
style_context
stringlengths
111
35.6k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A = logging.get_logger(__name__) A = { 'google/bit-50': 'https://huggingface.co/google/bit-50/resolve/main/config.json', } class _a ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): __magic_name__ = 'bit' __magic_name__ = ['preactivation', 'bottleneck'] __magic_name__ = ['SAME', 'VALID'] def __init__( self : Optional[Any] , _lowercase : Union[str, Any]=3 , _lowercase : List[str]=64 , _lowercase : Tuple=[256, 512, 1024, 2048] , _lowercase : Optional[Any]=[3, 4, 6, 3] , _lowercase : Optional[Any]="preactivation" , _lowercase : int="relu" , _lowercase : Optional[Any]=None , _lowercase : Tuple=32 , _lowercase : Tuple=0.0 , _lowercase : Optional[int]=False , _lowercase : Union[str, Any]=32 , _lowercase : List[str]=1 , _lowercase : Dict=None , _lowercase : Tuple=None , **_lowercase : Tuple , ) -> List[str]: super().__init__(**__SCREAMING_SNAKE_CASE ) if layer_type not in self.layer_types: raise ValueError(F'''layer_type={layer_type} is not one of {",".join(self.layer_types )}''' ) if global_padding is not None: if global_padding.upper() in self.supported_padding: snake_case : List[str] = global_padding.upper() else: raise ValueError(F'''Padding strategy {global_padding} not supported''' ) snake_case : int = num_channels snake_case : Optional[int] = embedding_size snake_case : Optional[int] = hidden_sizes snake_case : Union[str, Any] = depths snake_case : List[str] = layer_type snake_case : Tuple = hidden_act snake_case : int = global_padding snake_case : Tuple = num_groups snake_case : int = drop_path_rate snake_case : str = embedding_dynamic_padding snake_case : List[str] = output_stride snake_case : Dict = width_factor snake_case : Optional[int] = ["stem"] + [F'''stage{idx}''' for idx in range(1 , len(__SCREAMING_SNAKE_CASE ) + 1 )] snake_case , snake_case : List[str] = get_aligned_output_features_output_indices( out_features=__SCREAMING_SNAKE_CASE , out_indices=__SCREAMING_SNAKE_CASE , stage_names=self.stage_names )
449
from itertools import permutations def lowercase__ ( A_: tuple ) -> bool: """simple docstring""" if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False __UpperCAmelCase =[7, 11, 13, 17] for i, test in enumerate(A_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowercase__ ( A_: int = 10 ) -> int: """simple docstring""" return sum( int("""""".join(map(A_ , A_ ) ) ) for num in permutations(range(A_ ) ) if is_substring_divisible(A_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
68
0
'''simple docstring''' import os def __UpperCamelCase ( ) ->Optional[Any]: with open(os.path.dirname(A_ ) + '''/p022_names.txt''' ) as file: snake_case = str(file.readlines()[0] ) snake_case = names.replace('''\"''' , '''''' ).split(''',''' ) names.sort() snake_case = 0 snake_case = 0 for i, name in enumerate(A_ ): for letter in name: name_score += ord(A_ ) - 64 total_score += (i + 1) * name_score snake_case = 0 return total_score if __name__ == "__main__": print(solution())
342
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar __A = TypeVar("T") def lowercase__ ( A_: int ) -> int: """simple docstring""" return (position - 1) // 2 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 1 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 2 class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[str] ) -> None: __UpperCAmelCase =[] __UpperCAmelCase ={} __UpperCAmelCase =0 def __len__( self : str ) -> int: return self.elements def __repr__( self : Dict ) -> str: return str(self.heap ) def _a ( self : Optional[int] ) -> bool: # Check if the priority queue is empty return self.elements == 0 def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an element with given priority to the queue self.heap.append((elem, weight) ) __UpperCAmelCase =self.elements self.elements += 1 self._bubble_up(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> T: # Remove and return the element with lowest weight (highest priority) if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) __UpperCAmelCase , __UpperCAmelCase =self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: __UpperCAmelCase , __UpperCAmelCase =self.heap[0] self._bubble_down(__SCREAMING_SNAKE_CASE ) return elem def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Update the weight of the given key __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase =(elem, weight) if position > 0: __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._bubble_up(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (upward movement) [to be used internally # only] __UpperCAmelCase =self.position_map[elem] if curr_pos == 0: return None __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_up(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (downward movement) [to be used # internally only] __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase =get_child_left_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =get_child_right_position(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements and child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) else: return None if child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int ) -> None: # Swap the nodes at the given positions __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase , __UpperCAmelCase =( self.heap[nodea_pos], self.heap[nodea_pos], ) __UpperCAmelCase =nodea_pos __UpperCAmelCase =nodea_pos class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[Any] ) -> None: __UpperCAmelCase ={} __UpperCAmelCase =0 def __repr__( self : Tuple ) -> str: return str(self.connections ) def __len__( self : str ) -> int: return self.nodes def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : T ) -> None: # Add a node in the graph if it is not in the graph if node not in self.connections: __UpperCAmelCase ={} self.nodes += 1 def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an edge between 2 nodes in the graph self.add_node(__SCREAMING_SNAKE_CASE ) self.add_node(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =weight __UpperCAmelCase =weight def lowercase__ ( A_: GraphUndirectedWeighted[T] , ) -> tuple[dict[T, int], dict[T, T | None]]: """simple docstring""" __UpperCAmelCase ={node: maxsize for node in graph.connections} __UpperCAmelCase ={node: None for node in graph.connections} __UpperCAmelCase =MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(A_ , A_ ) if priority_queue.is_empty(): return dist, parent # initialization __UpperCAmelCase =priority_queue.extract_min() __UpperCAmelCase =0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node # running prim's algorithm while not priority_queue.is_empty(): __UpperCAmelCase =priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node return dist, parent
68
0
import re from filelock import FileLock try: import nltk __lowerCamelCase : List[str] = True except (ImportError, ModuleNotFoundError): __lowerCamelCase : int = False if NLTK_AVAILABLE: with FileLock(".lock") as lock: nltk.download("punkt", quiet=True) def lowerCamelCase_(lowerCamelCase_ ) -> str: re.sub("<n>" , "" , A_ ) # remove pegasus newline char assert NLTK_AVAILABLE, "nltk must be installed to separate newlines between sentences. (pip install nltk)" return "\n".join(nltk.sent_tokenize(A_ ) )
323
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __A = logging.get_logger(__name__) @dataclass class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[int] = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : Any , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Dict: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase =deprecated_arg[3:] __UpperCAmelCase =not kwargs.pop(__SCREAMING_SNAKE_CASE ) logger.warning( f'''{deprecated_arg} is depreciated. Please use --no-{positive_arg} or''' f''' {positive_arg}={kwargs[positive_arg]}''' ) __UpperCAmelCase =kwargs.pop("""tpu_name""" , self.tpu_name ) __UpperCAmelCase =kwargs.pop("""device_idx""" , self.device_idx ) __UpperCAmelCase =kwargs.pop("""eager_mode""" , self.eager_mode ) __UpperCAmelCase =kwargs.pop("""use_xla""" , self.use_xla ) super().__init__(**__SCREAMING_SNAKE_CASE ) lowerCamelCase : str = field( default=UpperCamelCase , metadata={'help': 'Name of TPU'} , ) lowerCamelCase : int = field( default=0 , metadata={'help': 'CPU / GPU device index. Defaults to 0.'} , ) lowerCamelCase : bool = field(default=UpperCamelCase , metadata={'help': 'Benchmark models in eager model.'} ) lowerCamelCase : bool = field( default=UpperCamelCase , metadata={ 'help': 'Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.' } , ) @cached_property def _a ( self : List[str] ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) __UpperCAmelCase =None if self.tpu: try: if self.tpu_name: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: __UpperCAmelCase =None return tpu @cached_property def _a ( self : Tuple ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) __UpperCAmelCase =tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , """GPU""" ) __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/gpu:{self.device_idx}''' ) else: tf.config.set_visible_devices([] , """GPU""" ) # disable GPU __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/cpu:{self.device_idx}''' ) return strategy @property def _a ( self : Optional[Any] ) -> bool: requires_backends(self , ["""tf"""] ) return self._setup_tpu is not None @property def _a ( self : str ) -> "tf.distribute.Strategy": requires_backends(self , ["""tf"""] ) return self._setup_strategy @property def _a ( self : Dict ) -> Optional[int]: requires_backends(self , ["""tf"""] ) return tf.config.list_physical_devices("""GPU""" ) @property def _a ( self : List[str] ) -> int: requires_backends(self , ["""tf"""] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _a ( self : List[str] ) -> bool: return self.n_gpu > 0
68
0
'''simple docstring''' import datasets import faiss import numpy as np import streamlit as st import torch from elasticsearch import Elasticsearch from elia_utils import ( embed_questions_for_retrieval, make_qa_sas_model, qa_sas_generate, query_es_index, query_qa_dense_index, ) import transformers from transformers import AutoModel, AutoModelForSeqaSeqLM, AutoTokenizer __lowerCAmelCase : Optional[Any] = "bart" __lowerCAmelCase : Union[str, Any] = True @st.cache(allow_output_mutation=A_ ) def lowerCAmelCase ( ): """simple docstring""" if LOAD_DENSE_INDEX: __UpperCAmelCase = AutoTokenizer.from_pretrained('''yjernite/retribert-base-uncased''' ) __UpperCAmelCase = AutoModel.from_pretrained('''yjernite/retribert-base-uncased''' ).to('''cuda:0''' ) __UpperCAmelCase = qar_model.eval() else: __UpperCAmelCase , __UpperCAmelCase = (None, None) if MODEL_TYPE == "bart": __UpperCAmelCase = AutoTokenizer.from_pretrained('''yjernite/bart_eli5''' ) __UpperCAmelCase = AutoModelForSeqaSeqLM.from_pretrained('''yjernite/bart_eli5''' ).to('''cuda:0''' ) __UpperCAmelCase = torch.load('''seq2seq_models/eli5_bart_model_blm_2.pth''' ) sas_model.load_state_dict(save_dict['''model'''] ) __UpperCAmelCase = sas_model.eval() else: __UpperCAmelCase , __UpperCAmelCase = make_qa_sas_model( model_name='''t5-small''' , from_file='''seq2seq_models/eli5_t5_model_1024_4.pth''' , device='''cuda:0''' ) return (qar_tokenizer, qar_model, sas_tokenizer, sas_model) @st.cache(allow_output_mutation=A_ ) def lowerCAmelCase ( ): """simple docstring""" if LOAD_DENSE_INDEX: __UpperCAmelCase = faiss.StandardGpuResources() __UpperCAmelCase = datasets.load_dataset(path='''wiki_snippets''' , name='''wiki40b_en_100_0''' )['''train'''] __UpperCAmelCase = np.memmap( '''wiki40b_passages_reps_32_l-8_h-768_b-512-512.dat''' , dtype='''float32''' , mode='''r''' , shape=(wikiaab_passages.num_rows, 1_2_8) , ) __UpperCAmelCase = faiss.IndexFlatIP(1_2_8 ) __UpperCAmelCase = faiss.index_cpu_to_gpu(A_ , 1 , A_ ) wikiaab_gpu_index_flat.add(A_ ) # TODO fix for larger GPU else: __UpperCAmelCase , __UpperCAmelCase = (None, None) __UpperCAmelCase = Elasticsearch([{'''host''': '''localhost''', '''port''': '''9200'''}] ) return (wikiaab_passages, wikiaab_gpu_index_flat, es_client) @st.cache(allow_output_mutation=A_ ) def lowerCAmelCase ( ): """simple docstring""" __UpperCAmelCase = datasets.load_dataset('''eli5''' , name='''LFQA_reddit''' ) __UpperCAmelCase = elia['''train_eli5'''] __UpperCAmelCase = np.memmap( '''eli5_questions_reps.dat''' , dtype='''float32''' , mode='''r''' , shape=(elia_train.num_rows, 1_2_8) ) __UpperCAmelCase = faiss.IndexFlatIP(1_2_8 ) eli5_train_q_index.add(A_ ) return (elia_train, eli5_train_q_index) __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : str = load_indexes() __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase : List[Any] = load_models() __lowerCAmelCase , __lowerCAmelCase : str = load_train_data() def lowerCAmelCase ( UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Optional[Any]=1_0 ): """simple docstring""" __UpperCAmelCase = embed_questions_for_retrieval([question] , A_ , A_ ) __UpperCAmelCase , __UpperCAmelCase = eli5_train_q_index.search(A_ , A_ ) __UpperCAmelCase = [elia_train[int(A_ )] for i in I[0]] return nn_examples def lowerCAmelCase ( UpperCamelCase__ : Union[str, Any] , UpperCamelCase__ : Any="wiki40b" , UpperCamelCase__ : Optional[int]="dense" , UpperCamelCase__ : Any=1_0 ): """simple docstring""" if source == "none": __UpperCAmelCase , __UpperCAmelCase = (''' <P> '''.join(['''''' for _ in range(1_1 )] ).strip(), []) else: if method == "dense": __UpperCAmelCase , __UpperCAmelCase = query_qa_dense_index( A_ , A_ , A_ , A_ , A_ , A_ ) else: __UpperCAmelCase , __UpperCAmelCase = query_es_index( A_ , A_ , index_name='''english_wiki40b_snippets_100w''' , n_results=A_ , ) __UpperCAmelCase = [ (res['''article_title'''], res['''section_title'''].strip(), res['''score'''], res['''passage_text''']) for res in hit_lst ] __UpperCAmelCase = '''question: {} context: {}'''.format(A_ , A_ ) return question_doc, support_list @st.cache( hash_funcs={ torch.Tensor: (lambda UpperCamelCase__ : None), transformers.models.bart.tokenization_bart.BartTokenizer: (lambda UpperCamelCase__ : None), } ) def lowerCAmelCase ( UpperCamelCase__ : Dict , UpperCamelCase__ : List[Any] , UpperCamelCase__ : Optional[Any] , UpperCamelCase__ : Tuple=6_4 , UpperCamelCase__ : int=2_5_6 , UpperCamelCase__ : Dict=False , UpperCamelCase__ : Dict=2 , UpperCamelCase__ : Dict=0.95 , UpperCamelCase__ : Optional[int]=0.8 ): """simple docstring""" with torch.no_grad(): __UpperCAmelCase = qa_sas_generate( A_ , A_ , A_ , num_answers=1 , num_beams=A_ , min_len=A_ , max_len=A_ , do_sample=A_ , temp=A_ , top_p=A_ , top_k=A_ , max_input_length=1_0_2_4 , device='''cuda:0''' , )[0] return (answer, support_list) st.title("Long Form Question Answering with ELI5") # Start sidebar __lowerCAmelCase : List[str] = "<img src='https://huggingface.co/front/assets/huggingface_logo.svg'>" __lowerCAmelCase : Union[str, Any] = "\n<html>\n <head>\n <style>\n .img-container {\n padding-left: 90px;\n padding-right: 90px;\n padding-top: 50px;\n padding-bottom: 50px;\n background-color: #f0f3f9;\n }\n </style>\n </head>\n <body>\n <span class=\"img-container\"> <!-- Inline parent element -->\n %s\n </span>\n </body>\n</html>\n" % ( header_html, ) st.sidebar.markdown( header_full, unsafe_allow_html=True, ) # Long Form QA with ELI5 and Wikipedia __lowerCAmelCase : Union[str, Any] = "\nThis demo presents a model trained to [provide long-form answers to open-domain questions](https://yjernite.github.io/lfqa.html).\nFirst, a document retriever fetches a set of relevant Wikipedia passages given the question from the [Wiki40b](https://research.google/pubs/pub49029/) dataset,\na pre-processed fixed snapshot of Wikipedia.\n" st.sidebar.markdown(description, unsafe_allow_html=True) __lowerCAmelCase : Tuple = [ "Answer the question", "View the retrieved document only", "View the most similar ELI5 question and answer", "Show me everything, please!", ] __lowerCAmelCase : int = st.sidebar.checkbox("Demo options") if demo_options: __lowerCAmelCase : Any = st.sidebar.selectbox( "", action_list, index=3, ) __lowerCAmelCase : Tuple = action_list.index(action_st) __lowerCAmelCase : Dict = st.sidebar.selectbox( "", ["Show full text of passages", "Show passage section titles"], index=0, ) __lowerCAmelCase : List[Any] = show_type == "Show full text of passages" else: __lowerCAmelCase : Tuple = 3 __lowerCAmelCase : Dict = True __lowerCAmelCase : Any = st.sidebar.checkbox("Retrieval options") if retrieval_options: __lowerCAmelCase : Union[str, Any] = "\n ### Information retriever options\n\n The **sparse** retriever uses ElasticSearch, while the **dense** retriever uses max-inner-product search between a question and passage embedding\n trained using the [ELI5](https://arxiv.org/abs/1907.09190) questions-answer pairs.\n The answer is then generated by sequence to sequence model which takes the question and retrieved document as input.\n " st.sidebar.markdown(retriever_info) __lowerCAmelCase : Dict = st.sidebar.selectbox("Which Wikipedia format should the model use?", ["wiki40b", "none"]) __lowerCAmelCase : str = st.sidebar.selectbox("Which Wikipedia indexer should the model use?", ["dense", "sparse", "mixed"]) else: __lowerCAmelCase : Optional[Any] = "wiki40b" __lowerCAmelCase : Any = "dense" __lowerCAmelCase : List[Any] = "beam" __lowerCAmelCase : Union[str, Any] = 2 __lowerCAmelCase : Dict = 64 __lowerCAmelCase : Union[str, Any] = 256 __lowerCAmelCase : int = None __lowerCAmelCase : int = None __lowerCAmelCase : Tuple = st.sidebar.checkbox("Generation options") if generate_options: __lowerCAmelCase : Dict = "\n ### Answer generation options\n\n The sequence-to-sequence model was initialized with [BART](https://huggingface.co/facebook/bart-large)\n weights and fine-tuned on the ELI5 QA pairs and retrieved documents. You can use the model for greedy decoding with\n **beam** search, or **sample** from the decoder's output probabilities.\n " st.sidebar.markdown(generate_info) __lowerCAmelCase : Union[str, Any] = st.sidebar.selectbox("Would you like to use beam search or sample an answer?", ["beam", "sampled"]) __lowerCAmelCase : List[str] = st.sidebar.slider( "Minimum generation length", min_value=8, max_value=256, value=64, step=8, format=None, key=None ) __lowerCAmelCase : int = st.sidebar.slider( "Maximum generation length", min_value=64, max_value=512, value=256, step=16, format=None, key=None ) if sampled == "beam": __lowerCAmelCase : str = st.sidebar.slider("Beam size", min_value=1, max_value=8, value=2, step=None, format=None, key=None) else: __lowerCAmelCase : List[str] = st.sidebar.slider( "Nucleus sampling p", min_value=0.1, max_value=1.0, value=0.9_5, step=0.0_1, format=None, key=None ) __lowerCAmelCase : Union[str, Any] = st.sidebar.slider( "Temperature", min_value=0.1, max_value=1.0, value=0.7, step=0.0_1, format=None, key=None ) __lowerCAmelCase : Optional[Any] = None # start main text __lowerCAmelCase : List[str] = [ "<MY QUESTION>", "How do people make chocolate?", "Why do we get a fever when we are sick?", "How can different animals perceive different colors?", "What is natural language processing?", "What's the best way to treat a sunburn?", "What exactly are vitamins ?", "How does nuclear energy provide electricity?", "What's the difference between viruses and bacteria?", "Why are flutes classified as woodwinds when most of them are made out of metal ?", "Why do people like drinking coffee even though it tastes so bad?", "What happens when wine ages? How does it make the wine taste better?", "If an animal is an herbivore, where does it get the protein that it needs to survive if it only eats grass?", "How can we set a date to the beginning or end of an artistic period? Doesn't the change happen gradually?", "How does New Zealand have so many large bird predators?", ] __lowerCAmelCase : List[Any] = st.selectbox( "What would you like to ask? ---- select <MY QUESTION> to enter a new query", questions_list, index=1, ) if question_s == "<MY QUESTION>": __lowerCAmelCase : Dict = st.text_input("Enter your question here:", "") else: __lowerCAmelCase : Dict = question_s if st.button("Show me!"): if action in [0, 1, 3]: if index_type == "mixed": __lowerCAmelCase , __lowerCAmelCase : Union[str, Any] = make_support(question, source=wiki_source, method="dense", n_results=10) __lowerCAmelCase , __lowerCAmelCase : Optional[int] = make_support(question, source=wiki_source, method="sparse", n_results=10) __lowerCAmelCase : List[str] = [] for res_d, res_s in zip(support_list_dense, support_list_sparse): if tuple(res_d) not in support_list: support_list += [tuple(res_d)] if tuple(res_s) not in support_list: support_list += [tuple(res_s)] __lowerCAmelCase : List[str] = support_list[:10] __lowerCAmelCase : Union[str, Any] = "<P> " + " <P> ".join([res[-1] for res in support_list]) else: __lowerCAmelCase , __lowerCAmelCase : Any = make_support(question, source=wiki_source, method=index_type, n_results=10) if action in [0, 3]: __lowerCAmelCase , __lowerCAmelCase : str = answer_question( question_doc, sas_model, sas_tokenizer, min_len=min_len, max_len=int(max_len), sampling=(sampled == "sampled"), n_beams=n_beams, top_p=top_p, temp=temp, ) st.markdown("### The model generated answer is:") st.write(answer) if action in [0, 1, 3] and wiki_source != "none": st.markdown("--- \n ### The model is drawing information from the following Wikipedia passages:") for i, res in enumerate(support_list): __lowerCAmelCase : Union[str, Any] = "https://en.wikipedia.org/wiki/{}".format(res[0].replace(" ", "_")) __lowerCAmelCase : List[str] = res[1].strip() if sec_titles == "": __lowerCAmelCase : Optional[Any] = "[{}]({})".format(res[0], wiki_url) else: __lowerCAmelCase : Any = sec_titles.split(" & ") __lowerCAmelCase : List[str] = " & ".join( ["[{}]({}#{})".format(sec.strip(), wiki_url, sec.strip().replace(" ", "_")) for sec in sec_list] ) st.markdown( "{0:02d} - **Article**: {1:<18} <br> _Section_: {2}".format(i + 1, res[0], sections), unsafe_allow_html=True, ) if show_passages: st.write( "> <span style=\"font-family:arial; font-size:10pt;\">" + res[-1] + "</span>", unsafe_allow_html=True ) if action in [2, 3]: __lowerCAmelCase : Tuple = find_nearest_training(question) __lowerCAmelCase : int = nn_train_list[0] st.markdown( "--- \n ### The most similar question in the ELI5 training set was: \n\n {}".format(train_exple["title"]) ) __lowerCAmelCase : List[str] = [ "{}. {}".format(i + 1, " \n".join([line.strip() for line in ans.split("\n") if line.strip() != ""])) for i, (ans, sc) in enumerate(zip(train_exple["answers"]["text"], train_exple["answers"]["score"])) if i == 0 or sc > 2 ] st.markdown("##### Its answers were: \n\n {}".format("\n".join(answers_st))) __lowerCAmelCase : List[Any] = "\n---\n\n**Disclaimer**\n\n*The intent of this app is to provide some (hopefully entertaining) insights into the behavior of a current LFQA system.\nEvaluating biases of such a model and ensuring factual generations are still very much open research problems.\nTherefore, until some significant progress is achieved, we caution against using the generated answers for practical purposes.*\n" st.sidebar.markdown(disclaimer, unsafe_allow_html=True)
262
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Dict: torch.manual_seed(0 ) __UpperCAmelCase =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def _a ( self : int ) -> Union[str, Any]: __UpperCAmelCase =self.dummy_uncond_unet __UpperCAmelCase =ScoreSdeVeScheduler() __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE )[ 0 ] __UpperCAmelCase =image[0, -3:, -3:, -1] __UpperCAmelCase =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCAmelCase =np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : Optional[int] ) -> int: __UpperCAmelCase ="""google/ncsnpp-church-256""" __UpperCAmelCase =UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVeScheduler.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=10 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCAmelCase =np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
68
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: __SCREAMING_SNAKE_CASE =None __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) __SCREAMING_SNAKE_CASE ={"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} __SCREAMING_SNAKE_CASE ={ "vocab_file": { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model", }, "tokenizer_file": { "xlnet-base-cased": "https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json", "xlnet-large-cased": "https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json", }, } __SCREAMING_SNAKE_CASE ={ "xlnet-base-cased": None, "xlnet-large-cased": None, } __SCREAMING_SNAKE_CASE ="▁" # Segments (not really needed) __SCREAMING_SNAKE_CASE =0 __SCREAMING_SNAKE_CASE =1 __SCREAMING_SNAKE_CASE =2 __SCREAMING_SNAKE_CASE =3 __SCREAMING_SNAKE_CASE =4 class UpperCamelCase ( lowercase_ ): lowercase = VOCAB_FILES_NAMES lowercase = PRETRAINED_VOCAB_FILES_MAP lowercase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowercase = 'left' lowercase = XLNetTokenizer def __init__( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=False ,__UpperCamelCase=True ,__UpperCamelCase=False ,__UpperCamelCase="<s>" ,__UpperCamelCase="</s>" ,__UpperCamelCase="<unk>" ,__UpperCamelCase="<sep>" ,__UpperCamelCase="<pad>" ,__UpperCamelCase="<cls>" ,__UpperCamelCase="<mask>" ,__UpperCamelCase=["<eop>", "<eod>"] ,**__UpperCamelCase ,) -> List[Any]: '''simple docstring''' lowercase_ : List[Any] = AddedToken(__SCREAMING_SNAKE_CASE ,lstrip=__SCREAMING_SNAKE_CASE ,rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) else mask_token super().__init__( vocab_file=__SCREAMING_SNAKE_CASE ,tokenizer_file=__SCREAMING_SNAKE_CASE ,do_lower_case=__SCREAMING_SNAKE_CASE ,remove_space=__SCREAMING_SNAKE_CASE ,keep_accents=__SCREAMING_SNAKE_CASE ,bos_token=__SCREAMING_SNAKE_CASE ,eos_token=__SCREAMING_SNAKE_CASE ,unk_token=__SCREAMING_SNAKE_CASE ,sep_token=__SCREAMING_SNAKE_CASE ,pad_token=__SCREAMING_SNAKE_CASE ,cls_token=__SCREAMING_SNAKE_CASE ,mask_token=__SCREAMING_SNAKE_CASE ,additional_special_tokens=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ,) lowercase_ : Optional[Any] = 3 lowercase_ : Optional[int] = do_lower_case lowercase_ : Tuple = remove_space lowercase_ : int = keep_accents lowercase_ : Union[str, Any] = vocab_file lowercase_ : str = False if not self.vocab_file else True def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase = None ) -> List[int]: '''simple docstring''' lowercase_ : List[str] = [self.sep_token_id] lowercase_ : List[Any] = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase = None ) -> List[int]: '''simple docstring''' lowercase_ : List[Any] = [self.sep_token_id] lowercase_ : str = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( 'Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ' 'tokenizer.' ) if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return lowercase_ : Any = os.path.join( __SCREAMING_SNAKE_CASE ,(filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file ,__SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
425
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __A = logging.get_logger(__name__) __A = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, "constant": get_constant_schedule, "constant_w_warmup": get_constant_schedule_with_warmup, } class _A ( UpperCamelCase ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : str=None , *__SCREAMING_SNAKE_CASE : Union[str, Any] , **__SCREAMING_SNAKE_CASE : List[Any] ) -> Any: super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if config is None: assert isinstance(self.model , __SCREAMING_SNAKE_CASE ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f''' {self.model.__class__}''' ) __UpperCAmelCase =self.model.config else: __UpperCAmelCase =config __UpperCAmelCase =data_args __UpperCAmelCase =self.config.tgt_vocab_size if isinstance(self.config , __SCREAMING_SNAKE_CASE ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' """ padding..""" ) if self.args.label_smoothing == 0: __UpperCAmelCase =torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss __UpperCAmelCase =label_smoothed_nll_loss def _a ( self : Any , __SCREAMING_SNAKE_CASE : int ) -> Any: if self.optimizer is None: __UpperCAmelCase =["""bias""", """LayerNorm.weight"""] __UpperCAmelCase =[ { """params""": [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], """weight_decay""": self.args.weight_decay, }, { """params""": [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] __UpperCAmelCase =Adafactor if self.args.adafactor else AdamW if self.args.adafactor: __UpperCAmelCase =Adafactor __UpperCAmelCase ={"""scale_parameter""": False, """relative_step""": False} else: __UpperCAmelCase =AdamW __UpperCAmelCase ={ """betas""": (self.args.adam_betaa, self.args.adam_betaa), """eps""": self.args.adam_epsilon, } __UpperCAmelCase =self.args.learning_rate if self.sharded_ddp: __UpperCAmelCase =OSS( params=__SCREAMING_SNAKE_CASE , optim=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) else: __UpperCAmelCase =optimizer_cls(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if self.lr_scheduler is None: __UpperCAmelCase =self._get_lr_scheduler(__SCREAMING_SNAKE_CASE ) else: # ignoring --lr_scheduler logger.warning("""scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.""" ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[int] ) -> Any: __UpperCAmelCase =arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": __UpperCAmelCase =schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": __UpperCAmelCase =schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: __UpperCAmelCase =schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__SCREAMING_SNAKE_CASE ) return scheduler def _a ( self : Optional[Any] ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Tuple: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models __UpperCAmelCase , __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[:2] else: # compute label smoothed loss __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =torch.nn.functional.log_softmax(__SCREAMING_SNAKE_CASE , dim=-1 ) __UpperCAmelCase , __UpperCAmelCase =self.loss_fn(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] ) -> Dict: __UpperCAmelCase =inputs.pop("""labels""" ) __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return loss def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : nn.Module , __SCREAMING_SNAKE_CASE : Dict[str, Union[torch.Tensor, Any]] , __SCREAMING_SNAKE_CASE : bool , __SCREAMING_SNAKE_CASE : Optional[List[str]] = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: __UpperCAmelCase =self._prepare_inputs(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """max_length""": self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, """num_beams""": self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: __UpperCAmelCase =self.model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , **__SCREAMING_SNAKE_CASE , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) __UpperCAmelCase =inputs.pop("""labels""" ) with torch.no_grad(): # compute loss on predict data __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) __UpperCAmelCase =generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) return (loss, logits, labels) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : int ) -> List[Any]: # If PAD token is not defined at least EOS token has to be defined __UpperCAmelCase =self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( """Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be""" f''' padded to `max_length`={max_length}''' ) __UpperCAmelCase =pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) __UpperCAmelCase =tensor return padded_tensor
68
0
"""simple docstring""" from .imports import is_rich_available if is_rich_available(): from rich.traceback import install install(show_locals=False) else: raise ModuleNotFoundError('''To use the rich extension, install rich with `pip install rich`''')
115
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : List[Any] = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str=0 ) -> Any: __UpperCAmelCase =floats_tensor((1, 3, 128, 128) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =np.random.RandomState(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """strength""": 0.75, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _a ( self : Optional[Any] ) -> int: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Optional[Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) # warmup pass to apply optimizations __UpperCAmelCase =pipe(**self.get_dummy_inputs() ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Optional[int]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _a ( self : Dict ) -> int: __UpperCAmelCase =ort.SessionOptions() __UpperCAmelCase =False return options def _a ( self : Dict ) -> Any: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) # using the PNDM scheduler by default __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def _a ( self : List[str] ) -> str: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) __UpperCAmelCase =LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
68
0
import argparse import torch from transformers import LxmertConfig, LxmertForPreTraining, load_tf_weights_in_lxmert from transformers.utils import logging logging.set_verbosity_info() def lowerCamelCase_ ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : Tuple ) -> Tuple: '''simple docstring''' A = LxmertConfig.from_json_file(A_ ) print(F'''Building PyTorch model from configuration: {config}''' ) A = LxmertForPreTraining(A_ ) # Load weights from tf checkpoint load_tf_weights_in_lxmert(A_ , A_ , A_ ) # Save pytorch-model print(F'''Save PyTorch model to {pytorch_dump_path}''' ) torch.save(model.state_dict() , A_ ) if __name__ == "__main__": __snake_case :Dict =argparse.ArgumentParser() # Required parameters parser.add_argument( '--tf_checkpoint_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help='The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.', ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) __snake_case :Optional[int] =parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
106
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __A = logging.getLogger(__name__) class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[Any] = 'sequence-classification' def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Tuple ) -> Optional[Any]: if type(__SCREAMING_SNAKE_CASE ) == dict: __UpperCAmelCase =Namespace(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =glue_output_modes[hparams.task] __UpperCAmelCase =glue_tasks_num_labels[hparams.task] super().__init__(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.mode ) def _a ( self : str , **__SCREAMING_SNAKE_CASE : Dict ) -> List[str]: return self.model(**__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Dict ) -> List[Any]: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =outputs[0] __UpperCAmelCase =self.trainer.lr_schedulers[0]["""scheduler"""] __UpperCAmelCase ={"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def _a ( self : Tuple ) -> List[Any]: __UpperCAmelCase =self.hparams __UpperCAmelCase =processors[args.task]() __UpperCAmelCase =processor.get_labels() for mode in ["train", "dev"]: __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) if os.path.exists(__SCREAMING_SNAKE_CASE ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) __UpperCAmelCase =( processor.get_dev_examples(args.data_dir ) if mode == """dev""" else processor.get_train_examples(args.data_dir ) ) __UpperCAmelCase =convert_examples_to_features( __SCREAMING_SNAKE_CASE , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , __SCREAMING_SNAKE_CASE ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : bool = False ) -> DataLoader: __UpperCAmelCase ="""dev""" if mode == """test""" else mode __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.load(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([f.input_ids for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , batch_size=__SCREAMING_SNAKE_CASE , shuffle=__SCREAMING_SNAKE_CASE , ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : int ) -> str: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =outputs[:2] __UpperCAmelCase =logits.detach().cpu().numpy() __UpperCAmelCase =inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Any ) -> tuple: __UpperCAmelCase =torch.stack([x["""val_loss"""] for x in outputs] ).mean().detach().cpu().item() __UpperCAmelCase =np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =np.argmax(__SCREAMING_SNAKE_CASE , axis=1 ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =np.squeeze(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =np.concatenate([x["""target"""] for x in outputs] , axis=0 ) __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase ={**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )} __UpperCAmelCase =dict(results.items() ) __UpperCAmelCase =results return ret, preds_list, out_label_list def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : list ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _a ( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[Any]: BaseTransformer.add_model_specific_args(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) parser.add_argument( """--max_seq_length""" , default=128 , type=__SCREAMING_SNAKE_CASE , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=__SCREAMING_SNAKE_CASE , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser def lowercase__ ( ) -> str: """simple docstring""" __UpperCAmelCase =argparse.ArgumentParser() add_generic_args(A_ , os.getcwd() ) __UpperCAmelCase =GLUETransformer.add_model_specific_args(A_ , os.getcwd() ) __UpperCAmelCase =parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: __UpperCAmelCase =os.path.join( """./results""" , F'''{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}''' , ) os.makedirs(args.output_dir ) __UpperCAmelCase =GLUETransformer(A_ ) __UpperCAmelCase =generic_train(A_ , A_ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: __UpperCAmelCase =sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=A_ ) ) __UpperCAmelCase =model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(A_ ) if __name__ == "__main__": main()
68
0
"""simple docstring""" from manim import * class __A ( A_ ): '''simple docstring''' def UpperCAmelCase ( self : int ) -> Dict: """simple docstring""" lowercase__ : Tuple = Rectangle(height=0.5 ,width=0.5 ) lowercase__ : Dict = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0 ) lowercase__ : List[str] = Rectangle(height=0.25 ,width=0.25 ) lowercase__ : Dict = [mem.copy() for i in range(6 )] lowercase__ : Optional[int] = [mem.copy() for i in range(6 )] lowercase__ : Optional[int] = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : Any = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : int = VGroup(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : Union[str, Any] = Text('''CPU''' ,font_size=24 ) lowercase__ : Dict = Group(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=__SCREAMING_SNAKE_CASE ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ : int = [mem.copy() for i in range(4 )] lowercase__ : List[str] = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : str = Text('''GPU''' ,font_size=24 ) lowercase__ : Dict = Group(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=__SCREAMING_SNAKE_CASE ) gpu.move_to([-1, -1, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ : Union[str, Any] = [mem.copy() for i in range(6 )] lowercase__ : Optional[Any] = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : Optional[int] = Text('''Model''' ,font_size=24 ) lowercase__ : int = Group(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=__SCREAMING_SNAKE_CASE ) model.move_to([3, -1.0, 0] ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ : Tuple = [] lowercase__ : Union[str, Any] = [] for i, rect in enumerate(__SCREAMING_SNAKE_CASE ): lowercase__ : int = fill.copy().set_fill(__SCREAMING_SNAKE_CASE ,opacity=0.8 ) target.move_to(__SCREAMING_SNAKE_CASE ) model_arr.append(__SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = Rectangle(height=0.46 ,width=0.46 ).set_stroke(width=0.0 ).set_fill(__SCREAMING_SNAKE_CASE ,opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__SCREAMING_SNAKE_CASE ) self.add(*__SCREAMING_SNAKE_CASE ,*__SCREAMING_SNAKE_CASE ) lowercase__ : Dict = [meta_mem.copy() for i in range(6 )] lowercase__ : str = [meta_mem.copy() for i in range(6 )] lowercase__ : Tuple = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : int = VGroup(*__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : Any = VGroup(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0 ) lowercase__ : List[str] = Text('''Disk''' ,font_size=24 ) lowercase__ : List[str] = Group(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ).arrange(__SCREAMING_SNAKE_CASE ,buff=0.5 ,aligned_edge=__SCREAMING_SNAKE_CASE ) disk.move_to([-4, -1.25, 0] ) self.add(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) lowercase__ : Any = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) lowercase__ : str = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor=\'{YELLOW}\'>●</span> Empty Model""" ,font_size=18 ,) key_text.move_to([-5, 2.4, 0] ) self.add(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) lowercase__ : Optional[int] = MarkupText( f"""<span fgcolor=\'{BLUE}\'>●</span> Checkpoint""" ,font_size=18 ,) blue_text.next_to(__SCREAMING_SNAKE_CASE ,DOWN * 2.4 ,aligned_edge=key_text.get_left() ) self.add(__SCREAMING_SNAKE_CASE ) lowercase__ : str = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__SCREAMING_SNAKE_CASE ) ) lowercase__ : Optional[Any] = Square(0.3 ) input.set_fill(__SCREAMING_SNAKE_CASE ,opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] ,__SCREAMING_SNAKE_CASE ,buff=0.5 ) self.play(Write(__SCREAMING_SNAKE_CASE ) ) input.generate_target() input.target.next_to(model_arr[0] ,direction=__SCREAMING_SNAKE_CASE ,buff=0.02 ) self.play(MoveToTarget(__SCREAMING_SNAKE_CASE ) ) self.play(FadeOut(__SCREAMING_SNAKE_CASE ) ) lowercase__ : List[Any] = Arrow(start=__SCREAMING_SNAKE_CASE ,end=__SCREAMING_SNAKE_CASE ,color=__SCREAMING_SNAKE_CASE ,buff=0.5 ) a.next_to(model_arr[0].get_left() ,__SCREAMING_SNAKE_CASE ,buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) lowercase__ : Any = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" ,font_size=24 ,) step_a.move_to([2, 2, 0] ) self.play(Write(__SCREAMING_SNAKE_CASE ,run_time=3 ) ) lowercase__ : Optional[int] = {'''run_time''': 1, '''fade_in''': True, '''fade_out''': True, '''buff''': 0.02} self.play( Write(__SCREAMING_SNAKE_CASE ) ,Circumscribe(model_arr[0] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(model_cpu_arr[0] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(gpu_rect[0] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,) self.play(MoveToTarget(model_cpu_arr[0] ) ) lowercase__ : Optional[int] = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 ,__SCREAMING_SNAKE_CASE ,buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) lowercase__ : Optional[Any] = AnimationGroup( FadeOut(__SCREAMING_SNAKE_CASE ,run_time=0.5 ) ,MoveToTarget(__SCREAMING_SNAKE_CASE ,run_time=0.5 ) ,FadeIn(__SCREAMING_SNAKE_CASE ,run_time=0.5 ) ,lag_ratio=0.2 ) self.play(__SCREAMING_SNAKE_CASE ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: lowercase__ : Union[str, Any] = 0.7 self.play( Circumscribe(model_arr[i] ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(cpu_left_col_base[i] ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(cpu_left_col_base[i + 1] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(gpu_rect[0] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(model_arr[i + 1] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) ,MoveToTarget(model_cpu_arr[i + 1] ) ,) else: self.play( MoveToTarget(model_cpu_arr[i] ,run_time=0.7 ) ,MoveToTarget(model_cpu_arr[i + 1] ,run_time=0.7 ) ,) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() ,RIGHT + 0.02 ,buff=0.2 ) self.play( Circumscribe(model_arr[-1] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(cpu_left_col_base[-1] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,Circumscribe(gpu_rect[0] ,color=__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) ,) self.play(MoveToTarget(model_cpu_arr[i] ) ) lowercase__ : List[Any] = a_c lowercase__ : int = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] ,RIGHT + 0.02 ,buff=0.5 ) self.play( FadeOut(__SCREAMING_SNAKE_CASE ) ,FadeOut(__SCREAMING_SNAKE_CASE ,run_time=0.5 ) ,) lowercase__ : List[str] = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" ,font_size=24 ) step_a.move_to([2, 2, 0] ) self.play(Write(__SCREAMING_SNAKE_CASE ,run_time=3 ) ,MoveToTarget(__SCREAMING_SNAKE_CASE ) ) self.wait()
560
def lowercase__ ( A_: int , A_: int ) -> int: """simple docstring""" return 1 if input_a == input_a else 0 def lowercase__ ( ) -> None: """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
68
0
'''simple docstring''' def __snake_case ( SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , SCREAMING_SNAKE_CASE_ : float , ) -> float: """simple docstring""" UpperCAmelCase = [redshift, radiation_density, matter_density, dark_energy] if any(p < 0 for p in parameters ): raise ValueError('''All input parameters must be positive''' ) if any(p > 1 for p in parameters[1:4] ): raise ValueError('''Relative densities cannot be greater than one''' ) else: UpperCAmelCase = 1 - (matter_density + radiation_density + dark_energy) UpperCAmelCase = ( radiation_density * (redshift + 1) ** 4 + matter_density * (redshift + 1) ** 3 + curvature * (redshift + 1) ** 2 + dark_energy ) UpperCAmelCase = hubble_constant * e_a ** (1 / 2) return hubble if __name__ == "__main__": import doctest # run doctest doctest.testmod() # demo LCDM approximation a__ : int = 0.3 print( hubble_parameter( hubble_constant=68.3, radiation_density=1e-4, matter_density=matter_density, dark_energy=1 - matter_density, redshift=0, ) )
51
from __future__ import annotations import bisect def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] < item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] <= item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_left(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_right(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =0 __UpperCAmelCase =len(A_ ) - 1 while left <= right: __UpperCAmelCase =left + (right - left) // 2 __UpperCAmelCase =sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: __UpperCAmelCase =midpoint - 1 else: __UpperCAmelCase =midpoint + 1 return None def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =bisect.bisect_left(A_ , A_ ) if index != len(A_ ) and sorted_collection[index] == item: return index return None def lowercase__ ( A_: list[int] , A_: int , A_: int , A_: int ) -> int | None: """simple docstring""" if right < left: return None __UpperCAmelCase =left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(A_ , A_ , A_ , midpoint - 1 ) else: return binary_search_by_recursion(A_ , A_ , midpoint + 1 , A_ ) if __name__ == "__main__": __A = input("Enter numbers separated by comma:\n").strip() __A = sorted(int(item) for item in user_input.split(",")) __A = int(input("Enter a single number to be found in the list:\n")) __A = binary_search(collection, target) if result is None: print(F"""{target} was not found in {collection}.""") else: print(F"""{target} was found at position {result} in {collection}.""")
68
0
"""simple docstring""" import inspect import unittest import numpy as np from transformers import BeitConfig from transformers.testing_utils import require_flax, require_vision, slow from transformers.utils import cached_property, is_flax_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor if is_flax_available(): import jax from transformers import FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling, FlaxBeitModel if is_vision_available(): from PIL import Image from transformers import BeitImageProcessor class _lowerCAmelCase ( unittest.TestCase ): def __init__( self , a_ , a_=100 , a_=13 , a_=30 , a_=2 , a_=3 , a_=True , a_=True , a_=32 , a_=5 , a_=4 , a_=37 , a_="gelu" , a_=0.1 , a_=0.1 , a_=10 , a_=0.02 , a_=3 , ) -> Optional[Any]: _UpperCAmelCase = parent _UpperCAmelCase = vocab_size _UpperCAmelCase = batch_size _UpperCAmelCase = image_size _UpperCAmelCase = patch_size _UpperCAmelCase = num_channels _UpperCAmelCase = is_training _UpperCAmelCase = use_labels _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_act _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = type_sequence_label_size _UpperCAmelCase = initializer_range # in BeiT, the seq length equals the number of patches + 1 (we add 1 for the [CLS] token) _UpperCAmelCase = (image_size // patch_size) ** 2 _UpperCAmelCase = num_patches + 1 def _a ( self ) -> Optional[Any]: _UpperCAmelCase = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _UpperCAmelCase = None if self.use_labels: _UpperCAmelCase = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _UpperCAmelCase = BeitConfig( vocab_size=self.vocab_size , image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, pixel_values, labels def _a ( self , a_ , a_ , a_ ) -> List[Any]: _UpperCAmelCase = FlaxBeitModel(config=__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _a ( self , a_ , a_ , a_ ) -> List[Any]: _UpperCAmelCase = FlaxBeitForMaskedImageModeling(config=__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length - 1, self.vocab_size) ) def _a ( self , a_ , a_ , a_ ) -> List[str]: _UpperCAmelCase = self.type_sequence_label_size _UpperCAmelCase = FlaxBeitForImageClassification(config=__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _UpperCAmelCase = 1 _UpperCAmelCase = FlaxBeitForImageClassification(__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _UpperCAmelCase = model(__SCREAMING_SNAKE_CASE ) def _a ( self ) -> Union[str, Any]: _UpperCAmelCase = self.prepare_config_and_inputs() ( ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ( _UpperCAmelCase ) , ) = config_and_inputs _UpperCAmelCase = {"pixel_values": pixel_values} return config, inputs_dict @require_flax class _lowerCAmelCase ( lowerCamelCase , unittest.TestCase ): lowercase_ : List[str] = ( (FlaxBeitModel, FlaxBeitForImageClassification, FlaxBeitForMaskedImageModeling) if is_flax_available() else () ) def _a ( self ) -> None: _UpperCAmelCase = FlaxBeitModelTester(self ) _UpperCAmelCase = ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , has_text_modality=__SCREAMING_SNAKE_CASE , hidden_size=37 ) def _a ( self ) -> List[str]: self.config_tester.run_common_tests() def _a ( self ) -> Any: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _UpperCAmelCase = model_class(__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _UpperCAmelCase = [*signature.parameters.keys()] _UpperCAmelCase = ["pixel_values"] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE ) def _a ( self ) -> List[str]: _UpperCAmelCase , _UpperCAmelCase = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): _UpperCAmelCase = self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) _UpperCAmelCase = model_class(__SCREAMING_SNAKE_CASE ) @jax.jit def model_jitted(a_ , **a_ ): return model(pixel_values=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) with self.subTest("JIT Enabled" ): _UpperCAmelCase = model_jitted(**__SCREAMING_SNAKE_CASE ).to_tuple() with self.subTest("JIT Disabled" ): with jax.disable_jit(): _UpperCAmelCase = model_jitted(**__SCREAMING_SNAKE_CASE ).to_tuple() self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , len(__SCREAMING_SNAKE_CASE ) ) for jitted_output, output in zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): self.assertEqual(jitted_output.shape , output.shape ) def _a ( self ) -> Optional[int]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) def _a ( self ) -> Optional[int]: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__SCREAMING_SNAKE_CASE ) def _a ( self ) -> Dict: _UpperCAmelCase = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__SCREAMING_SNAKE_CASE ) @slow def _a ( self ) -> Optional[int]: for model_class_name in self.all_model_classes: _UpperCAmelCase = model_class_name.from_pretrained("microsoft/beit-base-patch16-224" ) _UpperCAmelCase = model(np.ones((1, 3, 224, 224) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_vision @require_flax class _lowerCAmelCase ( unittest.TestCase ): @cached_property def _a ( self ) -> List[str]: return BeitImageProcessor.from_pretrained("microsoft/beit-base-patch16-224" ) if is_vision_available() else None @slow def _a ( self ) -> Tuple: _UpperCAmelCase = FlaxBeitForMaskedImageModeling.from_pretrained("microsoft/beit-base-patch16-224-pt22k" ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="np" ).pixel_values # prepare bool_masked_pos _UpperCAmelCase = np.ones((1, 196) , dtype=__SCREAMING_SNAKE_CASE ) # forward pass _UpperCAmelCase = model(pixel_values=__SCREAMING_SNAKE_CASE , bool_masked_pos=__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = outputs.logits # verify the logits _UpperCAmelCase = (1, 196, 8192) self.assertEqual(logits.shape , __SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.array( [[-3.2437, 0.5072, -13.9174], [-3.2456, 0.4948, -13.9401], [-3.2033, 0.5121, -13.8550]] ) self.assertTrue(np.allclose(logits[bool_masked_pos][:3, :3] , __SCREAMING_SNAKE_CASE , atol=1e-2 ) ) @slow def _a ( self ) -> Optional[Any]: _UpperCAmelCase = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-base-patch16-224" ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="np" ) # forward pass _UpperCAmelCase = model(**__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = outputs.logits # verify the logits _UpperCAmelCase = (1, 1000) self.assertEqual(logits.shape , __SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.array([-1.2385, -1.0987, -1.0108] ) self.assertTrue(np.allclose(logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) ) _UpperCAmelCase = 281 self.assertEqual(logits.argmax(-1 ).item() , __SCREAMING_SNAKE_CASE ) @slow def _a ( self ) -> Any: _UpperCAmelCase = FlaxBeitForImageClassification.from_pretrained("microsoft/beit-large-patch16-224-pt22k-ft22k" ) _UpperCAmelCase = self.default_image_processor _UpperCAmelCase = prepare_img() _UpperCAmelCase = image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="np" ) # forward pass _UpperCAmelCase = model(**__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = outputs.logits # verify the logits _UpperCAmelCase = (1, 21841) self.assertEqual(logits.shape , __SCREAMING_SNAKE_CASE ) _UpperCAmelCase = np.array([1.6881, -0.2787, 0.5901] ) self.assertTrue(np.allclose(logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) ) _UpperCAmelCase = 2396 self.assertEqual(logits.argmax(-1 ).item() , __SCREAMING_SNAKE_CASE )
657
from typing import List from .keymap import KEYMAP, get_character def lowercase__ ( A_: str ) -> str: """simple docstring""" def decorator(A_: int ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += [key] setattr(A_ , """handle_key""" , A_ ) return func return decorator def lowercase__ ( *A_: List[str] ) -> Optional[int]: """simple docstring""" def decorator(A_: Tuple ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += keys setattr(A_ , """handle_key""" , A_ ) return func return decorator class _A ( UpperCamelCase ): """simple docstring""" def __new__( cls : str , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> int: __UpperCAmelCase =super().__new__(cls , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if not hasattr(__SCREAMING_SNAKE_CASE , """key_handler""" ): setattr(__SCREAMING_SNAKE_CASE , """key_handler""" , {} ) setattr(__SCREAMING_SNAKE_CASE , """handle_input""" , KeyHandler.handle_input ) for value in attrs.values(): __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , """handle_key""" , [] ) for key in handled_keys: __UpperCAmelCase =value return new_cls @staticmethod def _a ( cls : Dict ) -> List[Any]: __UpperCAmelCase =get_character() if char != KEYMAP["undefined"]: __UpperCAmelCase =ord(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =cls.key_handler.get(__SCREAMING_SNAKE_CASE ) if handler: __UpperCAmelCase =char return handler(cls ) else: return None def lowercase__ ( cls: str ) -> int: """simple docstring""" return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
68
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available a_ = { 'configuration_bridgetower': [ 'BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BridgeTowerConfig', 'BridgeTowerTextConfig', 'BridgeTowerVisionConfig', ], 'processing_bridgetower': ['BridgeTowerProcessor'], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['BridgeTowerImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST', 'BridgeTowerForContrastiveLearning', 'BridgeTowerForImageAndTextRetrieval', 'BridgeTowerForMaskedLM', 'BridgeTowerModel', 'BridgeTowerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_bridgetower import ( BRIDGETOWER_PRETRAINED_CONFIG_ARCHIVE_MAP, BridgeTowerConfig, BridgeTowerTextConfig, BridgeTowerVisionConfig, ) from .processing_bridgetower import BridgeTowerProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_bridgetower import BridgeTowerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bridgetower import ( BRIDGETOWER_PRETRAINED_MODEL_ARCHIVE_LIST, BridgeTowerForContrastiveLearning, BridgeTowerForImageAndTextRetrieval, BridgeTowerForMaskedLM, BridgeTowerModel, BridgeTowerPreTrainedModel, ) else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure)
296
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
68
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class _a ( metaclass=SCREAMING_SNAKE_CASE__): __magic_name__ = ['torch', 'torchsde'] def __init__( self : str , *_lowercase : Dict , **_lowercase : Dict ) -> Union[str, Any]: requires_backends(self , ["torch", "torchsde"] ) @classmethod def __lowercase ( cls : int , *_lowercase : Dict , **_lowercase : int ) -> Union[str, Any]: requires_backends(cls , ["torch", "torchsde"] ) @classmethod def __lowercase ( cls : Any , *_lowercase : List[str] , **_lowercase : Optional[Any] ) -> str: requires_backends(cls , ["torch", "torchsde"] )
449
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class _A ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=7 , __SCREAMING_SNAKE_CASE : Tuple=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : int=True , __SCREAMING_SNAKE_CASE : Optional[Any]=99 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : int=5 , __SCREAMING_SNAKE_CASE : Dict=4 , __SCREAMING_SNAKE_CASE : str=37 , __SCREAMING_SNAKE_CASE : Union[str, Any]="gelu" , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : str=512 , __SCREAMING_SNAKE_CASE : Dict=16 , __SCREAMING_SNAKE_CASE : Any=2 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.02 , __SCREAMING_SNAKE_CASE : List[str]=4 , ) -> Optional[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =seq_length __UpperCAmelCase =is_training __UpperCAmelCase =use_attention_mask __UpperCAmelCase =use_token_type_ids __UpperCAmelCase =use_labels __UpperCAmelCase =vocab_size __UpperCAmelCase =hidden_size __UpperCAmelCase =num_hidden_layers __UpperCAmelCase =num_attention_heads __UpperCAmelCase =intermediate_size __UpperCAmelCase =hidden_act __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =max_position_embeddings __UpperCAmelCase =type_vocab_size __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =initializer_range __UpperCAmelCase =num_choices def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase =None if self.use_attention_mask: __UpperCAmelCase =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase =None if self.use_token_type_ids: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase =RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def _a ( self : List[str] ) -> Dict: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase =True __UpperCAmelCase =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = True lowerCamelCase : Union[str, Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =FlaxRobertaModelTester(self ) @slow def _a ( self : Optional[Any] ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase =model_class_name.from_pretrained("""roberta-base""" , from_pt=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model(np.ones((1, 1) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _lowercase = { 'configuration_mobilenet_v2': [ 'MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP', 'MobileNetV2Config', 'MobileNetV2OnnxConfig', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = ['MobileNetV2FeatureExtractor'] _lowercase = ['MobileNetV2ImageProcessor'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _lowercase = [ 'MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST', 'MobileNetV2ForImageClassification', 'MobileNetV2ForSemanticSegmentation', 'MobileNetV2Model', 'MobileNetV2PreTrainedModel', 'load_tf_weights_in_mobilenet_v2', ] if TYPE_CHECKING: from .configuration_mobilenet_va import ( MOBILENET_V2_PRETRAINED_CONFIG_ARCHIVE_MAP, MobileNetVaConfig, MobileNetVaOnnxConfig, ) try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_mobilenet_va import MobileNetVaFeatureExtractor from .image_processing_mobilenet_va import MobileNetVaImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mobilenet_va import ( MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel, MobileNetVaPreTrainedModel, load_tf_weights_in_mobilenet_va, ) else: import sys _lowercase = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
342
from __future__ import annotations def lowercase__ ( A_: list[list[int]] ) -> int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(A_ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
68
0
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __lowerCamelCase : Tuple = logging.get_logger(__name__) __lowerCamelCase : Any = { "camembert-base": "https://huggingface.co/camembert-base/resolve/main/config.json", "umberto-commoncrawl-cased-v1": ( "https://huggingface.co/Musixmatch/umberto-commoncrawl-cased-v1/resolve/main/config.json" ), "umberto-wikipedia-uncased-v1": ( "https://huggingface.co/Musixmatch/umberto-wikipedia-uncased-v1/resolve/main/config.json" ), } class __magic_name__ ( A__ ): lowercase : Optional[Any] ='camembert' def __init__( self : Dict , UpperCamelCase__ : Dict=3_05_22 , UpperCamelCase__ : Tuple=7_68 , UpperCamelCase__ : List[Any]=12 , UpperCamelCase__ : Any=12 , UpperCamelCase__ : Tuple=30_72 , UpperCamelCase__ : Dict="gelu" , UpperCamelCase__ : Optional[int]=0.1 , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Optional[Any]=5_12 , UpperCamelCase__ : Union[str, Any]=2 , UpperCamelCase__ : Union[str, Any]=0.02 , UpperCamelCase__ : List[str]=1e-1_2 , UpperCamelCase__ : List[str]=1 , UpperCamelCase__ : Any=0 , UpperCamelCase__ : str=2 , UpperCamelCase__ : Any="absolute" , UpperCamelCase__ : Optional[Any]=True , UpperCamelCase__ : Any=None , **UpperCamelCase__ : int , ) -> Optional[int]: '''simple docstring''' super().__init__(pad_token_id=__SCREAMING_SNAKE_CASE , bos_token_id=__SCREAMING_SNAKE_CASE , eos_token_id=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) UpperCAmelCase = vocab_size UpperCAmelCase = hidden_size UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = hidden_act UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = max_position_embeddings UpperCAmelCase = type_vocab_size UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = position_embedding_type UpperCAmelCase = use_cache UpperCAmelCase = classifier_dropout class __magic_name__ ( A__ ): @property def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task == "multiple-choice": UpperCAmelCase = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ] )
323
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowercase__ ( A_: int , A_: int , A_: int , A_: int , A_: int , A_: int ) -> np.ndarray: """simple docstring""" if (ksize % 2) == 0: __UpperCAmelCase =ksize + 1 __UpperCAmelCase =np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(A_ ): for x in range(A_ ): # distance from center __UpperCAmelCase =x - ksize // 2 __UpperCAmelCase =y - ksize // 2 # degree to radiant __UpperCAmelCase =theta / 180 * np.pi __UpperCAmelCase =np.cos(_theta ) __UpperCAmelCase =np.sin(_theta ) # get kernel x __UpperCAmelCase =cos_theta * px + sin_theta * py # get kernel y __UpperCAmelCase =-sin_theta * px + cos_theta * py # fill kernel __UpperCAmelCase =np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image __A = imread("../image_data/lena.jpg") # turn image in gray scale value __A = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges __A = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 1_20, 1_50]: __A = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) __A = out / out.max() * 2_55 __A = out.astype(np.uinta) imshow("Original", gray) imshow("Gabor filter with 20x20 mask and 6 directions", out) waitKey(0)
68
0
'''simple docstring''' import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging __lowerCAmelCase : List[Any] = logging.get_logger(__name__) class A ( UpperCAmelCase ): a_ = ['input_features'] def __init__( self : Union[str, Any] , __a : Union[str, Any]=8_0 , __a : List[Any]=1_6_0_0_0 , __a : int=1_6_0 , __a : Optional[int]=3_0 , __a : str=4_0_0 , __a : List[Any]=0.0 , __a : List[str]=False , **__a : List[str] , ) -> Union[str, Any]: super().__init__( feature_size=__SCREAMING_SNAKE_CASE , sampling_rate=__SCREAMING_SNAKE_CASE , padding_value=__SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) __UpperCAmelCase = n_fft __UpperCAmelCase = hop_length __UpperCAmelCase = chunk_length __UpperCAmelCase = chunk_length * sampling_rate __UpperCAmelCase = self.n_samples // hop_length __UpperCAmelCase = sampling_rate __UpperCAmelCase = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__SCREAMING_SNAKE_CASE , min_frequency=0.0 , max_frequency=8_0_0_0.0 , sampling_rate=__SCREAMING_SNAKE_CASE , norm='''slaney''' , mel_scale='''slaney''' , ) def snake_case__ ( self : Any , __a : np.array ) -> np.ndarray: __UpperCAmelCase = spectrogram( __SCREAMING_SNAKE_CASE , window_function(self.n_fft , '''hann''' ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel='''log10''' , ) __UpperCAmelCase = log_spec[:, :-1] __UpperCAmelCase = np.maximum(__SCREAMING_SNAKE_CASE , log_spec.max() - 8.0 ) __UpperCAmelCase = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def snake_case__ ( __a : List[np.ndarray] , __a : List[np.ndarray] , __a : float = 0.0 ) -> List[np.ndarray]: if attention_mask is not None: __UpperCAmelCase = np.array(__SCREAMING_SNAKE_CASE , np.intaa ) __UpperCAmelCase = [] for vector, length in zip(__SCREAMING_SNAKE_CASE , attention_mask.sum(-1 ) ): __UpperCAmelCase = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: __UpperCAmelCase = padding_value normed_input_values.append(__SCREAMING_SNAKE_CASE ) else: __UpperCAmelCase = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __call__( self : Optional[Any] , __a : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __a : bool = True , __a : Optional[int] = None , __a : Optional[Union[str, TensorType]] = None , __a : Optional[bool] = None , __a : Optional[str] = "max_length" , __a : Optional[int] = None , __a : Optional[int] = None , __a : Optional[bool] = None , **__a : Any , ) -> BatchFeature: if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a""" f""" sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input""" f""" was sampled with {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( '''It is strongly recommended to pass the `sampling_rate` argument to this function. ''' '''Failing to do so can result in silent errors that might be hard to debug.''' ) __UpperCAmelCase = isinstance(__SCREAMING_SNAKE_CASE , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) __UpperCAmelCase = is_batched_numpy or ( isinstance(__SCREAMING_SNAKE_CASE , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: __UpperCAmelCase = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__SCREAMING_SNAKE_CASE , np.ndarray ): __UpperCAmelCase = np.asarray(__SCREAMING_SNAKE_CASE , dtype=np.floataa ) elif isinstance(__SCREAMING_SNAKE_CASE , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): __UpperCAmelCase = raw_speech.astype(np.floataa ) # always return batch if not is_batched: __UpperCAmelCase = [np.asarray([raw_speech] ).T] __UpperCAmelCase = BatchFeature({'''input_features''': raw_speech} ) # convert into correct format for padding __UpperCAmelCase = self.pad( __SCREAMING_SNAKE_CASE , padding=__SCREAMING_SNAKE_CASE , max_length=max_length if max_length else self.n_samples , truncation=__SCREAMING_SNAKE_CASE , pad_to_multiple_of=__SCREAMING_SNAKE_CASE , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: __UpperCAmelCase = self.zero_mean_unit_var_norm( padded_inputs['''input_features'''] , attention_mask=padded_inputs['''attention_mask'''] , padding_value=self.padding_value , ) __UpperCAmelCase = np.stack(padded_inputs['''input_features'''] , axis=0 ) # make sure list is in array format __UpperCAmelCase = padded_inputs.get('''input_features''' ).transpose(2 , 0 , 1 ) __UpperCAmelCase = [self._np_extract_fbank_features(__SCREAMING_SNAKE_CASE ) for waveform in input_features[0]] if isinstance(input_features[0] , __SCREAMING_SNAKE_CASE ): __UpperCAmelCase = [np.asarray(__SCREAMING_SNAKE_CASE , dtype=np.floataa ) for feature in input_features] else: __UpperCAmelCase = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) __UpperCAmelCase = padded_inputs['''attention_mask'''][:, :: self.hop_length] if return_tensors is not None: __UpperCAmelCase = padded_inputs.convert_to_tensors(__SCREAMING_SNAKE_CASE ) return padded_inputs def snake_case__ ( self : int ) -> Dict[str, Any]: __UpperCAmelCase = copy.deepcopy(self.__dict__ ) __UpperCAmelCase = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
262
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _A : """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : Optional[int]=2 , __SCREAMING_SNAKE_CASE : Optional[int]=3 , __SCREAMING_SNAKE_CASE : List[str]=16 , __SCREAMING_SNAKE_CASE : Union[str, Any]=[1, 2, 1] , __SCREAMING_SNAKE_CASE : List[Any]=[2, 2, 4] , __SCREAMING_SNAKE_CASE : str=2 , __SCREAMING_SNAKE_CASE : Any=2.0 , __SCREAMING_SNAKE_CASE : Optional[int]=True , __SCREAMING_SNAKE_CASE : int=0.0 , __SCREAMING_SNAKE_CASE : Dict=0.0 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : Any="gelu" , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : List[Any]=0.02 , __SCREAMING_SNAKE_CASE : Tuple=1e-5 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : Optional[int]=None , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : Any=10 , __SCREAMING_SNAKE_CASE : Dict=8 , ) -> List[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =image_size __UpperCAmelCase =patch_size __UpperCAmelCase =num_channels __UpperCAmelCase =embed_dim __UpperCAmelCase =depths __UpperCAmelCase =num_heads __UpperCAmelCase =window_size __UpperCAmelCase =mlp_ratio __UpperCAmelCase =qkv_bias __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =drop_path_rate __UpperCAmelCase =hidden_act __UpperCAmelCase =use_absolute_embeddings __UpperCAmelCase =patch_norm __UpperCAmelCase =layer_norm_eps __UpperCAmelCase =initializer_range __UpperCAmelCase =is_training __UpperCAmelCase =scope __UpperCAmelCase =use_labels __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =encoder_stride def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase =None if self.use_labels: __UpperCAmelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase =self.get_config() return config, pixel_values, labels def _a ( self : List[Any] ) -> Optional[Any]: return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[int]: __UpperCAmelCase =SwinvaModel(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __UpperCAmelCase =int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : int ) -> Tuple: __UpperCAmelCase =SwinvaForMaskedImageModeling(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __UpperCAmelCase =1 __UpperCAmelCase =SwinvaForMaskedImageModeling(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Tuple: __UpperCAmelCase =self.type_sequence_label_size __UpperCAmelCase =SwinvaForImageClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : List[str] ) -> Tuple: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _A ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Optional[int] = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowerCamelCase : Tuple = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowerCamelCase : Dict = False lowerCamelCase : Tuple = False lowerCamelCase : List[str] = False lowerCamelCase : Tuple = False def _a ( self : str ) -> str: __UpperCAmelCase =SwinvaModelTester(self ) __UpperCAmelCase =ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , embed_dim=37 ) def _a ( self : List[Any] ) -> Optional[int]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : str ) -> str: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def _a ( self : Tuple ) -> Tuple: pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def _a ( self : Optional[Any] ) -> int: pass def _a ( self : Tuple ) -> int: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __UpperCAmelCase =model.get_output_embeddings() self.assertTrue(x is None or isinstance(__SCREAMING_SNAKE_CASE , nn.Linear ) ) def _a ( self : str ) -> List[str]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase =[*signature.parameters.keys()] __UpperCAmelCase =["""pixel_values"""] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =True for model_class in self.all_model_classes: __UpperCAmelCase =True __UpperCAmelCase =False __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions __UpperCAmelCase =len(self.model_tester.depths ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCAmelCase =True __UpperCAmelCase =config.window_size**2 __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) __UpperCAmelCase =len(__SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine __UpperCAmelCase =True __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if hasattr(self.model_tester , """num_hidden_states_types""" ): __UpperCAmelCase =self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states __UpperCAmelCase =2 self.assertEqual(out_len + added_hidden_states , len(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> int: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.hidden_states __UpperCAmelCase =getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # Swinv2 has a different seq_length __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) __UpperCAmelCase =outputs.reshaped_hidden_states self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =reshaped_hidden_states[0].shape __UpperCAmelCase =( reshaped_hidden_states[0].view(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =3 __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __UpperCAmelCase =image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Dict: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__SCREAMING_SNAKE_CASE ) @slow def _a ( self : int ) -> Dict: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase =SwinvaModel.from_pretrained(__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) def _a ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =_config_zero_init(__SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: __UpperCAmelCase =model_class(config=__SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class _A ( unittest.TestCase ): """simple docstring""" @cached_property def _a ( self : Tuple ) -> Dict: return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def _a ( self : int ) -> Optional[int]: __UpperCAmelCase =SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.default_image_processor __UpperCAmelCase =Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) __UpperCAmelCase =image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).to(__SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE ) # verify the logits __UpperCAmelCase =torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([-0.3_947, -0.4_306, 0.0_026] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) )
68
0
"""simple docstring""" import logging import os import sys import warnings from dataclasses import dataclass, field from random import randint from typing import Optional import datasets import evaluate import numpy as np from datasets import DatasetDict, load_dataset import transformers from transformers import ( AutoConfig, AutoFeatureExtractor, AutoModelForAudioClassification, HfArgumentParser, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version, send_example_telemetry from transformers.utils.versions import require_version __SCREAMING_SNAKE_CASE =logging.getLogger(__name__) # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version("4.31.0") require_version("datasets>=1.14.0", "To fix: pip install -r examples/pytorch/audio-classification/requirements.txt") def lowercase__( __SCREAMING_SNAKE_CASE : np.ndarray , __SCREAMING_SNAKE_CASE : float , __SCREAMING_SNAKE_CASE : int = 1_60_00 ): lowercase_ : Optional[Any] = int(round(sample_rate * max_length ) ) if len(A_ ) <= sample_length: return wav lowercase_ : List[Any] = randint(0 , len(A_ ) - sample_length - 1 ) return wav[random_offset : random_offset + sample_length] @dataclass class UpperCamelCase : lowercase = field(default=lowercase_ , metadata={'help': 'Name of a dataset from the datasets package'} ) lowercase = field( default=lowercase_ , metadata={'help': 'The configuration name of the dataset to use (via the datasets library).'} ) lowercase = field( default=lowercase_ , metadata={'help': 'A file containing the training audio paths and labels.'} ) lowercase = field( default=lowercase_ , metadata={'help': 'A file containing the validation audio paths and labels.'} ) lowercase = field( default='train' , metadata={ 'help': 'The name of the training data set split to use (via the datasets library). Defaults to \'train\'' } , ) lowercase = field( default='validation' , metadata={ 'help': ( 'The name of the training data set split to use (via the datasets library). Defaults to \'validation\'' ) } , ) lowercase = field( default='audio' , metadata={'help': 'The name of the dataset column containing the audio data. Defaults to \'audio\''} , ) lowercase = field( default='label' , metadata={'help': 'The name of the dataset column containing the labels. Defaults to \'label\''} ) lowercase = field( default=lowercase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of training examples to this ' 'value if set.' ) } , ) lowercase = field( default=lowercase_ , metadata={ 'help': ( 'For debugging purposes or quicker training, truncate the number of evaluation examples to this ' 'value if set.' ) } , ) lowercase = field( default=2_0 , metadata={'help': 'Audio clips will be randomly cut to this length during training if the value is set.'} , ) @dataclass class UpperCamelCase : lowercase = field( default='facebook/wav2vec2-base' , metadata={'help': 'Path to pretrained model or model identifier from huggingface.co/models'} , ) lowercase = field( default=lowercase_ , metadata={'help': 'Pretrained config name or path if not the same as model_name'} ) lowercase = field( default=lowercase_ , metadata={'help': 'Where do you want to store the pretrained models downloaded from the Hub'} ) lowercase = field( default='main' , metadata={'help': 'The specific model version to use (can be a branch name, tag name or commit id).'} , ) lowercase = field( default=lowercase_ , metadata={'help': 'Name or path of preprocessor config.'} ) lowercase = field( default=lowercase_ , metadata={'help': 'Whether to freeze the feature encoder layers of the model.'} ) lowercase = field( default=lowercase_ , metadata={'help': 'Whether to generate an attention mask in the feature extractor.'} ) lowercase = field( default=lowercase_ , metadata={ 'help': ( 'Will use the token generated when running `huggingface-cli login` (necessary to use this script ' 'with private models).' ) } , ) lowercase = field( default=lowercase_ , metadata={'help': 'Whether to freeze the feature extractor layers of the model.'} ) lowercase = field( default=lowercase_ , metadata={'help': 'Will enable to load a pretrained model whose head dimensions are different.'} , ) def _UpperCAmelCase ( self ) -> Union[str, Any]: '''simple docstring''' if not self.freeze_feature_extractor and self.freeze_feature_encoder: warnings.warn( 'The argument `--freeze_feature_extractor` is deprecated and ' 'will be removed in a future version. Use `--freeze_feature_encoder`' 'instead. Setting `freeze_feature_encoder==True`.' ,__SCREAMING_SNAKE_CASE ,) if self.freeze_feature_extractor and not self.freeze_feature_encoder: raise ValueError( 'The argument `--freeze_feature_extractor` is deprecated and ' 'should not be used in combination with `--freeze_feature_encoder`.' 'Only make use of `--freeze_feature_encoder`.' ) def lowercase__( ): lowercase_ : Tuple = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith('.json' ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. lowercase_ , lowercase_ , lowercase_ : Dict = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: lowercase_ , lowercase_ , lowercase_ : Dict = parser.parse_args_into_dataclasses() # Sending telemetry. Tracking the example usage helps us better allocate resources to maintain them. The # information sent is the one passed as arguments along with your Python/PyTorch versions. send_example_telemetry('run_audio_classification' , A_ , A_ ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , handlers=[logging.StreamHandler(sys.stdout )] , ) if training_args.should_log: # The default of training_args.log_level is passive, so we set log level at info here to have that default. transformers.utils.logging.set_verbosity_info() lowercase_ : Optional[Any] = training_args.get_process_log_level() logger.setLevel(A_ ) transformers.utils.logging.set_verbosity(A_ ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'''Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu} ''' + F'''distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}''' ) logger.info(F'''Training/evaluation parameters {training_args}''' ) # Set seed before initializing model. set_seed(training_args.seed ) # Detecting last checkpoint. lowercase_ : Any = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: lowercase_ : Tuple = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'''Output directory ({training_args.output_dir}) already exists and is not empty. ''' 'Use --overwrite_output_dir to train from scratch.' ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'''Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ''' 'the `--output_dir` or add `--overwrite_output_dir` to train from scratch.' ) # Initialize our dataset and prepare it for the audio classification task. lowercase_ : Dict = DatasetDict() lowercase_ : List[Any] = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.train_split_name , use_auth_token=True if model_args.use_auth_token else None , ) lowercase_ : Any = load_dataset( data_args.dataset_name , data_args.dataset_config_name , split=data_args.eval_split_name , use_auth_token=True if model_args.use_auth_token else None , ) if data_args.audio_column_name not in raw_datasets["train"].column_names: raise ValueError( F'''--audio_column_name {data_args.audio_column_name} not found in dataset \'{data_args.dataset_name}\'. ''' 'Make sure to set `--audio_column_name` to the correct audio column - one of ' F'''{", ".join(raw_datasets["train"].column_names )}.''' ) if data_args.label_column_name not in raw_datasets["train"].column_names: raise ValueError( F'''--label_column_name {data_args.label_column_name} not found in dataset \'{data_args.dataset_name}\'. ''' 'Make sure to set `--label_column_name` to the correct text column - one of ' F'''{", ".join(raw_datasets["train"].column_names )}.''' ) # Setting `return_attention_mask=True` is the way to get a correctly masked mean-pooling over # transformer outputs in the classifier, but it doesn't always lead to better accuracy lowercase_ : Tuple = AutoFeatureExtractor.from_pretrained( model_args.feature_extractor_name or model_args.model_name_or_path , return_attention_mask=model_args.attention_mask , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # `datasets` takes care of automatically loading and resampling the audio, # so we just need to set the correct target sampling rate. lowercase_ : Optional[int] = raw_datasets.cast_column( data_args.audio_column_name , datasets.features.Audio(sampling_rate=feature_extractor.sampling_rate ) ) lowercase_ : List[Any] = feature_extractor.model_input_names[0] def train_transforms(__SCREAMING_SNAKE_CASE : Any ): lowercase_ : Optional[int] = [] for audio in batch[data_args.audio_column_name]: lowercase_ : List[str] = random_subsample( audio['array'] , max_length=data_args.max_length_seconds , sample_rate=feature_extractor.sampling_rate ) subsampled_wavs.append(A_ ) lowercase_ : str = feature_extractor(A_ , sampling_rate=feature_extractor.sampling_rate ) lowercase_ : Optional[Any] = {model_input_name: inputs.get(A_ )} lowercase_ : Union[str, Any] = list(batch[data_args.label_column_name] ) return output_batch def val_transforms(__SCREAMING_SNAKE_CASE : str ): lowercase_ : Any = [audio['array'] for audio in batch[data_args.audio_column_name]] lowercase_ : Tuple = feature_extractor(A_ , sampling_rate=feature_extractor.sampling_rate ) lowercase_ : Any = {model_input_name: inputs.get(A_ )} lowercase_ : str = list(batch[data_args.label_column_name] ) return output_batch # Prepare label mappings. # We'll include these in the model's config to get human readable labels in the Inference API. lowercase_ : Tuple = raw_datasets['train'].features[data_args.label_column_name].names lowercase_ , lowercase_ : Union[str, Any] = {}, {} for i, label in enumerate(A_ ): lowercase_ : int = str(A_ ) lowercase_ : Union[str, Any] = label # Load the accuracy metric from the datasets package lowercase_ : str = evaluate.load('accuracy' ) # Define our compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with # `predictions` and `label_ids` fields) and has to return a dictionary string to float. def compute_metrics(__SCREAMING_SNAKE_CASE : str ): lowercase_ : List[Any] = np.argmax(eval_pred.predictions , axis=1 ) return metric.compute(predictions=A_ , references=eval_pred.label_ids ) lowercase_ : str = AutoConfig.from_pretrained( model_args.config_name or model_args.model_name_or_path , num_labels=len(A_ ) , labelaid=A_ , idalabel=A_ , finetuning_task='audio-classification' , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) lowercase_ : Optional[int] = AutoModelForAudioClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=A_ , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ignore_mismatched_sizes=model_args.ignore_mismatched_sizes , ) # freeze the convolutional waveform encoder if model_args.freeze_feature_encoder: model.freeze_feature_encoder() if training_args.do_train: if data_args.max_train_samples is not None: lowercase_ : List[str] = ( raw_datasets['train'].shuffle(seed=training_args.seed ).select(range(data_args.max_train_samples ) ) ) # Set the training transforms raw_datasets["train"].set_transform(A_ , output_all_columns=A_ ) if training_args.do_eval: if data_args.max_eval_samples is not None: lowercase_ : int = ( raw_datasets['eval'].shuffle(seed=training_args.seed ).select(range(data_args.max_eval_samples ) ) ) # Set the validation transforms raw_datasets["eval"].set_transform(A_ , output_all_columns=A_ ) # Initialize our trainer lowercase_ : Tuple = Trainer( model=A_ , args=A_ , train_dataset=raw_datasets['train'] if training_args.do_train else None , eval_dataset=raw_datasets['eval'] if training_args.do_eval else None , compute_metrics=A_ , tokenizer=A_ , ) # Training if training_args.do_train: lowercase_ : List[str] = None if training_args.resume_from_checkpoint is not None: lowercase_ : int = training_args.resume_from_checkpoint elif last_checkpoint is not None: lowercase_ : Dict = last_checkpoint lowercase_ : Dict = trainer.train(resume_from_checkpoint=A_ ) trainer.save_model() trainer.log_metrics('train' , train_result.metrics ) trainer.save_metrics('train' , train_result.metrics ) trainer.save_state() # Evaluation if training_args.do_eval: lowercase_ : Tuple = trainer.evaluate() trainer.log_metrics('eval' , A_ ) trainer.save_metrics('eval' , A_ ) # Write model card and (optionally) push to hub lowercase_ : Dict = { 'finetuned_from': model_args.model_name_or_path, 'tasks': 'audio-classification', 'dataset': data_args.dataset_name, 'tags': ['audio-classification'], } if training_args.push_to_hub: trainer.push_to_hub(**A_ ) else: trainer.create_model_card(**A_ ) if __name__ == "__main__": main()
425
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A = logging.get_logger(__name__) __A = {"vocab_file": "spiece.model"} __A = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } __A = { "AI-Sweden/gpt-sw3-126m": 20_48, "AI-Sweden/gpt-sw3-350m": 20_48, "AI-Sweden/gpt-sw3-1.6b": 20_48, "AI-Sweden/gpt-sw3-6.7b": 20_48, "AI-Sweden/gpt-sw3-20b": 20_48, } class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : int = VOCAB_FILES_NAMES lowerCamelCase : int = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = ['input_ids', 'attention_mask'] def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : Union[str, Any]=False , __SCREAMING_SNAKE_CASE : List[str]=False , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Optional[Dict[str, Any]] = None , **__SCREAMING_SNAKE_CASE : Optional[Any] , ) -> None: __UpperCAmelCase ={} if sp_model_kwargs is None else sp_model_kwargs __UpperCAmelCase =kwargs.get("""name_or_path""" ) if name_or_path is None: logger.warning( """name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,""" """ you are testing the model, this can safely be ignored""" ) __UpperCAmelCase ="""None""" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __UpperCAmelCase ="""<|endoftext|>""" if eos_token is None else eos_token __UpperCAmelCase ="""<unk>""" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __UpperCAmelCase =unk_token if pad_token is None else pad_token __UpperCAmelCase =eos_token if bos_token is None else bos_token else: __UpperCAmelCase ="""<pad>""" if pad_token is None else pad_token __UpperCAmelCase ="""<s>""" if bos_token is None else bos_token super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) __UpperCAmelCase =do_lower_case __UpperCAmelCase =remove_space __UpperCAmelCase =keep_accents __UpperCAmelCase =vocab_file __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) # Used for whitespace normalization in input texts # fmt : off __UpperCAmelCase ={""" """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """""", """„"""} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __UpperCAmelCase =re.compile( f'''[{"".join(map(__SCREAMING_SNAKE_CASE , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]''' ) def __getstate__( self : Any ) -> str: __UpperCAmelCase =self.__dict__.copy() __UpperCAmelCase =None return state def __setstate__( self : str , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Union[str, Any]: __UpperCAmelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __UpperCAmelCase ={} __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def _a ( self : Union[str, Any] ) -> int: return len(self.sp_model ) def _a ( self : Dict , __SCREAMING_SNAKE_CASE : str ) -> str: __UpperCAmelCase =self.non_printing_characters_re.sub("""""" , __SCREAMING_SNAKE_CASE ) # Normalize whitespaces __UpperCAmelCase ="""""".join([char if char not in self.whitespaces else """ """ for char in text] ) # NFC Unicode normalization __UpperCAmelCase =unicodedata.normalize("""NFC""" , __SCREAMING_SNAKE_CASE ) return text def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> List[str]: __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) return self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str ) -> int: return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : int ) -> str: return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) @staticmethod def _a ( __SCREAMING_SNAKE_CASE : str ) -> str: return out_string def _a ( self : Any , __SCREAMING_SNAKE_CASE : List[str] ) -> str: __UpperCAmelCase =[] __UpperCAmelCase ="""""" __UpperCAmelCase =False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) + token __UpperCAmelCase =True __UpperCAmelCase =[] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =False out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) return out_string def _a ( self : Any ) -> Dict[str, int]: __UpperCAmelCase ={self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _a ( self : Any , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __UpperCAmelCase =os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE , """wb""" ) as fi: __UpperCAmelCase =self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, List[str]] , __SCREAMING_SNAKE_CASE : Union[str, bool] = False ) -> Union[List[int], List[List[int]], "torch.Tensor"]: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) else: __UpperCAmelCase =[self.preprocess_text(__SCREAMING_SNAKE_CASE ) for t in text] __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) if return_tensors is True or return_tensors == "pt": __UpperCAmelCase =torch.tensor(__SCREAMING_SNAKE_CASE ) return token_ids def _a ( self : str , __SCREAMING_SNAKE_CASE : Union[int, List[int]] ) -> str: return self.sp_model.decode(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : "Conversation" ) -> List[int]: __UpperCAmelCase =[f'''User: {text}''' if is_user else f'''Bot: {text}''' for is_user, text in conversation.iter_texts()] __UpperCAmelCase =( f'''{self.eos_token}{self.bos_token}''' + f'''{self.bos_token}'''.join(__SCREAMING_SNAKE_CASE ) + f'''{self.bos_token}Bot:''' ) return self.encode(text=__SCREAMING_SNAKE_CASE )
68
0
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets UpperCamelCase_ : Any = datasets.logging.get_logger(__name__) UpperCamelCase_ : Dict = '''\\n@inproceedings{rei-EtAl:2020:WMT,\n author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon},\n title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task},\n booktitle = {Proceedings of the Fifth Conference on Machine Translation},\n month = {November},\n year = {2020},\n address = {Online},\n publisher = {Association for Computational Linguistics},\n pages = {909--918},\n}\n@inproceedings{rei-etal-2020-comet,\n title = \"{COMET}: A Neural Framework for {MT} Evaluation\",\n author = \"Rei, Ricardo and\n Stewart, Craig and\n Farinha, Ana C and\n Lavie, Alon\",\n booktitle = \"Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)\",\n month = nov,\n year = \"2020\",\n address = \"Online\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/2020.emnlp-main.213\",\n pages = \"2685--2702\",\n}\n''' UpperCamelCase_ : int = '''\\nCrosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM).\nWith the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition.\n\nSee the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information.\n''' UpperCamelCase_ : Union[str, Any] = '''\nCOMET score.\n\nArgs:\n\n`sources` (list of str): Source sentences\n`predictions` (list of str): candidate translations\n`references` (list of str): reference translations\n`cuda` (bool): If set to True, runs COMET using GPU\n`show_progress` (bool): Shows progress\n`model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None.\n\nReturns:\n `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`.\n `scores`: List of scores.\n\nExamples:\n\n >>> comet_metric = datasets.load_metric(\'comet\')\n >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use\n >>> source = [\"Dem Feuer konnte Einhalt geboten werden\", \"Schulen und Kindergärten wurden eröffnet.\"]\n >>> hypothesis = [\"The fire could be stopped\", \"Schools and kindergartens were open\"]\n >>> reference = [\"They were able to control the fire.\", \"Schools and kindergartens opened\"]\n >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source)\n >>> print([round(v, 2) for v in results[\"scores\"]])\n [0.19, 0.92]\n''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowerCAmelCase ( datasets.Metric ): """simple docstring""" def lowerCamelCase__ ( self : Optional[Any] ) -> str: """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence" ), "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def lowerCamelCase__ ( self : List[str] , _snake_case : int ) -> Union[str, Any]: """simple docstring""" if self.config_name == "default": A_ = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da" ) ) else: A_ = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def lowerCamelCase__ ( self : Optional[int] , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Optional[int] , _snake_case : Optional[Any]=None , _snake_case : List[str]=False ) -> Union[str, Any]: """simple docstring""" if gpus is None: A_ = 1 if torch.cuda.is_available() else 0 A_ = {"src": sources, "mt": predictions, "ref": references} A_ = [dict(zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) for t in zip(*data.values() )] A_ , A_ = self.scorer.predict(__SCREAMING_SNAKE_CASE , gpus=__SCREAMING_SNAKE_CASE , progress_bar=__SCREAMING_SNAKE_CASE ) return {"mean_score": mean_score, "scores": scores}
115
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __A = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Path , __SCREAMING_SNAKE_CASE : Union[str, None] = None , __SCREAMING_SNAKE_CASE : Union[List[str], None] = None , __SCREAMING_SNAKE_CASE : Union[str, List[str], None] = None , __SCREAMING_SNAKE_CASE : bool = True , ) -> List[str]: __UpperCAmelCase =[file for file in os.listdir(__SCREAMING_SNAKE_CASE ) if os.path.isfile(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) )] if identifier is not None: __UpperCAmelCase =[file for file in files if identifier in file] if n_identifier is not None: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): for n_ in n_identifier: __UpperCAmelCase =[file for file in files if n_ not in file] else: __UpperCAmelCase =[file for file in files if n_identifier not in file] __UpperCAmelCase =ignore_files or [] ignore_files.append("""__init__.py""" ) __UpperCAmelCase =[file for file in files if file not in ignore_files] for file in files: # Open all files print("""Testing""" , __SCREAMING_SNAKE_CASE ) if only_modules: __UpperCAmelCase =file.split(""".""" )[0] try: __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =doctest.DocTestSuite(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =unittest.TextTestRunner().run(__SCREAMING_SNAKE_CASE ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f'''{module_identifier} is not a module.''' ) else: __UpperCAmelCase =doctest.testfile(str("""..""" / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def _a ( self : Optional[Any] ) -> List[str]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""modeling""" __UpperCAmelCase =[ """modeling_ctrl.py""", """modeling_tf_ctrl.py""", ] self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE , ignore_files=__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""tokenization""" self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] ) -> Optional[Any]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""configuration""" self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : List[Any] ) -> Tuple: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase =["""configuration""", """modeling""", """tokenization"""] self.analyze_directory(__SCREAMING_SNAKE_CASE , n_identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : Any ) -> Tuple: __UpperCAmelCase =Path("""docs/source""" ) __UpperCAmelCase =["""favicon.ico"""] self.analyze_directory(__SCREAMING_SNAKE_CASE , ignore_files=__SCREAMING_SNAKE_CASE , only_modules=__SCREAMING_SNAKE_CASE )
68
0
import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() __snake_case :int =logging.get_logger(__name__) def lowerCamelCase_ ( lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : List[Any] , lowerCAmelCase__ : Dict ) -> str: '''simple docstring''' A = WavaVecaForSequenceClassification.from_pretrained(A_ , config=A_ ) A = downstream_dict['projector.weight'] A = downstream_dict['projector.bias'] A = downstream_dict['model.post_net.linear.weight'] A = downstream_dict['model.post_net.linear.bias'] return model def lowerCamelCase_ ( lowerCAmelCase__ : Any , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Any ) -> Tuple: '''simple docstring''' A = WavaVecaForAudioFrameClassification.from_pretrained(A_ , config=A_ ) A = downstream_dict['model.linear.weight'] A = downstream_dict['model.linear.bias'] return model def lowerCamelCase_ ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Optional[int] , lowerCAmelCase__ : List[Any] ) -> List[Any]: '''simple docstring''' A = WavaVecaForXVector.from_pretrained(A_ , config=A_ ) A = downstream_dict['connector.weight'] A = downstream_dict['connector.bias'] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): A = downstream_dict[ F'''model.framelevel_feature_extractor.module.{i}.kernel.weight''' ] A = downstream_dict[F'''model.framelevel_feature_extractor.module.{i}.kernel.bias'''] A = downstream_dict['model.utterancelevel_feature_extractor.linear1.weight'] A = downstream_dict['model.utterancelevel_feature_extractor.linear1.bias'] A = downstream_dict['model.utterancelevel_feature_extractor.linear2.weight'] A = downstream_dict['model.utterancelevel_feature_extractor.linear2.bias'] A = downstream_dict['objective.W'] return model @torch.no_grad() def lowerCamelCase_ ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Any , lowerCAmelCase__ : List[str] ) -> Tuple: '''simple docstring''' A = torch.load(A_ , map_location='cpu' ) A = checkpoint['Downstream'] A = WavaVecaConfig.from_pretrained(A_ ) A = WavaVecaFeatureExtractor.from_pretrained( A_ , return_attention_mask=A_ , do_normalize=A_ ) A = hf_config.architectures[0] if arch.endswith('ForSequenceClassification' ): A = convert_classification(A_ , A_ , A_ ) elif arch.endswith('ForAudioFrameClassification' ): A = convert_diarization(A_ , A_ , A_ ) elif arch.endswith('ForXVector' ): A = convert_xvector(A_ , A_ , A_ ) else: raise NotImplementedError(F'''S3PRL weights conversion is not supported for {arch}''' ) if hf_config.use_weighted_layer_sum: A = checkpoint['Featurizer']['weights'] hf_feature_extractor.save_pretrained(A_ ) hf_model.save_pretrained(A_ ) if __name__ == "__main__": __snake_case :Any =argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') __snake_case :Dict =parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
106
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model __A = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def lowercase__ ( A_: int , A_: Optional[Any] , A_: List[str]=None ) -> List[str]: """simple docstring""" if rng is None: __UpperCAmelCase =random.Random() __UpperCAmelCase =1 for dim in shape: total_dims *= dim __UpperCAmelCase =[] for _ in range(A_ ): values.append(rng.randint(0 , vocab_size - 1 ) ) __UpperCAmelCase =np.array(A_ , dtype=jnp.intaa ).reshape(A_ ) return output def lowercase__ ( A_: List[str] , A_: List[str]=None ) -> Any: """simple docstring""" __UpperCAmelCase =ids_tensor(A_ , vocab_size=2 , rng=A_ ) # make sure that at least one token is attended to for each batch __UpperCAmelCase =1 return attn_mask @require_flax class _A : """simple docstring""" lowerCamelCase : Optional[Any] = None lowerCamelCase : int = () def _a ( self : str ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 __UpperCAmelCase =2 __UpperCAmelCase =inputs["""input_ids"""].shape[-1] // 2 __UpperCAmelCase =inputs["""input_ids"""][:max_batch_size, :sequence_length] __UpperCAmelCase =jnp.ones_like(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens __UpperCAmelCase =input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` __UpperCAmelCase =config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def _a ( self : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =0 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model_class.__name__[4:] # Skip the "Flax" at the beginning __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =pt_model_class(__SCREAMING_SNAKE_CASE ).eval() __UpperCAmelCase =load_flax_weights_in_pytorch_model(__SCREAMING_SNAKE_CASE , flax_model.params ) __UpperCAmelCase =flax_model.generate(__SCREAMING_SNAKE_CASE ).sequences __UpperCAmelCase =pt_model.generate(torch.tensor(__SCREAMING_SNAKE_CASE , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: __UpperCAmelCase =flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def _a ( self : Optional[int] ) -> Dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Union[str, Any] ) -> List[str]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =True __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : List[Any] ) -> Any: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =2 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Any ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =2 __UpperCAmelCase =2 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def _a ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =True __UpperCAmelCase =max_length __UpperCAmelCase =0.8 __UpperCAmelCase =10 __UpperCAmelCase =0.3 __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =max_length __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Optional[int] ) -> Any: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =max_length __UpperCAmelCase =2 __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : List[str] ) -> Dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =False __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Dict ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =True __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Dict ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =2 __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : int ) -> Any: __UpperCAmelCase =AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-bert""" ) __UpperCAmelCase =FlaxAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) __UpperCAmelCase ="""Hello world""" __UpperCAmelCase =tokenizer(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(__SCREAMING_SNAKE_CASE , """do_samples""" ): model.generate(__SCREAMING_SNAKE_CASE , do_samples=__SCREAMING_SNAKE_CASE ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(__SCREAMING_SNAKE_CASE , """foo""" ): __UpperCAmelCase ={"""foo""": """bar"""} model.generate(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE )
68
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
560
from __future__ import annotations from collections.abc import Iterator class _A : """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : int ) -> None: __UpperCAmelCase =value __UpperCAmelCase =None __UpperCAmelCase =None class _A : """simple docstring""" def __init__( self : List[Any] , __SCREAMING_SNAKE_CASE : Node ) -> None: __UpperCAmelCase =tree def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Node | None ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : int ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
68
0
'''simple docstring''' from typing import List from .keymap import KEYMAP, get_character def __snake_case ( SCREAMING_SNAKE_CASE_ : str ) -> str: """simple docstring""" def decorator(SCREAMING_SNAKE_CASE_ : int ): UpperCAmelCase = getattr(A_ , '''handle_key''' , [] ) handle += [key] setattr(A_ , '''handle_key''' , A_ ) return func return decorator def __snake_case ( *SCREAMING_SNAKE_CASE_ : List[str] ) -> Optional[int]: """simple docstring""" def decorator(SCREAMING_SNAKE_CASE_ : Tuple ): UpperCAmelCase = getattr(A_ , '''handle_key''' , [] ) handle += keys setattr(A_ , '''handle_key''' , A_ ) return func return decorator class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' def __new__( cls : str , a__ : Tuple , a__ : Tuple , a__ : List[str] ): UpperCAmelCase = super().__new__(cls , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if not hasattr(__SCREAMING_SNAKE_CASE , '''key_handler''' ): setattr(__SCREAMING_SNAKE_CASE , '''key_handler''' , {} ) setattr(__SCREAMING_SNAKE_CASE , '''handle_input''' , KeyHandler.handle_input ) for value in attrs.values(): UpperCAmelCase = getattr(__SCREAMING_SNAKE_CASE , '''handle_key''' , [] ) for key in handled_keys: UpperCAmelCase = value return new_cls @staticmethod def __snake_case ( cls : Dict ): UpperCAmelCase = get_character() if char != KEYMAP["undefined"]: UpperCAmelCase = ord(__SCREAMING_SNAKE_CASE ) UpperCAmelCase = cls.key_handler.get(__SCREAMING_SNAKE_CASE ) if handler: UpperCAmelCase = char return handler(cls ) else: return None def __snake_case ( cls : str ) -> int: """simple docstring""" return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
51
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def lowercase__ ( A_: Union[str, Any] ) -> List[Any]: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ {"""Effect""": """Allow""", """Principal""": {"""Service""": """sagemaker.amazonaws.com"""}, """Action""": """sts:AssumeRole"""} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=A_ , AssumeRolePolicyDocument=json.dumps(A_ , indent=2 ) ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ { """Effect""": """Allow""", """Action""": [ """sagemaker:*""", """ecr:GetDownloadUrlForLayer""", """ecr:BatchGetImage""", """ecr:BatchCheckLayerAvailability""", """ecr:GetAuthorizationToken""", """cloudwatch:PutMetricData""", """cloudwatch:GetMetricData""", """cloudwatch:GetMetricStatistics""", """cloudwatch:ListMetrics""", """logs:CreateLogGroup""", """logs:CreateLogStream""", """logs:DescribeLogStreams""", """logs:PutLogEvents""", """logs:GetLogEvents""", """s3:CreateBucket""", """s3:ListBucket""", """s3:GetBucketLocation""", """s3:GetObject""", """s3:PutObject""", ], """Resource""": """*""", } ], } # attach policy to role iam_client.put_role_policy( RoleName=A_ , PolicyName=F'''{role_name}_policy_permission''' , PolicyDocument=json.dumps(A_ , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(F'''role {role_name} already exists. Using existing one''' ) def lowercase__ ( A_: Dict ) -> Any: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) return iam_client.get_role(RoleName=A_ )["Role"]["Arn"] def lowercase__ ( ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase =_ask_options( """How do you want to authorize?""" , ["""AWS Profile""", """Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) """] , A_ , ) __UpperCAmelCase =None if credentials_configuration == 0: __UpperCAmelCase =_ask_field("""Enter your AWS Profile name: [default] """ , default="""default""" ) __UpperCAmelCase =aws_profile else: print( """Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,""" """`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`""" ) __UpperCAmelCase =_ask_field("""AWS Access Key ID: """ ) __UpperCAmelCase =aws_access_key_id __UpperCAmelCase =_ask_field("""AWS Secret Access Key: """ ) __UpperCAmelCase =aws_secret_access_key __UpperCAmelCase =_ask_field("""Enter your AWS Region: [us-east-1]""" , default="""us-east-1""" ) __UpperCAmelCase =aws_region __UpperCAmelCase =_ask_options( """Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?""" , ["""Provide IAM Role name""", """Create new IAM role using credentials"""] , A_ , ) if role_management == 0: __UpperCAmelCase =_ask_field("""Enter your IAM role name: """ ) else: __UpperCAmelCase ="""accelerate_sagemaker_execution_role""" print(F'''Accelerate will create an iam role "{iam_role_name}" using the provided credentials''' ) _create_iam_role_for_sagemaker(A_ ) __UpperCAmelCase =_ask_field( """Do you want to use custom Docker image? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_custom_docker_image: __UpperCAmelCase =_ask_field("""Enter your Docker image: """ , lambda A_ : str(A_ ).lower() ) __UpperCAmelCase =_ask_field( """Do you want to provide SageMaker input channels with data locations? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_inputs_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_field( """Do you want to enable SageMaker metrics? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_metrics_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_options( """What is the distributed mode?""" , ["""No distributed training""", """Data parallelism"""] , _convert_sagemaker_distributed_mode , ) __UpperCAmelCase ={} __UpperCAmelCase =_ask_field( """Do you wish to optimize your script with torch dynamo?[yes/NO]:""" , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_dynamo: __UpperCAmelCase ="""dynamo_""" __UpperCAmelCase =_ask_options( """Which dynamo backend would you like to use?""" , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) __UpperCAmelCase =_ask_field( """Do you want to customize the defaults sent to torch.compile? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_custom_options: __UpperCAmelCase =_ask_options( """Which mode do you want to use?""" , A_ , lambda A_ : TORCH_DYNAMO_MODES[int(A_ )] , default="""default""" , ) __UpperCAmelCase =_ask_field( """Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =_ask_field( """Do you want to enable dynamic shape tracing? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase ="""Which EC2 instance type you want to use for your training?""" if distributed_type != SageMakerDistributedType.NO: __UpperCAmelCase =_ask_options( A_ , A_ , lambda A_ : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(A_ )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" __UpperCAmelCase =_ask_field(A_ , lambda A_ : str(A_ ).lower() , default="""ml.p3.2xlarge""" ) __UpperCAmelCase =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): __UpperCAmelCase =_ask_field( """How many machines do you want use? [1]: """ , A_ , default=1 , ) __UpperCAmelCase =_ask_options( """Do you wish to use FP16 or BF16 (mixed precision)?""" , ["""no""", """fp16""", """bf16""", """fp8"""] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( """Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.""" ) return SageMakerConfig( image_uri=A_ , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=A_ , use_cpu=A_ , dynamo_config=A_ , eca_instance_type=A_ , profile=A_ , region=A_ , iam_role_name=A_ , mixed_precision=A_ , num_machines=A_ , sagemaker_inputs_file=A_ , sagemaker_metrics_file=A_ , )
68
0
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ = { '''configuration_informer''': [ '''INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''InformerConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''InformerForPrediction''', '''InformerModel''', '''InformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
657
from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = {"ctrl": "https://huggingface.co/ctrl/resolve/main/config.json"} class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Tuple = 'ctrl' lowerCamelCase : Any = ['past_key_values'] lowerCamelCase : Optional[int] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=246534 , __SCREAMING_SNAKE_CASE : int=256 , __SCREAMING_SNAKE_CASE : Optional[Any]=1280 , __SCREAMING_SNAKE_CASE : Optional[Any]=8192 , __SCREAMING_SNAKE_CASE : int=48 , __SCREAMING_SNAKE_CASE : Union[str, Any]=16 , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : Dict=0.1 , __SCREAMING_SNAKE_CASE : List[Any]=1e-6 , __SCREAMING_SNAKE_CASE : List[str]=0.02 , __SCREAMING_SNAKE_CASE : Union[str, Any]=True , **__SCREAMING_SNAKE_CASE : int , ) -> Any: __UpperCAmelCase =vocab_size __UpperCAmelCase =n_positions __UpperCAmelCase =n_embd __UpperCAmelCase =n_layer __UpperCAmelCase =n_head __UpperCAmelCase =dff __UpperCAmelCase =resid_pdrop __UpperCAmelCase =embd_pdrop __UpperCAmelCase =layer_norm_epsilon __UpperCAmelCase =initializer_range __UpperCAmelCase =use_cache super().__init__(**__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) a_ = { 'configuration_llama': ['LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'LlamaConfig'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['LlamaTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = ['LlamaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ = [ 'LlamaForCausalLM', 'LlamaModel', 'LlamaPreTrainedModel', 'LlamaForSequenceClassification', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys a_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
296
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowercase__ ( A_: Optional[Any] ) -> Union[str, Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase =k.replace(A_ , A_ ) if k.startswith("""encoder""" ): __UpperCAmelCase =k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase =k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase =k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm3""" , """final_layer_norm""" ) return k def lowercase__ ( A_: Tuple ) -> str: """simple docstring""" __UpperCAmelCase =[ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase =sd.pop(A_ ) __UpperCAmelCase =k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase =v __A = ["START"] @torch.no_grad() def lowercase__ ( A_: List[Any] , A_: str , A_: int ) -> Optional[int]: """simple docstring""" __UpperCAmelCase =torch.load(A_ , map_location="""cpu""" ) __UpperCAmelCase =model["""model"""] __UpperCAmelCase =BlenderbotConfig.from_json_file(A_ ) __UpperCAmelCase =BlenderbotForConditionalGeneration(A_ ) __UpperCAmelCase =m.model.state_dict().keys() __UpperCAmelCase =[] __UpperCAmelCase ={} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase =rename_state_dict_key(A_ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase =v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(A_ ) m.model.load_state_dict(A_ , strict=A_ ) m.half() m.save_pretrained(A_ ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) __A = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
68
0
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( lowerCamelCase_: int = 3 , lowerCamelCase_: int = 7 , lowerCamelCase_: int = 1_0_0_0_0_0_0 ): """simple docstring""" snake_case : List[Any] = 0 snake_case : Dict = 1 for current_denominator in range(1 , limit + 1 ): snake_case : Tuple = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: snake_case : Optional[int] = current_numerator snake_case : Tuple = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_0_0_0_0_0_0))
449
from itertools import permutations def lowercase__ ( A_: tuple ) -> bool: """simple docstring""" if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False __UpperCAmelCase =[7, 11, 13, 17] for i, test in enumerate(A_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowercase__ ( A_: int = 10 ) -> int: """simple docstring""" return sum( int("""""".join(map(A_ , A_ ) ) ) for num in permutations(range(A_ ) ) if is_substring_divisible(A_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
68
0
'''simple docstring''' import argparse import copy def __UpperCamelCase ( a : Tuple ) ->Dict: snake_case = {} with open(A_ ) as f: for line in f: if line.split()[0] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[1], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[0]].append( [line.split()[1], line.split()[2]] ) if line.split()[1] not in dict_of_neighbours: snake_case = [] _list.append([line.split()[0], line.split()[2]] ) snake_case = _list else: dict_of_neighbours[line.split()[1]].append( [line.split()[0], line.split()[2]] ) return dict_of_neighbours def __UpperCamelCase ( a : str , a : List[str] ) ->Optional[Any]: with open(A_ ) as f: snake_case = f.read(1 ) snake_case = start_node snake_case = [] snake_case = start_node snake_case = 0 while visiting not in first_solution: snake_case = 1_0000 for k in dict_of_neighbours[visiting]: if int(k[1] ) < int(A_ ) and k[0] not in first_solution: snake_case = k[1] snake_case = k[0] first_solution.append(A_ ) snake_case = distance_of_first_solution + int(A_ ) snake_case = best_node first_solution.append(A_ ) snake_case = 0 for k in dict_of_neighbours[first_solution[-2]]: if k[0] == start_node: break position += 1 snake_case = ( distance_of_first_solution + int(dict_of_neighbours[first_solution[-2]][position][1] ) - 1_0000 ) return first_solution, distance_of_first_solution def __UpperCamelCase ( a : Any , a : Optional[Any] ) ->str: snake_case = [] for n in solution[1:-1]: snake_case = solution.index(A_ ) for kn in solution[1:-1]: snake_case = solution.index(A_ ) if n == kn: continue snake_case = copy.deepcopy(A_ ) snake_case = kn snake_case = n snake_case = 0 for k in _tmp[:-1]: snake_case = _tmp[_tmp.index(A_ ) + 1] for i in dict_of_neighbours[k]: if i[0] == next_node: snake_case = distance + int(i[1] ) _tmp.append(A_ ) if _tmp not in neighborhood_of_solution: neighborhood_of_solution.append(_tmp ) snake_case = len(neighborhood_of_solution[0] ) - 1 neighborhood_of_solution.sort(key=lambda a : x[index_of_last_item_in_the_list] ) return neighborhood_of_solution def __UpperCamelCase ( a : List[Any] , a : List[str] , a : Union[str, Any] , a : Optional[Any] , a : Optional[Any] ) ->List[Any]: snake_case = 1 snake_case = first_solution snake_case = [] snake_case = distance_of_first_solution snake_case = solution while count <= iters: snake_case = find_neighborhood(A_ , A_ ) snake_case = 0 snake_case = neighborhood[index_of_best_solution] snake_case = len(A_ ) - 1 snake_case = False while not found: snake_case = 0 while i < len(A_ ): if best_solution[i] != solution[i]: snake_case = best_solution[i] snake_case = solution[i] break snake_case = i + 1 if [first_exchange_node, second_exchange_node] not in tabu_list and [ second_exchange_node, first_exchange_node, ] not in tabu_list: tabu_list.append([first_exchange_node, second_exchange_node] ) snake_case = True snake_case = best_solution[:-1] snake_case = neighborhood[index_of_best_solution][best_cost_index] if cost < best_cost: snake_case = cost snake_case = solution else: snake_case = index_of_best_solution + 1 snake_case = neighborhood[index_of_best_solution] if len(A_ ) >= size: tabu_list.pop(0 ) snake_case = count + 1 return best_solution_ever, best_cost def __UpperCamelCase ( a : Union[str, Any]=None ) ->Tuple: snake_case = generate_neighbours(args.File ) snake_case , snake_case = generate_first_solution( args.File , A_ ) snake_case , snake_case = tabu_search( A_ , A_ , A_ , args.Iterations , args.Size , ) print(f"""Best solution: {best_sol}, with total distance: {best_cost}.""" ) if __name__ == "__main__": _lowercase = argparse.ArgumentParser(description='Tabu Search') parser.add_argument( '-f', '--File', type=str, help='Path to the file containing the data', required=True, ) parser.add_argument( '-i', '--Iterations', type=int, help='How many iterations the algorithm should perform', required=True, ) parser.add_argument( '-s', '--Size', type=int, help='Size of the tabu list', required=True ) # Pass the arguments to main method main(parser.parse_args())
342
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar __A = TypeVar("T") def lowercase__ ( A_: int ) -> int: """simple docstring""" return (position - 1) // 2 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 1 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 2 class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[str] ) -> None: __UpperCAmelCase =[] __UpperCAmelCase ={} __UpperCAmelCase =0 def __len__( self : str ) -> int: return self.elements def __repr__( self : Dict ) -> str: return str(self.heap ) def _a ( self : Optional[int] ) -> bool: # Check if the priority queue is empty return self.elements == 0 def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an element with given priority to the queue self.heap.append((elem, weight) ) __UpperCAmelCase =self.elements self.elements += 1 self._bubble_up(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> T: # Remove and return the element with lowest weight (highest priority) if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) __UpperCAmelCase , __UpperCAmelCase =self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: __UpperCAmelCase , __UpperCAmelCase =self.heap[0] self._bubble_down(__SCREAMING_SNAKE_CASE ) return elem def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Update the weight of the given key __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase =(elem, weight) if position > 0: __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._bubble_up(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (upward movement) [to be used internally # only] __UpperCAmelCase =self.position_map[elem] if curr_pos == 0: return None __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_up(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (downward movement) [to be used # internally only] __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase =get_child_left_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =get_child_right_position(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements and child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) else: return None if child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int ) -> None: # Swap the nodes at the given positions __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase , __UpperCAmelCase =( self.heap[nodea_pos], self.heap[nodea_pos], ) __UpperCAmelCase =nodea_pos __UpperCAmelCase =nodea_pos class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[Any] ) -> None: __UpperCAmelCase ={} __UpperCAmelCase =0 def __repr__( self : Tuple ) -> str: return str(self.connections ) def __len__( self : str ) -> int: return self.nodes def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : T ) -> None: # Add a node in the graph if it is not in the graph if node not in self.connections: __UpperCAmelCase ={} self.nodes += 1 def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an edge between 2 nodes in the graph self.add_node(__SCREAMING_SNAKE_CASE ) self.add_node(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =weight __UpperCAmelCase =weight def lowercase__ ( A_: GraphUndirectedWeighted[T] , ) -> tuple[dict[T, int], dict[T, T | None]]: """simple docstring""" __UpperCAmelCase ={node: maxsize for node in graph.connections} __UpperCAmelCase ={node: None for node in graph.connections} __UpperCAmelCase =MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(A_ , A_ ) if priority_queue.is_empty(): return dist, parent # initialization __UpperCAmelCase =priority_queue.extract_min() __UpperCAmelCase =0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node # running prim's algorithm while not priority_queue.is_empty(): __UpperCAmelCase =priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node return dist, parent
68
0
import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig __lowerCamelCase : str = logging.get_logger(__name__) __lowerCamelCase : Dict = { "Intel/dpt-large": "https://huggingface.co/Intel/dpt-large/resolve/main/config.json", # See all DPT models at https://huggingface.co/models?filter=dpt } class __magic_name__ ( A__ ): lowercase : Optional[Any] ='dpt' def __init__( self : List[Any] , UpperCamelCase__ : str=7_68 , UpperCamelCase__ : Optional[int]=12 , UpperCamelCase__ : int=12 , UpperCamelCase__ : Any=30_72 , UpperCamelCase__ : int="gelu" , UpperCamelCase__ : Any=0.0 , UpperCamelCase__ : Any=0.0 , UpperCamelCase__ : Dict=0.02 , UpperCamelCase__ : List[str]=1e-1_2 , UpperCamelCase__ : Optional[Any]=3_84 , UpperCamelCase__ : List[str]=16 , UpperCamelCase__ : int=3 , UpperCamelCase__ : Dict=False , UpperCamelCase__ : List[Any]=True , UpperCamelCase__ : Any=[2, 5, 8, 11] , UpperCamelCase__ : Dict="project" , UpperCamelCase__ : int=[4, 2, 1, 0.5] , UpperCamelCase__ : Union[str, Any]=[96, 1_92, 3_84, 7_68] , UpperCamelCase__ : str=2_56 , UpperCamelCase__ : int=-1 , UpperCamelCase__ : List[str]=False , UpperCamelCase__ : Tuple=True , UpperCamelCase__ : str=0.4 , UpperCamelCase__ : Any=2_55 , UpperCamelCase__ : Tuple=0.1 , UpperCamelCase__ : Union[str, Any]=[1, 10_24, 24, 24] , UpperCamelCase__ : Tuple=[0, 1] , UpperCamelCase__ : Union[str, Any]=None , **UpperCamelCase__ : List[Any] , ) -> Optional[int]: '''simple docstring''' super().__init__(**__SCREAMING_SNAKE_CASE ) UpperCAmelCase = hidden_size UpperCAmelCase = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info("Initializing the config with a `BiT` backbone." ) UpperCAmelCase = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, } UpperCAmelCase = BitConfig(**__SCREAMING_SNAKE_CASE ) elif isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): logger.info("Initializing the config with a `BiT` backbone." ) UpperCAmelCase = BitConfig(**__SCREAMING_SNAKE_CASE ) elif isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): UpperCAmelCase = backbone_config else: raise ValueError( F'backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}.' ) UpperCAmelCase = backbone_featmap_shape UpperCAmelCase = neck_ignore_stages if readout_type != "project": raise ValueError("Readout type must be 'project' when using `DPT-hybrid` mode." ) else: UpperCAmelCase = None UpperCAmelCase = None UpperCAmelCase = [] UpperCAmelCase = num_hidden_layers UpperCAmelCase = num_attention_heads UpperCAmelCase = intermediate_size UpperCAmelCase = hidden_act UpperCAmelCase = hidden_dropout_prob UpperCAmelCase = attention_probs_dropout_prob UpperCAmelCase = initializer_range UpperCAmelCase = layer_norm_eps UpperCAmelCase = image_size UpperCAmelCase = patch_size UpperCAmelCase = num_channels UpperCAmelCase = qkv_bias UpperCAmelCase = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError("Readout_type must be one of ['ignore', 'add', 'project']" ) UpperCAmelCase = readout_type UpperCAmelCase = reassemble_factors UpperCAmelCase = neck_hidden_sizes UpperCAmelCase = fusion_hidden_size UpperCAmelCase = head_in_index UpperCAmelCase = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) UpperCAmelCase = use_auxiliary_head UpperCAmelCase = auxiliary_loss_weight UpperCAmelCase = semantic_loss_ignore_index UpperCAmelCase = semantic_classifier_dropout def SCREAMING_SNAKE_CASE_ ( self : str ) -> Any: '''simple docstring''' UpperCAmelCase = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: UpperCAmelCase = self.backbone_config.to_dict() UpperCAmelCase = self.__class__.model_type return output
323
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __A = logging.get_logger(__name__) @dataclass class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[int] = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : Any , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Dict: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase =deprecated_arg[3:] __UpperCAmelCase =not kwargs.pop(__SCREAMING_SNAKE_CASE ) logger.warning( f'''{deprecated_arg} is depreciated. Please use --no-{positive_arg} or''' f''' {positive_arg}={kwargs[positive_arg]}''' ) __UpperCAmelCase =kwargs.pop("""tpu_name""" , self.tpu_name ) __UpperCAmelCase =kwargs.pop("""device_idx""" , self.device_idx ) __UpperCAmelCase =kwargs.pop("""eager_mode""" , self.eager_mode ) __UpperCAmelCase =kwargs.pop("""use_xla""" , self.use_xla ) super().__init__(**__SCREAMING_SNAKE_CASE ) lowerCamelCase : str = field( default=UpperCamelCase , metadata={'help': 'Name of TPU'} , ) lowerCamelCase : int = field( default=0 , metadata={'help': 'CPU / GPU device index. Defaults to 0.'} , ) lowerCamelCase : bool = field(default=UpperCamelCase , metadata={'help': 'Benchmark models in eager model.'} ) lowerCamelCase : bool = field( default=UpperCamelCase , metadata={ 'help': 'Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.' } , ) @cached_property def _a ( self : List[str] ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) __UpperCAmelCase =None if self.tpu: try: if self.tpu_name: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: __UpperCAmelCase =None return tpu @cached_property def _a ( self : Tuple ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) __UpperCAmelCase =tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , """GPU""" ) __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/gpu:{self.device_idx}''' ) else: tf.config.set_visible_devices([] , """GPU""" ) # disable GPU __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/cpu:{self.device_idx}''' ) return strategy @property def _a ( self : Optional[Any] ) -> bool: requires_backends(self , ["""tf"""] ) return self._setup_tpu is not None @property def _a ( self : str ) -> "tf.distribute.Strategy": requires_backends(self , ["""tf"""] ) return self._setup_strategy @property def _a ( self : Dict ) -> Optional[int]: requires_backends(self , ["""tf"""] ) return tf.config.list_physical_devices("""GPU""" ) @property def _a ( self : List[str] ) -> int: requires_backends(self , ["""tf"""] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _a ( self : List[str] ) -> bool: return self.n_gpu > 0
68
0
'''simple docstring''' import math import sys def lowerCAmelCase ( UpperCamelCase__ : str ): """simple docstring""" __UpperCAmelCase = '''''' try: with open(A_ , '''rb''' ) as binary_file: __UpperCAmelCase = binary_file.read() for dat in data: __UpperCAmelCase = f"""{dat:08b}""" result += curr_byte return result except OSError: print('''File not accessible''' ) sys.exit() def lowerCAmelCase ( UpperCamelCase__ : str ): """simple docstring""" __UpperCAmelCase = {'''0''': '''0''', '''1''': '''1'''} __UpperCAmelCase , __UpperCAmelCase = '''''', '''''' __UpperCAmelCase = len(A_ ) for i in range(len(A_ ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue __UpperCAmelCase = lexicon[curr_string] result += last_match_id __UpperCAmelCase = last_match_id + '''0''' if math.loga(A_ ).is_integer(): __UpperCAmelCase = {} for curr_key in list(A_ ): __UpperCAmelCase = lexicon.pop(A_ ) __UpperCAmelCase = new_lex __UpperCAmelCase = last_match_id + '''1''' index += 1 __UpperCAmelCase = '''''' return result def lowerCAmelCase ( UpperCamelCase__ : str , UpperCamelCase__ : str ): """simple docstring""" __UpperCAmelCase = 8 try: with open(A_ , '''wb''' ) as opened_file: __UpperCAmelCase = [ to_write[i : i + byte_length] for i in range(0 , len(A_ ) , A_ ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append('''10000000''' ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(A_ , 2 ).to_bytes(1 , byteorder='''big''' ) ) except OSError: print('''File not accessible''' ) sys.exit() def lowerCAmelCase ( UpperCamelCase__ : str ): """simple docstring""" __UpperCAmelCase = 0 for letter in data_bits: if letter == "1": break counter += 1 __UpperCAmelCase = data_bits[counter:] __UpperCAmelCase = data_bits[counter + 1 :] return data_bits def lowerCAmelCase ( UpperCamelCase__ : str , UpperCamelCase__ : str ): """simple docstring""" __UpperCAmelCase = read_file_binary(A_ ) __UpperCAmelCase = remove_prefix(A_ ) __UpperCAmelCase = decompress_data(A_ ) write_file_binary(A_ , A_ ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
262
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Dict: torch.manual_seed(0 ) __UpperCAmelCase =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def _a ( self : int ) -> Union[str, Any]: __UpperCAmelCase =self.dummy_uncond_unet __UpperCAmelCase =ScoreSdeVeScheduler() __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE )[ 0 ] __UpperCAmelCase =image[0, -3:, -3:, -1] __UpperCAmelCase =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCAmelCase =np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : Optional[int] ) -> int: __UpperCAmelCase ="""google/ncsnpp-church-256""" __UpperCAmelCase =UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVeScheduler.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=10 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCAmelCase =np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
68
0
"""simple docstring""" from typing import List, Optional, Union import torch from transformers import ( XLMRobertaTokenizer, ) from ...models import UNetaDConditionModel, VQModel from ...pipelines import DiffusionPipeline from ...pipelines.pipeline_utils import ImagePipelineOutput from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import ( is_accelerate_available, is_accelerate_version, logging, randn_tensor, replace_example_docstring, ) from .text_encoder import MultilingualCLIP __SCREAMING_SNAKE_CASE =logging.get_logger(__name__) # pylint: disable=invalid-name __SCREAMING_SNAKE_CASE ="\n Examples:\n ```py\n >>> from diffusers import KandinskyPipeline, KandinskyPriorPipeline\n >>> import torch\n\n >>> pipe_prior = KandinskyPriorPipeline.from_pretrained(\"kandinsky-community/Kandinsky-2-1-prior\")\n >>> pipe_prior.to(\"cuda\")\n\n >>> prompt = \"red cat, 4k photo\"\n >>> out = pipe_prior(prompt)\n >>> image_emb = out.image_embeds\n >>> negative_image_emb = out.negative_image_embeds\n\n >>> pipe = KandinskyPipeline.from_pretrained(\"kandinsky-community/kandinsky-2-1\")\n >>> pipe.to(\"cuda\")\n\n >>> image = pipe(\n ... prompt,\n ... image_embeds=image_emb,\n ... negative_image_embeds=negative_image_emb,\n ... height=768,\n ... width=768,\n ... num_inference_steps=100,\n ... ).images\n\n >>> image[0].save(\"cat.png\")\n ```\n" def lowercase__( __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : str=8 ): lowercase_ : List[Any] = h // scale_factor**2 if h % scale_factor**2 != 0: new_h += 1 lowercase_ : int = w // scale_factor**2 if w % scale_factor**2 != 0: new_w += 1 return new_h * scale_factor, new_w * scale_factor class UpperCamelCase ( lowercase_ ): def __init__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,) -> Optional[Any]: '''simple docstring''' super().__init__() self.register_modules( text_encoder=__SCREAMING_SNAKE_CASE ,tokenizer=__SCREAMING_SNAKE_CASE ,unet=__SCREAMING_SNAKE_CASE ,scheduler=__SCREAMING_SNAKE_CASE ,movq=__SCREAMING_SNAKE_CASE ,) lowercase_ : str = 2 ** (len(self.movq.config.block_out_channels ) - 1) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) -> List[str]: '''simple docstring''' if latents is None: lowercase_ : Union[str, Any] = randn_tensor(__SCREAMING_SNAKE_CASE ,generator=__SCREAMING_SNAKE_CASE ,device=__SCREAMING_SNAKE_CASE ,dtype=__SCREAMING_SNAKE_CASE ) else: if latents.shape != shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {shape}''' ) lowercase_ : List[str] = latents.to(__SCREAMING_SNAKE_CASE ) lowercase_ : Dict = latents * scheduler.init_noise_sigma return latents def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=None ,) -> Any: '''simple docstring''' lowercase_ : Optional[Any] = len(__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) else 1 # get prompt text embeddings lowercase_ : Any = self.tokenizer( __SCREAMING_SNAKE_CASE ,padding='max_length' ,truncation=__SCREAMING_SNAKE_CASE ,max_length=77 ,return_attention_mask=__SCREAMING_SNAKE_CASE ,add_special_tokens=__SCREAMING_SNAKE_CASE ,return_tensors='pt' ,) lowercase_ : List[str] = text_inputs.input_ids lowercase_ : List[Any] = self.tokenizer(__SCREAMING_SNAKE_CASE ,padding='longest' ,return_tensors='pt' ).input_ids if untruncated_ids.shape[-1] >= text_input_ids.shape[-1] and not torch.equal(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Dict = self.tokenizer.batch_decode(untruncated_ids[:, self.tokenizer.model_max_length - 1 : -1] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f''' {self.tokenizer.model_max_length} tokens: {removed_text}''' ) lowercase_ : List[Any] = text_input_ids.to(__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = text_inputs.attention_mask.to(__SCREAMING_SNAKE_CASE ) lowercase_ , lowercase_ : Tuple = self.text_encoder( input_ids=__SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = prompt_embeds.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) lowercase_ : int = text_encoder_hidden_states.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) lowercase_ : Any = text_mask.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) if do_classifier_free_guidance: lowercase_ : List[str] = 42 if negative_prompt is None: lowercase_ : Optional[Any] = [''] * batch_size elif type(__SCREAMING_SNAKE_CASE ) is not type(__SCREAMING_SNAKE_CASE ): raise TypeError( f'''`negative_prompt` should be the same type to `prompt`, but got {type(__SCREAMING_SNAKE_CASE )} !=''' f''' {type(__SCREAMING_SNAKE_CASE )}.''' ) elif isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Optional[Any] = [negative_prompt] elif batch_size != len(__SCREAMING_SNAKE_CASE ): raise ValueError( f'''`negative_prompt`: {negative_prompt} has batch size {len(__SCREAMING_SNAKE_CASE )}, but `prompt`:''' f''' {prompt} has batch size {batch_size}. Please make sure that passed `negative_prompt` matches''' ' the batch size of `prompt`.' ) else: lowercase_ : List[Any] = negative_prompt lowercase_ : Tuple = self.tokenizer( __SCREAMING_SNAKE_CASE ,padding='max_length' ,max_length=77 ,truncation=__SCREAMING_SNAKE_CASE ,return_attention_mask=__SCREAMING_SNAKE_CASE ,add_special_tokens=__SCREAMING_SNAKE_CASE ,return_tensors='pt' ,) lowercase_ : Optional[Any] = uncond_input.input_ids.to(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = uncond_input.attention_mask.to(__SCREAMING_SNAKE_CASE ) lowercase_ , lowercase_ : int = self.text_encoder( input_ids=__SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method lowercase_ : Union[str, Any] = negative_prompt_embeds.shape[1] lowercase_ : Tuple = negative_prompt_embeds.repeat(1 ,__SCREAMING_SNAKE_CASE ) lowercase_ : int = negative_prompt_embeds.view(batch_size * num_images_per_prompt ,__SCREAMING_SNAKE_CASE ) lowercase_ : Any = uncond_text_encoder_hidden_states.shape[1] lowercase_ : List[Any] = uncond_text_encoder_hidden_states.repeat(1 ,__SCREAMING_SNAKE_CASE ,1 ) lowercase_ : Any = uncond_text_encoder_hidden_states.view( batch_size * num_images_per_prompt ,__SCREAMING_SNAKE_CASE ,-1 ) lowercase_ : Optional[Any] = uncond_text_mask.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) # done duplicates # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes lowercase_ : Union[str, Any] = torch.cat([negative_prompt_embeds, prompt_embeds] ) lowercase_ : Dict = torch.cat([uncond_text_encoder_hidden_states, text_encoder_hidden_states] ) lowercase_ : Union[str, Any] = torch.cat([uncond_text_mask, text_mask] ) return prompt_embeds, text_encoder_hidden_states, text_mask def _UpperCAmelCase ( self ,__UpperCamelCase=0 ) -> Dict: '''simple docstring''' if is_accelerate_available(): from accelerate import cpu_offload else: raise ImportError('Please install accelerate via `pip install accelerate`' ) lowercase_ : Tuple = torch.device(f'''cuda:{gpu_id}''' ) lowercase_ : Tuple = [ self.unet, self.text_encoder, self.movq, ] for cpu_offloaded_model in models: if cpu_offloaded_model is not None: cpu_offload(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) def _UpperCAmelCase ( self ,__UpperCamelCase=0 ) -> Any: '''simple docstring''' if is_accelerate_available() and is_accelerate_version('>=' ,'0.17.0.dev0' ): from accelerate import cpu_offload_with_hook else: raise ImportError('`enable_model_cpu_offload` requires `accelerate v0.17.0` or higher.' ) lowercase_ : List[Any] = torch.device(f'''cuda:{gpu_id}''' ) if self.device.type != "cpu": self.to('cpu' ,silence_dtype_warnings=__SCREAMING_SNAKE_CASE ) torch.cuda.empty_cache() # otherwise we don't see the memory savings (but they probably exist) lowercase_ : List[str] = None for cpu_offloaded_model in [self.text_encoder, self.unet, self.movq]: lowercase_ , lowercase_ : Optional[int] = cpu_offload_with_hook(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,prev_module_hook=__SCREAMING_SNAKE_CASE ) if self.safety_checker is not None: lowercase_ , lowercase_ : Any = cpu_offload_with_hook(self.safety_checker ,__SCREAMING_SNAKE_CASE ,prev_module_hook=__SCREAMING_SNAKE_CASE ) # We'll offload the last model manually. lowercase_ : List[Any] = hook @property # Copied from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion.StableDiffusionPipeline._execution_device def _UpperCAmelCase ( self ) -> Dict: '''simple docstring''' if not hasattr(self.unet ,'_hf_hook' ): return self.device for module in self.unet.modules(): if ( hasattr(__SCREAMING_SNAKE_CASE ,'_hf_hook' ) and hasattr(module._hf_hook ,'execution_device' ) and module._hf_hook.execution_device is not None ): return torch.device(module._hf_hook.execution_device ) return self.device @torch.no_grad() @replace_example_docstring(__SCREAMING_SNAKE_CASE ) def __call__( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase = None ,__UpperCamelCase = 512 ,__UpperCamelCase = 512 ,__UpperCamelCase = 100 ,__UpperCamelCase = 4.0 ,__UpperCamelCase = 1 ,__UpperCamelCase = None ,__UpperCamelCase = None ,__UpperCamelCase = "pil" ,__UpperCamelCase = True ,) -> Optional[int]: '''simple docstring''' if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Optional[Any] = 1 elif isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Tuple = len(__SCREAMING_SNAKE_CASE ) else: raise ValueError(f'''`prompt` has to be of type `str` or `list` but is {type(__SCREAMING_SNAKE_CASE )}''' ) lowercase_ : str = self._execution_device lowercase_ : Union[str, Any] = batch_size * num_images_per_prompt lowercase_ : List[Any] = guidance_scale > 1.0 lowercase_ , lowercase_ , lowercase_ : Dict = self._encode_prompt( __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Optional[Any] = torch.cat(__SCREAMING_SNAKE_CASE ,dim=0 ) if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ): lowercase_ : Union[str, Any] = torch.cat(__SCREAMING_SNAKE_CASE ,dim=0 ) if do_classifier_free_guidance: lowercase_ : Optional[Any] = image_embeds.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) lowercase_ : Any = negative_image_embeds.repeat_interleave(__SCREAMING_SNAKE_CASE ,dim=0 ) lowercase_ : int = torch.cat([negative_image_embeds, image_embeds] ,dim=0 ).to( dtype=prompt_embeds.dtype ,device=__SCREAMING_SNAKE_CASE ) self.scheduler.set_timesteps(__SCREAMING_SNAKE_CASE ,device=__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = self.scheduler.timesteps lowercase_ : int = self.unet.config.in_channels lowercase_ , lowercase_ : List[Any] = get_new_h_w(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,self.movq_scale_factor ) # create initial latent lowercase_ : str = self.prepare_latents( (batch_size, num_channels_latents, height, width) ,text_encoder_hidden_states.dtype ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,self.scheduler ,) for i, t in enumerate(self.progress_bar(__SCREAMING_SNAKE_CASE ) ): # expand the latents if we are doing classifier free guidance lowercase_ : List[Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents lowercase_ : Optional[int] = {'text_embeds': prompt_embeds, 'image_embeds': image_embeds} lowercase_ : Any = self.unet( sample=__SCREAMING_SNAKE_CASE ,timestep=__SCREAMING_SNAKE_CASE ,encoder_hidden_states=__SCREAMING_SNAKE_CASE ,added_cond_kwargs=__SCREAMING_SNAKE_CASE ,return_dict=__SCREAMING_SNAKE_CASE ,)[0] if do_classifier_free_guidance: lowercase_ , lowercase_ : List[str] = noise_pred.split(latents.shape[1] ,dim=1 ) lowercase_ , lowercase_ : Optional[Any] = noise_pred.chunk(2 ) lowercase_ , lowercase_ : str = variance_pred.chunk(2 ) lowercase_ : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) lowercase_ : List[Any] = torch.cat([noise_pred, variance_pred_text] ,dim=1 ) if not ( hasattr(self.scheduler.config ,'variance_type' ) and self.scheduler.config.variance_type in ["learned", "learned_range"] ): lowercase_ , lowercase_ : List[str] = noise_pred.split(latents.shape[1] ,dim=1 ) # compute the previous noisy sample x_t -> x_t-1 lowercase_ : Optional[Any] = self.scheduler.step( __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,generator=__SCREAMING_SNAKE_CASE ,).prev_sample # post-processing lowercase_ : Any = self.movq.decode(__SCREAMING_SNAKE_CASE ,force_not_quantize=__SCREAMING_SNAKE_CASE )['sample'] if output_type not in ["pt", "np", "pil"]: raise ValueError(f'''Only the output types `pt`, `pil` and `np` are supported not output_type={output_type}''' ) if output_type in ["np", "pil"]: lowercase_ : str = image * 0.5 + 0.5 lowercase_ : Any = image.clamp(0 ,1 ) lowercase_ : Optional[Any] = image.cpu().permute(0 ,2 ,3 ,1 ).float().numpy() if output_type == "pil": lowercase_ : Dict = self.numpy_to_pil(__SCREAMING_SNAKE_CASE ) if not return_dict: return (image,) return ImagePipelineOutput(images=__SCREAMING_SNAKE_CASE )
425
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __A = logging.get_logger(__name__) __A = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, "constant": get_constant_schedule, "constant_w_warmup": get_constant_schedule_with_warmup, } class _A ( UpperCamelCase ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : str=None , *__SCREAMING_SNAKE_CASE : Union[str, Any] , **__SCREAMING_SNAKE_CASE : List[Any] ) -> Any: super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if config is None: assert isinstance(self.model , __SCREAMING_SNAKE_CASE ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f''' {self.model.__class__}''' ) __UpperCAmelCase =self.model.config else: __UpperCAmelCase =config __UpperCAmelCase =data_args __UpperCAmelCase =self.config.tgt_vocab_size if isinstance(self.config , __SCREAMING_SNAKE_CASE ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' """ padding..""" ) if self.args.label_smoothing == 0: __UpperCAmelCase =torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss __UpperCAmelCase =label_smoothed_nll_loss def _a ( self : Any , __SCREAMING_SNAKE_CASE : int ) -> Any: if self.optimizer is None: __UpperCAmelCase =["""bias""", """LayerNorm.weight"""] __UpperCAmelCase =[ { """params""": [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], """weight_decay""": self.args.weight_decay, }, { """params""": [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] __UpperCAmelCase =Adafactor if self.args.adafactor else AdamW if self.args.adafactor: __UpperCAmelCase =Adafactor __UpperCAmelCase ={"""scale_parameter""": False, """relative_step""": False} else: __UpperCAmelCase =AdamW __UpperCAmelCase ={ """betas""": (self.args.adam_betaa, self.args.adam_betaa), """eps""": self.args.adam_epsilon, } __UpperCAmelCase =self.args.learning_rate if self.sharded_ddp: __UpperCAmelCase =OSS( params=__SCREAMING_SNAKE_CASE , optim=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) else: __UpperCAmelCase =optimizer_cls(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if self.lr_scheduler is None: __UpperCAmelCase =self._get_lr_scheduler(__SCREAMING_SNAKE_CASE ) else: # ignoring --lr_scheduler logger.warning("""scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.""" ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[int] ) -> Any: __UpperCAmelCase =arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": __UpperCAmelCase =schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": __UpperCAmelCase =schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: __UpperCAmelCase =schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__SCREAMING_SNAKE_CASE ) return scheduler def _a ( self : Optional[Any] ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Tuple: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models __UpperCAmelCase , __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[:2] else: # compute label smoothed loss __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =torch.nn.functional.log_softmax(__SCREAMING_SNAKE_CASE , dim=-1 ) __UpperCAmelCase , __UpperCAmelCase =self.loss_fn(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] ) -> Dict: __UpperCAmelCase =inputs.pop("""labels""" ) __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return loss def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : nn.Module , __SCREAMING_SNAKE_CASE : Dict[str, Union[torch.Tensor, Any]] , __SCREAMING_SNAKE_CASE : bool , __SCREAMING_SNAKE_CASE : Optional[List[str]] = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: __UpperCAmelCase =self._prepare_inputs(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """max_length""": self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, """num_beams""": self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: __UpperCAmelCase =self.model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , **__SCREAMING_SNAKE_CASE , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) __UpperCAmelCase =inputs.pop("""labels""" ) with torch.no_grad(): # compute loss on predict data __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) __UpperCAmelCase =generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) return (loss, logits, labels) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : int ) -> List[Any]: # If PAD token is not defined at least EOS token has to be defined __UpperCAmelCase =self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( """Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be""" f''' padded to `max_length`={max_length}''' ) __UpperCAmelCase =pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) __UpperCAmelCase =tensor return padded_tensor
68
0
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL UpperCamelCase_ : List[str] = logging.get_logger(__name__) def A_ (__a ): '''simple docstring''' if isinstance(A_ , (list, tuple) ) and isinstance(videos[0] , (list, tuple) ) and is_valid_image(videos[0][0] ): return videos elif isinstance(A_ , (list, tuple) ) and is_valid_image(videos[0] ): return [videos] elif is_valid_image(A_ ): return [[videos]] raise ValueError(f'Could not make batched video from {videos}' ) class __lowerCAmelCase ( _lowercase ): """simple docstring""" snake_case = ['pixel_values'] def __init__( self : List[Any] , _snake_case : bool = True , _snake_case : Dict[str, int] = None , _snake_case : PILImageResampling = PILImageResampling.BILINEAR , _snake_case : bool = True , _snake_case : Dict[str, int] = None , _snake_case : bool = True , _snake_case : Union[int, float] = 1 / 255 , _snake_case : bool = True , _snake_case : bool = True , _snake_case : Optional[Union[float, List[float]]] = None , _snake_case : Optional[Union[float, List[float]]] = None , **_snake_case : Optional[int] , ) -> None: """simple docstring""" super().__init__(**__SCREAMING_SNAKE_CASE ) A_ = size if size is not None else {"shortest_edge": 256} A_ = get_size_dict(__SCREAMING_SNAKE_CASE , default_to_square=__SCREAMING_SNAKE_CASE ) A_ = crop_size if crop_size is not None else {"height": 224, "width": 224} A_ = get_size_dict(__SCREAMING_SNAKE_CASE , param_name="crop_size" ) A_ = do_resize A_ = size A_ = do_center_crop A_ = crop_size A_ = resample A_ = do_rescale A_ = rescale_factor A_ = offset A_ = do_normalize A_ = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN A_ = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCamelCase__ ( self : str , _snake_case : np.ndarray , _snake_case : Dict[str, int] , _snake_case : PILImageResampling = PILImageResampling.BILINEAR , _snake_case : Optional[Union[str, ChannelDimension]] = None , **_snake_case : Tuple , ) -> np.ndarray: """simple docstring""" A_ = get_size_dict(__SCREAMING_SNAKE_CASE , default_to_square=__SCREAMING_SNAKE_CASE ) if "shortest_edge" in size: A_ = get_resize_output_image_size(__SCREAMING_SNAKE_CASE , size["shortest_edge"] , default_to_square=__SCREAMING_SNAKE_CASE ) elif "height" in size and "width" in size: A_ = (size["height"], size["width"]) else: raise ValueError(F'Size must have \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}' ) return resize(__SCREAMING_SNAKE_CASE , size=__SCREAMING_SNAKE_CASE , resample=__SCREAMING_SNAKE_CASE , data_format=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : Dict , _snake_case : np.ndarray , _snake_case : Dict[str, int] , _snake_case : Optional[Union[str, ChannelDimension]] = None , **_snake_case : List[str] , ) -> np.ndarray: """simple docstring""" A_ = get_size_dict(__SCREAMING_SNAKE_CASE ) if "height" not in size or "width" not in size: raise ValueError(F'Size must have \'height\' and \'width\' as keys. Got {size.keys()}' ) return center_crop(__SCREAMING_SNAKE_CASE , size=(size["height"], size["width"]) , data_format=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : Any , _snake_case : np.ndarray , _snake_case : Union[int, float] , _snake_case : bool = True , _snake_case : Optional[Union[str, ChannelDimension]] = None , **_snake_case : Any , ) -> Optional[int]: """simple docstring""" A_ = image.astype(np.floataa ) if offset: A_ = image - (scale / 2) return rescale(__SCREAMING_SNAKE_CASE , scale=__SCREAMING_SNAKE_CASE , data_format=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : List[Any] , _snake_case : np.ndarray , _snake_case : Union[float, List[float]] , _snake_case : Union[float, List[float]] , _snake_case : Optional[Union[str, ChannelDimension]] = None , **_snake_case : Optional[Any] , ) -> np.ndarray: """simple docstring""" return normalize(__SCREAMING_SNAKE_CASE , mean=__SCREAMING_SNAKE_CASE , std=__SCREAMING_SNAKE_CASE , data_format=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : List[Any] , _snake_case : ImageInput , _snake_case : bool = None , _snake_case : Dict[str, int] = None , _snake_case : PILImageResampling = None , _snake_case : bool = None , _snake_case : Dict[str, int] = None , _snake_case : bool = None , _snake_case : float = None , _snake_case : bool = None , _snake_case : bool = None , _snake_case : Optional[Union[float, List[float]]] = None , _snake_case : Optional[Union[float, List[float]]] = None , _snake_case : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: """simple docstring""" if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) if offset and not do_rescale: raise ValueError("For offset, do_rescale must also be set to True." ) # All transformations expect numpy arrays. A_ = to_numpy_array(__SCREAMING_SNAKE_CASE ) if do_resize: A_ = self.resize(image=__SCREAMING_SNAKE_CASE , size=__SCREAMING_SNAKE_CASE , resample=__SCREAMING_SNAKE_CASE ) if do_center_crop: A_ = self.center_crop(__SCREAMING_SNAKE_CASE , size=__SCREAMING_SNAKE_CASE ) if do_rescale: A_ = self.rescale(image=__SCREAMING_SNAKE_CASE , scale=__SCREAMING_SNAKE_CASE , offset=__SCREAMING_SNAKE_CASE ) if do_normalize: A_ = self.normalize(image=__SCREAMING_SNAKE_CASE , mean=__SCREAMING_SNAKE_CASE , std=__SCREAMING_SNAKE_CASE ) A_ = to_channel_dimension_format(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return image def lowerCamelCase__ ( self : Any , _snake_case : ImageInput , _snake_case : bool = None , _snake_case : Dict[str, int] = None , _snake_case : PILImageResampling = None , _snake_case : bool = None , _snake_case : Dict[str, int] = None , _snake_case : bool = None , _snake_case : float = None , _snake_case : bool = None , _snake_case : bool = None , _snake_case : Optional[Union[float, List[float]]] = None , _snake_case : Optional[Union[float, List[float]]] = None , _snake_case : Optional[Union[str, TensorType]] = None , _snake_case : ChannelDimension = ChannelDimension.FIRST , **_snake_case : Optional[Any] , ) -> PIL.Image.Image: """simple docstring""" A_ = do_resize if do_resize is not None else self.do_resize A_ = resample if resample is not None else self.resample A_ = do_center_crop if do_center_crop is not None else self.do_center_crop A_ = do_rescale if do_rescale is not None else self.do_rescale A_ = rescale_factor if rescale_factor is not None else self.rescale_factor A_ = offset if offset is not None else self.offset A_ = do_normalize if do_normalize is not None else self.do_normalize A_ = image_mean if image_mean is not None else self.image_mean A_ = image_std if image_std is not None else self.image_std A_ = size if size is not None else self.size A_ = get_size_dict(__SCREAMING_SNAKE_CASE , default_to_square=__SCREAMING_SNAKE_CASE ) A_ = crop_size if crop_size is not None else self.crop_size A_ = get_size_dict(__SCREAMING_SNAKE_CASE , param_name="crop_size" ) if not valid_images(__SCREAMING_SNAKE_CASE ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) A_ = make_batched(__SCREAMING_SNAKE_CASE ) A_ = [ [ self._preprocess_image( image=__SCREAMING_SNAKE_CASE , do_resize=__SCREAMING_SNAKE_CASE , size=__SCREAMING_SNAKE_CASE , resample=__SCREAMING_SNAKE_CASE , do_center_crop=__SCREAMING_SNAKE_CASE , crop_size=__SCREAMING_SNAKE_CASE , do_rescale=__SCREAMING_SNAKE_CASE , rescale_factor=__SCREAMING_SNAKE_CASE , offset=__SCREAMING_SNAKE_CASE , do_normalize=__SCREAMING_SNAKE_CASE , image_mean=__SCREAMING_SNAKE_CASE , image_std=__SCREAMING_SNAKE_CASE , data_format=__SCREAMING_SNAKE_CASE , ) for img in video ] for video in videos ] A_ = {"pixel_values": videos} return BatchFeature(data=__SCREAMING_SNAKE_CASE , tensor_type=__SCREAMING_SNAKE_CASE )
115
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : List[Any] = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str=0 ) -> Any: __UpperCAmelCase =floats_tensor((1, 3, 128, 128) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =np.random.RandomState(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """strength""": 0.75, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _a ( self : Optional[Any] ) -> int: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Optional[Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) # warmup pass to apply optimizations __UpperCAmelCase =pipe(**self.get_dummy_inputs() ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Optional[int]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _a ( self : Dict ) -> int: __UpperCAmelCase =ort.SessionOptions() __UpperCAmelCase =False return options def _a ( self : Dict ) -> Any: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) # using the PNDM scheduler by default __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def _a ( self : List[str] ) -> str: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) __UpperCAmelCase =LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
68
0
from __future__ import annotations def lowerCamelCase_ ( lowerCAmelCase__ : list[int] ) -> bool: '''simple docstring''' return len(set(A_ ) ) == len(A_ ) if __name__ == "__main__": import doctest doctest.testmod()
106
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __A = logging.getLogger(__name__) class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[Any] = 'sequence-classification' def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Tuple ) -> Optional[Any]: if type(__SCREAMING_SNAKE_CASE ) == dict: __UpperCAmelCase =Namespace(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =glue_output_modes[hparams.task] __UpperCAmelCase =glue_tasks_num_labels[hparams.task] super().__init__(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.mode ) def _a ( self : str , **__SCREAMING_SNAKE_CASE : Dict ) -> List[str]: return self.model(**__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Dict ) -> List[Any]: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =outputs[0] __UpperCAmelCase =self.trainer.lr_schedulers[0]["""scheduler"""] __UpperCAmelCase ={"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def _a ( self : Tuple ) -> List[Any]: __UpperCAmelCase =self.hparams __UpperCAmelCase =processors[args.task]() __UpperCAmelCase =processor.get_labels() for mode in ["train", "dev"]: __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) if os.path.exists(__SCREAMING_SNAKE_CASE ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) __UpperCAmelCase =( processor.get_dev_examples(args.data_dir ) if mode == """dev""" else processor.get_train_examples(args.data_dir ) ) __UpperCAmelCase =convert_examples_to_features( __SCREAMING_SNAKE_CASE , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , __SCREAMING_SNAKE_CASE ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : bool = False ) -> DataLoader: __UpperCAmelCase ="""dev""" if mode == """test""" else mode __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.load(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([f.input_ids for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , batch_size=__SCREAMING_SNAKE_CASE , shuffle=__SCREAMING_SNAKE_CASE , ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : int ) -> str: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =outputs[:2] __UpperCAmelCase =logits.detach().cpu().numpy() __UpperCAmelCase =inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Any ) -> tuple: __UpperCAmelCase =torch.stack([x["""val_loss"""] for x in outputs] ).mean().detach().cpu().item() __UpperCAmelCase =np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =np.argmax(__SCREAMING_SNAKE_CASE , axis=1 ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =np.squeeze(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =np.concatenate([x["""target"""] for x in outputs] , axis=0 ) __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase ={**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )} __UpperCAmelCase =dict(results.items() ) __UpperCAmelCase =results return ret, preds_list, out_label_list def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : list ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _a ( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[Any]: BaseTransformer.add_model_specific_args(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) parser.add_argument( """--max_seq_length""" , default=128 , type=__SCREAMING_SNAKE_CASE , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=__SCREAMING_SNAKE_CASE , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser def lowercase__ ( ) -> str: """simple docstring""" __UpperCAmelCase =argparse.ArgumentParser() add_generic_args(A_ , os.getcwd() ) __UpperCAmelCase =GLUETransformer.add_model_specific_args(A_ , os.getcwd() ) __UpperCAmelCase =parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: __UpperCAmelCase =os.path.join( """./results""" , F'''{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}''' , ) os.makedirs(args.output_dir ) __UpperCAmelCase =GLUETransformer(A_ ) __UpperCAmelCase =generic_train(A_ , A_ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: __UpperCAmelCase =sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=A_ ) ) __UpperCAmelCase =model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(A_ ) if __name__ == "__main__": main()
68
0
"""simple docstring""" import argparse import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.local_sgd import LocalSGD ######################################################################## # This is a fully working simple example to use Accelerate # with LocalSGD, which is a method to synchronize model # parameters every K batches. It is different, but complementary # to gradient accumulation. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## lowerCAmelCase_ = 16 lowerCAmelCase_ = 32 def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase = 16 ) -> Union[str, Any]: lowercase__ : List[str] = AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowercase__ : Optional[int] = load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__lowerCamelCase ): # max_length=None => use the model max length (it's actually the default) lowercase__ : str = tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=A_ , max_length=A_ ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowercase__ : Union[str, Any] = datasets.map( A_ , batched=A_ , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__ : Tuple = tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__lowerCamelCase ): # On TPU it's best to pad everything to the same length or training will be very slow. lowercase__ : Union[str, Any] = 1_28 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowercase__ : List[Any] = 16 elif accelerator.mixed_precision != "no": lowercase__ : Optional[int] = 8 else: lowercase__ : Optional[int] = None return tokenizer.pad( A_ , padding='''longest''' , max_length=A_ , pad_to_multiple_of=A_ , return_tensors='''pt''' , ) # Instantiate dataloaders. lowercase__ : str = DataLoader( tokenized_datasets['''train'''] , shuffle=A_ , collate_fn=A_ , batch_size=A_ ) lowercase__ : Union[str, Any] = DataLoader( tokenized_datasets['''validation'''] , shuffle=A_ , collate_fn=A_ , batch_size=A_ ) return train_dataloader, eval_dataloader # For testing only if os.environ.get('TESTING_MOCKED_DATALOADERS', None) == "1": from accelerate.test_utils.training import mocked_dataloaders lowerCAmelCase_ = mocked_dataloaders # noqa: F811 def __UpperCAmelCase ( __lowerCamelCase , __lowerCamelCase ) -> List[str]: if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , A_ ) == "1": lowercase__ : Optional[Any] = 2 # New Code # lowercase__ : Tuple = int(args.gradient_accumulation_steps ) lowercase__ : Tuple = int(args.local_sgd_steps ) # Initialize accelerator lowercase__ : Optional[Any] = Accelerator( cpu=args.cpu , mixed_precision=args.mixed_precision , gradient_accumulation_steps=A_ ) if accelerator.distributed_type not in [DistributedType.NO, DistributedType.MULTI_CPU, DistributedType.MULTI_GPU]: raise NotImplementedError('''LocalSGD is supported only for CPUs and GPUs (no DeepSpeed or MegatronLM)''' ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__ : Tuple = config['''lr'''] lowercase__ : Optional[Any] = int(config['''num_epochs'''] ) lowercase__ : Tuple = int(config['''seed'''] ) lowercase__ : List[str] = int(config['''batch_size'''] ) lowercase__ : Optional[int] = evaluate.load('''glue''' , '''mrpc''' ) set_seed(A_ ) lowercase__ , lowercase__ : int = get_dataloaders(A_ , A_ ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__ : str = AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=A_ ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowercase__ : int = model.to(accelerator.device ) # Instantiate optimizer lowercase__ : Optional[Any] = AdamW(params=model.parameters() , lr=A_ ) # Instantiate scheduler lowercase__ : str = get_linear_schedule_with_warmup( optimizer=A_ , num_warmup_steps=1_00 , num_training_steps=(len(A_ ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : List[Any] = accelerator.prepare( A_ , A_ , A_ , A_ , A_ ) # Now we train the model for epoch in range(A_ ): model.train() with LocalSGD( accelerator=A_ , model=A_ , local_sgd_steps=A_ , enabled=local_sgd_steps is not None ) as local_sgd: for step, batch in enumerate(A_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) # New code # # We use the new `accumulate` context manager to perform gradient accumulation # We also currently do not support TPUs nor advise it as bugs were found on the XLA side when running our tests. with accelerator.accumulate(A_ ): lowercase__ : Any = model(**A_ ) lowercase__ : List[Any] = output.loss accelerator.backward(A_ ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() # LocalSGD-specific line local_sgd.step() model.eval() for step, batch in enumerate(A_ ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase__ : List[str] = model(**A_ ) lowercase__ : str = outputs.logits.argmax(dim=-1 ) lowercase__ , lowercase__ : Tuple = accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=A_ , references=A_ , ) lowercase__ : Optional[Any] = metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(f"""epoch {epoch}:""" , A_ ) def __UpperCAmelCase ( ) -> Tuple: lowercase__ : int = argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=A_ , default=A_ , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) # New Code # parser.add_argument( '''--gradient_accumulation_steps''' , type=A_ , default=1 , help='''The number of minibatches to be ran before gradients are accumulated.''' , ) parser.add_argument( '''--local_sgd_steps''' , type=A_ , default=8 , help='''Number of local SGD steps or None to disable local SGD''' ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) lowercase__ : List[str] = parser.parse_args() lowercase__ : List[str] = {'''lr''': 2E-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(A_ , A_ ) if __name__ == "__main__": main()
560
def lowercase__ ( A_: int , A_: int ) -> int: """simple docstring""" return 1 if input_a == input_a else 0 def lowercase__ ( ) -> None: """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
68
0
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a__ : List[str] = logging.get_logger(__name__) a__ : Dict = {'openai-gpt': 'https://huggingface.co/openai-gpt/resolve/main/config.json'} class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' _lowerCamelCase ='openai-gpt' _lowerCamelCase ={ 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : List[str] , a__ : str=40478 , a__ : Optional[Any]=512 , a__ : Optional[int]=768 , a__ : Dict=12 , a__ : Optional[Any]=12 , a__ : List[Any]="gelu" , a__ : str=0.1 , a__ : Optional[int]=0.1 , a__ : Optional[int]=0.1 , a__ : str=1e-5 , a__ : Tuple=0.02 , a__ : Tuple="cls_index" , a__ : Any=True , a__ : Optional[int]=None , a__ : Optional[Any]=True , a__ : int=0.1 , **a__ : List[str] , ): UpperCAmelCase = vocab_size UpperCAmelCase = n_positions UpperCAmelCase = n_embd UpperCAmelCase = n_layer UpperCAmelCase = n_head UpperCAmelCase = afn UpperCAmelCase = resid_pdrop UpperCAmelCase = embd_pdrop UpperCAmelCase = attn_pdrop UpperCAmelCase = layer_norm_epsilon UpperCAmelCase = initializer_range UpperCAmelCase = summary_type UpperCAmelCase = summary_use_proj UpperCAmelCase = summary_activation UpperCAmelCase = summary_first_dropout UpperCAmelCase = summary_proj_to_labels super().__init__(**__SCREAMING_SNAKE_CASE )
51
from __future__ import annotations import bisect def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] < item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] <= item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_left(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_right(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =0 __UpperCAmelCase =len(A_ ) - 1 while left <= right: __UpperCAmelCase =left + (right - left) // 2 __UpperCAmelCase =sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: __UpperCAmelCase =midpoint - 1 else: __UpperCAmelCase =midpoint + 1 return None def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =bisect.bisect_left(A_ , A_ ) if index != len(A_ ) and sorted_collection[index] == item: return index return None def lowercase__ ( A_: list[int] , A_: int , A_: int , A_: int ) -> int | None: """simple docstring""" if right < left: return None __UpperCAmelCase =left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(A_ , A_ , A_ , midpoint - 1 ) else: return binary_search_by_recursion(A_ , A_ , midpoint + 1 , A_ ) if __name__ == "__main__": __A = input("Enter numbers separated by comma:\n").strip() __A = sorted(int(item) for item in user_input.split(",")) __A = int(input("Enter a single number to be found in the list:\n")) __A = binary_search(collection, target) if result is None: print(F"""{target} was not found in {collection}.""") else: print(F"""{target} was found at position {result} in {collection}.""")
68
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __magic_name__ = {'''configuration_ibert''': ['''IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''IBertConfig''', '''IBertOnnxConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __magic_name__ = [ '''IBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''IBertForMaskedLM''', '''IBertForMultipleChoice''', '''IBertForQuestionAnswering''', '''IBertForSequenceClassification''', '''IBertForTokenClassification''', '''IBertModel''', '''IBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_ibert import IBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, IBertConfig, IBertOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_ibert import ( IBERT_PRETRAINED_MODEL_ARCHIVE_LIST, IBertForMaskedLM, IBertForMultipleChoice, IBertForQuestionAnswering, IBertForSequenceClassification, IBertForTokenClassification, IBertModel, IBertPreTrainedModel, ) else: import sys __magic_name__ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
657
from typing import List from .keymap import KEYMAP, get_character def lowercase__ ( A_: str ) -> str: """simple docstring""" def decorator(A_: int ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += [key] setattr(A_ , """handle_key""" , A_ ) return func return decorator def lowercase__ ( *A_: List[str] ) -> Optional[int]: """simple docstring""" def decorator(A_: Tuple ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += keys setattr(A_ , """handle_key""" , A_ ) return func return decorator class _A ( UpperCamelCase ): """simple docstring""" def __new__( cls : str , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> int: __UpperCAmelCase =super().__new__(cls , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if not hasattr(__SCREAMING_SNAKE_CASE , """key_handler""" ): setattr(__SCREAMING_SNAKE_CASE , """key_handler""" , {} ) setattr(__SCREAMING_SNAKE_CASE , """handle_input""" , KeyHandler.handle_input ) for value in attrs.values(): __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , """handle_key""" , [] ) for key in handled_keys: __UpperCAmelCase =value return new_cls @staticmethod def _a ( cls : Dict ) -> List[Any]: __UpperCAmelCase =get_character() if char != KEYMAP["undefined"]: __UpperCAmelCase =ord(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =cls.key_handler.get(__SCREAMING_SNAKE_CASE ) if handler: __UpperCAmelCase =char return handler(cls ) else: return None def lowercase__ ( cls: str ) -> int: """simple docstring""" return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
68
0
'''simple docstring''' from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS a_ = logging.get_logger(__name__) a_ = { 'linear': get_linear_schedule_with_warmup, 'cosine': get_cosine_schedule_with_warmup, 'cosine_w_restarts': get_cosine_with_hard_restarts_schedule_with_warmup, 'polynomial': get_polynomial_decay_schedule_with_warmup, 'constant': get_constant_schedule, 'constant_w_warmup': get_constant_schedule_with_warmup, } class __SCREAMING_SNAKE_CASE ( lowerCamelCase ): def __init__( self : List[str] , __lowercase : Union[str, Any]=None , __lowercase : str=None , *__lowercase : Union[str, Any] , **__lowercase : List[Any] ) -> Any: super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if config is None: assert isinstance(self.model , __SCREAMING_SNAKE_CASE ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" F" {self.model.__class__}" ) SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.model.config else: SCREAMING_SNAKE_CASE__ : Optional[int] =config SCREAMING_SNAKE_CASE__ : str =data_args SCREAMING_SNAKE_CASE__ : Optional[Any] =self.config.tgt_vocab_size if isinstance(self.config , __SCREAMING_SNAKE_CASE ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( F"The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for" ''' padding..''' ) if self.args.label_smoothing == 0: SCREAMING_SNAKE_CASE__ : List[Any] =torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss SCREAMING_SNAKE_CASE__ : str =label_smoothed_nll_loss def __magic_name__ ( self : Any , __lowercase : int ) -> Any: if self.optimizer is None: SCREAMING_SNAKE_CASE__ : Dict =['''bias''', '''LayerNorm.weight'''] SCREAMING_SNAKE_CASE__ : int =[ { '''params''': [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], '''weight_decay''': self.args.weight_decay, }, { '''params''': [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], '''weight_decay''': 0.0, }, ] SCREAMING_SNAKE_CASE__ : Optional[int] =Adafactor if self.args.adafactor else AdamW if self.args.adafactor: SCREAMING_SNAKE_CASE__ : Optional[int] =Adafactor SCREAMING_SNAKE_CASE__ : Optional[int] ={'''scale_parameter''': False, '''relative_step''': False} else: SCREAMING_SNAKE_CASE__ : List[str] =AdamW SCREAMING_SNAKE_CASE__ : int ={ '''betas''': (self.args.adam_betaa, self.args.adam_betaa), '''eps''': self.args.adam_epsilon, } SCREAMING_SNAKE_CASE__ : Any =self.args.learning_rate if self.sharded_ddp: SCREAMING_SNAKE_CASE__ : Optional[Any] =OSS( params=__SCREAMING_SNAKE_CASE , optim=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) else: SCREAMING_SNAKE_CASE__ : Dict =optimizer_cls(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if self.lr_scheduler is None: SCREAMING_SNAKE_CASE__ : Dict =self._get_lr_scheduler(__SCREAMING_SNAKE_CASE ) else: # ignoring --lr_scheduler logger.warning('''scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.''' ) def __magic_name__ ( self : Optional[Any] , __lowercase : Optional[int] ) -> Any: SCREAMING_SNAKE_CASE__ : Union[str, Any] =arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": SCREAMING_SNAKE_CASE__ : int =schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": SCREAMING_SNAKE_CASE__ : Optional[Any] =schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: SCREAMING_SNAKE_CASE__ : List[Any] =schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__SCREAMING_SNAKE_CASE ) return scheduler def __magic_name__ ( self : Optional[Any] ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def __magic_name__ ( self : List[Any] , __lowercase : List[str] , __lowercase : List[str] , __lowercase : Optional[Any] ) -> Tuple: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token SCREAMING_SNAKE_CASE__ : str =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] SCREAMING_SNAKE_CASE__ : Dict =self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Optional[Any] =model(**__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[:2] else: # compute label smoothed loss SCREAMING_SNAKE_CASE__ : int =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] SCREAMING_SNAKE_CASE__ : int =torch.nn.functional.log_softmax(__SCREAMING_SNAKE_CASE , dim=-1 ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.loss_fn(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def __magic_name__ ( self : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : List[Any] ) -> Dict: SCREAMING_SNAKE_CASE__ : Any =inputs.pop('''labels''' ) SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : List[Any] =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return loss def __magic_name__ ( self : List[str] , __lowercase : nn.Module , __lowercase : Dict[str, Union[torch.Tensor, Any]] , __lowercase : bool , __lowercase : Optional[List[str]] = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: SCREAMING_SNAKE_CASE__ : Dict =self._prepare_inputs(__SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ : List[Any] ={ '''max_length''': self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, '''num_beams''': self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: SCREAMING_SNAKE_CASE__ : int =self.model.generate( inputs['''input_ids'''] , attention_mask=inputs['''attention_mask'''] , **__SCREAMING_SNAKE_CASE , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ : Optional[Any] =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs['''max_length'''] ) SCREAMING_SNAKE_CASE__ : List[str] =inputs.pop('''labels''' ) with torch.no_grad(): # compute loss on predict data SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__ : Any =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) SCREAMING_SNAKE_CASE__ : Any =loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) SCREAMING_SNAKE_CASE__ : Optional[Any] =generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: SCREAMING_SNAKE_CASE__ : List[str] =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs['''max_length'''] ) return (loss, logits, labels) def __magic_name__ ( self : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : int ) -> List[Any]: # If PAD token is not defined at least EOS token has to be defined SCREAMING_SNAKE_CASE__ : Union[str, Any] =self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( '''Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be''' F" padded to `max_length`={max_length}" ) SCREAMING_SNAKE_CASE__ : Tuple =pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) SCREAMING_SNAKE_CASE__ : Dict =tensor return padded_tensor
296
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
68
0
"""simple docstring""" import os import unittest from transformers import BertTokenizerFast from transformers.models.bert.tokenization_bert import ( VOCAB_FILES_NAMES, BasicTokenizer, BertTokenizer, WordpieceTokenizer, _is_control, _is_punctuation, _is_whitespace, ) from transformers.testing_utils import require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin, filter_non_english @require_tokenizers class _a ( SCREAMING_SNAKE_CASE__ , unittest.TestCase): __magic_name__ = BertTokenizer __magic_name__ = BertTokenizerFast __magic_name__ = True __magic_name__ = True __magic_name__ = filter_non_english def __lowercase ( self : Dict ) -> str: super().setUp() snake_case : List[str] = [ "[UNK]", "[CLS]", "[SEP]", "[PAD]", "[MASK]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] snake_case : Dict = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def __lowercase ( self : Dict , _lowercase : Dict ) -> Tuple: snake_case : List[str] = "UNwant\u00E9d,running" snake_case : Optional[int] = "unwanted, running" return input_text, output_text def __lowercase ( self : List[str] ) -> Union[str, Any]: snake_case : List[str] = self.tokenizer_class(self.vocab_file ) snake_case : Optional[Any] = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(__SCREAMING_SNAKE_CASE , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(__SCREAMING_SNAKE_CASE ) , [9, 6, 7, 12, 10, 11] ) def __lowercase ( self : int ) -> Dict: if not self.test_rust_tokenizer: return snake_case : int = self.get_tokenizer() snake_case : Optional[Any] = self.get_rust_tokenizer() snake_case : Optional[int] = "UNwant\u00E9d,running" snake_case : Optional[Any] = tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) snake_case : List[Any] = rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case : int = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : List[str] = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case : Optional[Any] = self.get_rust_tokenizer() snake_case : Optional[Any] = tokenizer.encode(__SCREAMING_SNAKE_CASE ) snake_case : List[str] = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # With lower casing snake_case : List[str] = self.get_tokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) snake_case : int = self.get_rust_tokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = "UNwant\u00E9d,running" snake_case : str = tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) snake_case : Dict = rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case : Dict = tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : List[str] = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case : List[Any] = self.get_rust_tokenizer() snake_case : int = tokenizer.encode(__SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = rust_tokenizer.encode(__SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def __lowercase ( self : Union[str, Any] ) -> Dict: snake_case : str = BasicTokenizer() self.assertListEqual(tokenizer.tokenize("ah\u535A\u63A8zz" ) , ["ah", "\u535A", "\u63A8", "zz"] ) def __lowercase ( self : Dict ) -> List[Any]: snake_case : Dict = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["hello", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __lowercase ( self : Optional[Any] ) -> int: snake_case : Dict = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hällo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["h\u00E9llo"] ) def __lowercase ( self : Optional[int] ) -> Any: snake_case : List[str] = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __lowercase ( self : Dict ) -> int: snake_case : int = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["hallo", "!", "how", "are", "you", "?"] ) self.assertListEqual(tokenizer.tokenize("H\u00E9llo" ) , ["hello"] ) def __lowercase ( self : int ) -> Tuple: snake_case : Dict = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? " ) , ["HeLLo", "!", "how", "Are", "yoU", "?"] ) def __lowercase ( self : Optional[int] ) -> Optional[int]: snake_case : Optional[Any] = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HäLLo", "!", "how", "Are", "yoU", "?"] ) def __lowercase ( self : Any ) -> int: snake_case : Any = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE ) self.assertListEqual( tokenizer.tokenize(" \tHäLLo!how \n Are yoU? " ) , ["HaLLo", "!", "how", "Are", "yoU", "?"] ) def __lowercase ( self : List[Any] ) -> Optional[int]: snake_case : Dict = BasicTokenizer(do_lower_case=__SCREAMING_SNAKE_CASE , never_split=["[UNK]"] ) self.assertListEqual( tokenizer.tokenize(" \tHeLLo!how \n Are yoU? [UNK]" ) , ["HeLLo", "!", "how", "Are", "yoU", "?", "[UNK]"] ) def __lowercase ( self : int ) -> Any: snake_case : Dict = BasicTokenizer() snake_case : List[str] = "a\n'll !!to?'d of, can't." snake_case : List[Any] = ["a", "'", "ll", "!", "!", "to", "?", "'", "d", "of", ",", "can", "'", "t", "."] self.assertListEqual(tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) def __lowercase ( self : Optional[Any] ) -> Tuple: snake_case : Optional[int] = ["[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing"] snake_case : int = {} for i, token in enumerate(__SCREAMING_SNAKE_CASE ): snake_case : Dict = i snake_case : Tuple = WordpieceTokenizer(vocab=__SCREAMING_SNAKE_CASE , unk_token="[UNK]" ) self.assertListEqual(tokenizer.tokenize("" ) , [] ) self.assertListEqual(tokenizer.tokenize("unwanted running" ) , ["un", "##want", "##ed", "runn", "##ing"] ) self.assertListEqual(tokenizer.tokenize("unwantedX running" ) , ["[UNK]", "runn", "##ing"] ) def __lowercase ( self : Optional[Any] ) -> Dict: self.assertTrue(_is_whitespace(" " ) ) self.assertTrue(_is_whitespace("\t" ) ) self.assertTrue(_is_whitespace("\r" ) ) self.assertTrue(_is_whitespace("\n" ) ) self.assertTrue(_is_whitespace("\u00A0" ) ) self.assertFalse(_is_whitespace("A" ) ) self.assertFalse(_is_whitespace("-" ) ) def __lowercase ( self : List[Any] ) -> List[Any]: self.assertTrue(_is_control("\u0005" ) ) self.assertFalse(_is_control("A" ) ) self.assertFalse(_is_control(" " ) ) self.assertFalse(_is_control("\t" ) ) self.assertFalse(_is_control("\r" ) ) def __lowercase ( self : Optional[int] ) -> Union[str, Any]: self.assertTrue(_is_punctuation("-" ) ) self.assertTrue(_is_punctuation("$" ) ) self.assertTrue(_is_punctuation("`" ) ) self.assertTrue(_is_punctuation("." ) ) self.assertFalse(_is_punctuation("A" ) ) self.assertFalse(_is_punctuation(" " ) ) def __lowercase ( self : Tuple ) -> Optional[int]: snake_case : Optional[Any] = self.get_tokenizer() snake_case : Any = self.get_rust_tokenizer() # Example taken from the issue https://github.com/huggingface/tokenizers/issues/340 self.assertListEqual([tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) self.assertListEqual( [rust_tokenizer.tokenize(__SCREAMING_SNAKE_CASE ) for t in ["Test", "\xad", "test"]] , [["[UNK]"], [], ["[UNK]"]] ) @slow def __lowercase ( self : Any ) -> List[Any]: snake_case : List[str] = self.tokenizer_class.from_pretrained("bert-base-uncased" ) snake_case : Dict = tokenizer.encode("sequence builders" , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : List[str] = tokenizer.encode("multi-sequence build" , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : Optional[Any] = tokenizer.build_inputs_with_special_tokens(__SCREAMING_SNAKE_CASE ) snake_case : Any = tokenizer.build_inputs_with_special_tokens(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) assert encoded_sentence == [101] + text + [102] assert encoded_pair == [101] + text + [102] + text_a + [102] def __lowercase ( self : List[str] ) -> Optional[int]: for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): snake_case : Optional[int] = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case : List[Any] = F'''A, naïve {tokenizer_r.mask_token} AllenNLP sentence.''' snake_case : List[str] = tokenizer_r.encode_plus( __SCREAMING_SNAKE_CASE , return_attention_mask=__SCREAMING_SNAKE_CASE , return_token_type_ids=__SCREAMING_SNAKE_CASE , return_offsets_mapping=__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE , ) snake_case : Optional[int] = tokenizer_r.do_lower_case if hasattr(__SCREAMING_SNAKE_CASE , "do_lower_case" ) else False snake_case : Tuple = ( [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "A"), ((1, 2), ","), ((3, 5), "na"), ((5, 6), "##ï"), ((6, 8), "##ve"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "Allen"), ((21, 23), "##NL"), ((23, 24), "##P"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] if not do_lower_case else [ ((0, 0), tokenizer_r.cls_token), ((0, 1), "a"), ((1, 2), ","), ((3, 8), "naive"), ((9, 15), tokenizer_r.mask_token), ((16, 21), "allen"), ((21, 23), "##nl"), ((23, 24), "##p"), ((25, 33), "sentence"), ((33, 34), "."), ((0, 0), tokenizer_r.sep_token), ] ) self.assertEqual( [e[1] for e in expected_results] , tokenizer_r.convert_ids_to_tokens(tokens["input_ids"] ) ) self.assertEqual([e[0] for e in expected_results] , tokens["offset_mapping"] ) def __lowercase ( self : Dict ) -> Optional[int]: snake_case : int = ["的", "人", "有"] snake_case : Dict = "".join(__SCREAMING_SNAKE_CASE ) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): snake_case : Optional[Any] = True snake_case : Tuple = self.tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case : Any = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case : Tuple = tokenizer_p.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : int = tokenizer_r.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : Optional[int] = tokenizer_r.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) snake_case : Dict = tokenizer_p.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) # it is expected that each Chinese character is not preceded by "##" self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = False snake_case : int = self.rust_tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = self.tokenizer_class.from_pretrained(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) snake_case : str = tokenizer_r.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : Any = tokenizer_p.encode(__SCREAMING_SNAKE_CASE , add_special_tokens=__SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = tokenizer_r.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) snake_case : List[str] = tokenizer_p.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ) # it is expected that only the first Chinese character is not preceded by "##". snake_case : List[Any] = [ F'''##{token}''' if idx != 0 else token for idx, token in enumerate(__SCREAMING_SNAKE_CASE ) ] self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) self.assertListEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
449
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class _A ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=7 , __SCREAMING_SNAKE_CASE : Tuple=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : int=True , __SCREAMING_SNAKE_CASE : Optional[Any]=99 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : int=5 , __SCREAMING_SNAKE_CASE : Dict=4 , __SCREAMING_SNAKE_CASE : str=37 , __SCREAMING_SNAKE_CASE : Union[str, Any]="gelu" , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : str=512 , __SCREAMING_SNAKE_CASE : Dict=16 , __SCREAMING_SNAKE_CASE : Any=2 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.02 , __SCREAMING_SNAKE_CASE : List[str]=4 , ) -> Optional[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =seq_length __UpperCAmelCase =is_training __UpperCAmelCase =use_attention_mask __UpperCAmelCase =use_token_type_ids __UpperCAmelCase =use_labels __UpperCAmelCase =vocab_size __UpperCAmelCase =hidden_size __UpperCAmelCase =num_hidden_layers __UpperCAmelCase =num_attention_heads __UpperCAmelCase =intermediate_size __UpperCAmelCase =hidden_act __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =max_position_embeddings __UpperCAmelCase =type_vocab_size __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =initializer_range __UpperCAmelCase =num_choices def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase =None if self.use_attention_mask: __UpperCAmelCase =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase =None if self.use_token_type_ids: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase =RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def _a ( self : List[str] ) -> Dict: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase =True __UpperCAmelCase =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = True lowerCamelCase : Union[str, Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =FlaxRobertaModelTester(self ) @slow def _a ( self : Optional[Any] ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase =model_class_name.from_pretrained("""roberta-base""" , from_pt=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model(np.ones((1, 1) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' def __UpperCamelCase ( a : float , a : float , a : int ) ->float: if principal <= 0: raise Exception('''Principal borrowed must be > 0''' ) if rate_per_annum < 0: raise Exception('''Rate of interest must be >= 0''' ) if years_to_repay <= 0 or not isinstance(A_ , A_ ): raise Exception('''Years to repay must be an integer > 0''' ) # Yearly rate is divided by 12 to get monthly rate snake_case = rate_per_annum / 12 # Years to repay is multiplied by 12 to get number of payments as payment is monthly snake_case = years_to_repay * 12 return ( principal * rate_per_month * (1 + rate_per_month) ** number_of_payments / ((1 + rate_per_month) ** number_of_payments - 1) ) if __name__ == "__main__": import doctest doctest.testmod()
342
from __future__ import annotations def lowercase__ ( A_: list[list[int]] ) -> int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(A_ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
68
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ChineseCLIPImageProcessor class __magic_name__ ( unittest.TestCase ): def __init__( self : str , UpperCamelCase__ : Tuple , UpperCamelCase__ : str=7 , UpperCamelCase__ : Union[str, Any]=3 , UpperCamelCase__ : Tuple=18 , UpperCamelCase__ : Tuple=30 , UpperCamelCase__ : Union[str, Any]=4_00 , UpperCamelCase__ : str=True , UpperCamelCase__ : Dict=None , UpperCamelCase__ : List[str]=True , UpperCamelCase__ : int=None , UpperCamelCase__ : Optional[int]=True , UpperCamelCase__ : Optional[Any]=[0.48_14_54_66, 0.4_57_82_75, 0.40_82_10_73] , UpperCamelCase__ : int=[0.26_86_29_54, 0.26_13_02_58, 0.27_57_77_11] , UpperCamelCase__ : Dict=True , ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = size if size is not None else {"height": 2_24, "width": 2_24} UpperCAmelCase = crop_size if crop_size is not None else {"height": 18, "width": 18} UpperCAmelCase = parent UpperCAmelCase = batch_size UpperCAmelCase = num_channels UpperCAmelCase = image_size UpperCAmelCase = min_resolution UpperCAmelCase = max_resolution UpperCAmelCase = do_resize UpperCAmelCase = size UpperCAmelCase = do_center_crop UpperCAmelCase = crop_size UpperCAmelCase = do_normalize UpperCAmelCase = image_mean UpperCAmelCase = image_std UpperCAmelCase = do_convert_rgb def SCREAMING_SNAKE_CASE_ ( self : Any ) -> Any: '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_convert_rgb": self.do_convert_rgb, } def SCREAMING_SNAKE_CASE_ ( self : str , UpperCamelCase__ : str=False , UpperCamelCase__ : List[str]=False , UpperCamelCase__ : int=False ) -> List[Any]: '''simple docstring''' assert not (numpify and torchify), "You cannot specify both numpy and PyTorch tensors at the same time" if equal_resolution: UpperCAmelCase = [] for i in range(self.batch_size ): image_inputs.append( np.random.randint( 2_55 , size=(self.num_channels, self.max_resolution, self.max_resolution) , dtype=np.uinta ) ) else: UpperCAmelCase = [] for i in range(self.batch_size ): UpperCAmelCase , UpperCAmelCase = np.random.choice(np.arange(self.min_resolution , self.max_resolution ) , 2 ) image_inputs.append(np.random.randint(2_55 , size=(self.num_channels, width, height) , dtype=np.uinta ) ) if not numpify and not torchify: # PIL expects the channel dimension as last dimension UpperCAmelCase = [Image.fromarray(np.moveaxis(__SCREAMING_SNAKE_CASE , 0 , -1 ) ) for x in image_inputs] if torchify: UpperCAmelCase = [torch.from_numpy(__SCREAMING_SNAKE_CASE ) for x in image_inputs] return image_inputs @require_torch @require_vision class __magic_name__ ( A__, unittest.TestCase ): lowercase : Dict =ChineseCLIPImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> List[Any]: '''simple docstring''' UpperCAmelCase = ChineseCLIPImageProcessingTester(self , do_center_crop=__SCREAMING_SNAKE_CASE ) @property def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> List[Any]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_resize" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "size" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_center_crop" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "center_crop" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_normalize" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "image_mean" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "image_std" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_convert_rgb" ) ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> str: '''simple docstring''' UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 2_24, "width": 2_24} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) UpperCAmelCase = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCAmelCase = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=__SCREAMING_SNAKE_CASE , numpify=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , np.ndarray ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCAmelCase = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Any: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=__SCREAMING_SNAKE_CASE , torchify=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , torch.Tensor ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCAmelCase = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) @require_torch @require_vision class __magic_name__ ( A__, unittest.TestCase ): lowercase : Optional[int] =ChineseCLIPImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE_ ( self : Tuple ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase = ChineseCLIPImageProcessingTester(self , num_channels=4 , do_center_crop=__SCREAMING_SNAKE_CASE ) UpperCAmelCase = 3 @property def SCREAMING_SNAKE_CASE_ ( self : Optional[Any] ) -> Optional[int]: '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE_ ( self : Dict ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_resize" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "size" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_center_crop" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "center_crop" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_normalize" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "image_mean" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "image_std" ) ) self.assertTrue(hasattr(__SCREAMING_SNAKE_CASE , "do_convert_rgb" ) ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Tuple: '''simple docstring''' pass def SCREAMING_SNAKE_CASE_ ( self : List[Any] ) -> Dict: '''simple docstring''' UpperCAmelCase = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase = self.image_processor_tester.prepare_inputs(equal_resolution=__SCREAMING_SNAKE_CASE ) for image in image_inputs: self.assertIsInstance(__SCREAMING_SNAKE_CASE , Image.Image ) # Test not batched input UpperCAmelCase = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched UpperCAmelCase = image_processing(__SCREAMING_SNAKE_CASE , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.expected_encoded_image_num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
323
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowercase__ ( A_: int , A_: int , A_: int , A_: int , A_: int , A_: int ) -> np.ndarray: """simple docstring""" if (ksize % 2) == 0: __UpperCAmelCase =ksize + 1 __UpperCAmelCase =np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(A_ ): for x in range(A_ ): # distance from center __UpperCAmelCase =x - ksize // 2 __UpperCAmelCase =y - ksize // 2 # degree to radiant __UpperCAmelCase =theta / 180 * np.pi __UpperCAmelCase =np.cos(_theta ) __UpperCAmelCase =np.sin(_theta ) # get kernel x __UpperCAmelCase =cos_theta * px + sin_theta * py # get kernel y __UpperCAmelCase =-sin_theta * px + cos_theta * py # fill kernel __UpperCAmelCase =np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image __A = imread("../image_data/lena.jpg") # turn image in gray scale value __A = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges __A = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 1_20, 1_50]: __A = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) __A = out / out.max() * 2_55 __A = out.astype(np.uinta) imshow("Original", gray) imshow("Gabor filter with 20x20 mask and 6 directions", out) waitKey(0)
68
0
'''simple docstring''' from __future__ import annotations def lowerCAmelCase ( UpperCamelCase__ : int ): """simple docstring""" __UpperCAmelCase = str(A_ ) return len(A_ ) == 9 and set(A_ ) == set('''123456789''' ) def lowerCAmelCase ( ): """simple docstring""" for base_num in range(9_9_9_9 , 4_9_9_9 , -1 ): __UpperCAmelCase = 1_0_0_0_0_2 * base_num if is_9_pandigital(A_ ): return candidate for base_num in range(3_3_3 , 9_9 , -1 ): __UpperCAmelCase = 1_0_0_2_0_0_3 * base_num if is_9_pandigital(A_ ): return candidate return None if __name__ == "__main__": print(F"""{solution() = }""")
262
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _A : """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : Optional[int]=2 , __SCREAMING_SNAKE_CASE : Optional[int]=3 , __SCREAMING_SNAKE_CASE : List[str]=16 , __SCREAMING_SNAKE_CASE : Union[str, Any]=[1, 2, 1] , __SCREAMING_SNAKE_CASE : List[Any]=[2, 2, 4] , __SCREAMING_SNAKE_CASE : str=2 , __SCREAMING_SNAKE_CASE : Any=2.0 , __SCREAMING_SNAKE_CASE : Optional[int]=True , __SCREAMING_SNAKE_CASE : int=0.0 , __SCREAMING_SNAKE_CASE : Dict=0.0 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : Any="gelu" , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : List[Any]=0.02 , __SCREAMING_SNAKE_CASE : Tuple=1e-5 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : Optional[int]=None , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : Any=10 , __SCREAMING_SNAKE_CASE : Dict=8 , ) -> List[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =image_size __UpperCAmelCase =patch_size __UpperCAmelCase =num_channels __UpperCAmelCase =embed_dim __UpperCAmelCase =depths __UpperCAmelCase =num_heads __UpperCAmelCase =window_size __UpperCAmelCase =mlp_ratio __UpperCAmelCase =qkv_bias __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =drop_path_rate __UpperCAmelCase =hidden_act __UpperCAmelCase =use_absolute_embeddings __UpperCAmelCase =patch_norm __UpperCAmelCase =layer_norm_eps __UpperCAmelCase =initializer_range __UpperCAmelCase =is_training __UpperCAmelCase =scope __UpperCAmelCase =use_labels __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =encoder_stride def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase =None if self.use_labels: __UpperCAmelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase =self.get_config() return config, pixel_values, labels def _a ( self : List[Any] ) -> Optional[Any]: return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[int]: __UpperCAmelCase =SwinvaModel(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __UpperCAmelCase =int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : int ) -> Tuple: __UpperCAmelCase =SwinvaForMaskedImageModeling(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __UpperCAmelCase =1 __UpperCAmelCase =SwinvaForMaskedImageModeling(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Tuple: __UpperCAmelCase =self.type_sequence_label_size __UpperCAmelCase =SwinvaForImageClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : List[str] ) -> Tuple: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _A ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Optional[int] = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowerCamelCase : Tuple = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowerCamelCase : Dict = False lowerCamelCase : Tuple = False lowerCamelCase : List[str] = False lowerCamelCase : Tuple = False def _a ( self : str ) -> str: __UpperCAmelCase =SwinvaModelTester(self ) __UpperCAmelCase =ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , embed_dim=37 ) def _a ( self : List[Any] ) -> Optional[int]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : str ) -> str: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def _a ( self : Tuple ) -> Tuple: pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def _a ( self : Optional[Any] ) -> int: pass def _a ( self : Tuple ) -> int: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __UpperCAmelCase =model.get_output_embeddings() self.assertTrue(x is None or isinstance(__SCREAMING_SNAKE_CASE , nn.Linear ) ) def _a ( self : str ) -> List[str]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase =[*signature.parameters.keys()] __UpperCAmelCase =["""pixel_values"""] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =True for model_class in self.all_model_classes: __UpperCAmelCase =True __UpperCAmelCase =False __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions __UpperCAmelCase =len(self.model_tester.depths ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCAmelCase =True __UpperCAmelCase =config.window_size**2 __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) __UpperCAmelCase =len(__SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine __UpperCAmelCase =True __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if hasattr(self.model_tester , """num_hidden_states_types""" ): __UpperCAmelCase =self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states __UpperCAmelCase =2 self.assertEqual(out_len + added_hidden_states , len(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> int: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.hidden_states __UpperCAmelCase =getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # Swinv2 has a different seq_length __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) __UpperCAmelCase =outputs.reshaped_hidden_states self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =reshaped_hidden_states[0].shape __UpperCAmelCase =( reshaped_hidden_states[0].view(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =3 __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __UpperCAmelCase =image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Dict: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__SCREAMING_SNAKE_CASE ) @slow def _a ( self : int ) -> Dict: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase =SwinvaModel.from_pretrained(__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) def _a ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =_config_zero_init(__SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: __UpperCAmelCase =model_class(config=__SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class _A ( unittest.TestCase ): """simple docstring""" @cached_property def _a ( self : Tuple ) -> Dict: return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def _a ( self : int ) -> Optional[int]: __UpperCAmelCase =SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.default_image_processor __UpperCAmelCase =Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) __UpperCAmelCase =image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).to(__SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE ) # verify the logits __UpperCAmelCase =torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([-0.3_947, -0.4_306, 0.0_026] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) )
68
0
"""simple docstring""" import logging import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEncoder, BertModel, BertPreTrainedModel, ) __SCREAMING_SNAKE_CASE =logging.getLogger(__name__) class UpperCamelCase ( lowercase_ ): def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=None ,__UpperCamelCase=None ) -> List[str]: '''simple docstring''' lowercase_ : Optional[int] = self.layer[current_layer](__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,head_mask[current_layer] ) lowercase_ : Optional[Any] = layer_outputs[0] return hidden_states @add_start_docstrings( 'The bare Bert Model transformer with PABEE outputting raw hidden-states without any specific head on top.' , lowercase_ , ) class UpperCamelCase ( lowercase_ ): def __init__( self ,__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' super().__init__(__SCREAMING_SNAKE_CASE ) lowercase_ : str = BertEncoderWithPabee(__SCREAMING_SNAKE_CASE ) self.init_weights() lowercase_ : List[str] = 0 lowercase_ : str = 0 lowercase_ : List[Any] = 0 lowercase_ : Tuple = 0 def _UpperCAmelCase ( self ,__UpperCamelCase ) -> int: '''simple docstring''' lowercase_ : Dict = threshold def _UpperCAmelCase ( self ,__UpperCamelCase ) -> Union[str, Any]: '''simple docstring''' lowercase_ : Optional[Any] = patience def _UpperCAmelCase ( self ) -> Tuple: '''simple docstring''' lowercase_ : Tuple = 0 lowercase_ : List[str] = 0 def _UpperCAmelCase ( self ) -> str: '''simple docstring''' lowercase_ : Optional[int] = self.inference_layers_num / self.inference_instances_num lowercase_ : int = ( f'''*** Patience = {self.patience} Avg. Inference Layers = {avg_inf_layers:.2f} Speed Up =''' f''' {1 - avg_inf_layers / self.config.num_hidden_layers:.2f} ***''' ) print(__SCREAMING_SNAKE_CASE ) @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) def _UpperCAmelCase ( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=False ,) -> str: '''simple docstring''' if input_ids is not None and inputs_embeds is not None: raise ValueError('You cannot specify both input_ids and inputs_embeds at the same time' ) elif input_ids is not None: lowercase_ : Any = input_ids.size() elif inputs_embeds is not None: lowercase_ : Any = inputs_embeds.size()[:-1] else: raise ValueError('You have to specify either input_ids or inputs_embeds' ) lowercase_ : List[str] = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: lowercase_ : int = torch.ones(__SCREAMING_SNAKE_CASE ,device=__SCREAMING_SNAKE_CASE ) if token_type_ids is None: lowercase_ : Tuple = torch.zeros(__SCREAMING_SNAKE_CASE ,dtype=torch.long ,device=__SCREAMING_SNAKE_CASE ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. lowercase_ : Tuple = self.get_extended_attention_mask(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if self.config.is_decoder and encoder_hidden_states is not None: lowercase_ , lowercase_ , lowercase_ : int = encoder_hidden_states.size() lowercase_ : Optional[Any] = (encoder_batch_size, encoder_sequence_length) if encoder_attention_mask is None: lowercase_ : int = torch.ones(__SCREAMING_SNAKE_CASE ,device=__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = self.invert_attention_mask(__SCREAMING_SNAKE_CASE ) else: lowercase_ : Any = None # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] lowercase_ : List[Any] = self.get_head_mask(__SCREAMING_SNAKE_CASE ,self.config.num_hidden_layers ) lowercase_ : Union[str, Any] = self.embeddings( input_ids=__SCREAMING_SNAKE_CASE ,position_ids=__SCREAMING_SNAKE_CASE ,token_type_ids=__SCREAMING_SNAKE_CASE ,inputs_embeds=__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = embedding_output if self.training: lowercase_ : str = [] for i in range(self.config.num_hidden_layers ): lowercase_ : int = self.encoder.adaptive_forward( __SCREAMING_SNAKE_CASE ,current_layer=__SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ,head_mask=__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = self.pooler(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = output_layers[i](output_dropout(__SCREAMING_SNAKE_CASE ) ) res.append(__SCREAMING_SNAKE_CASE ) elif self.patience == 0: # Use all layers for inference lowercase_ : str = self.encoder( __SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ,head_mask=__SCREAMING_SNAKE_CASE ,encoder_hidden_states=__SCREAMING_SNAKE_CASE ,encoder_attention_mask=__SCREAMING_SNAKE_CASE ,) lowercase_ : List[str] = self.pooler(encoder_outputs[0] ) lowercase_ : str = [output_layers[self.config.num_hidden_layers - 1](__SCREAMING_SNAKE_CASE )] else: lowercase_ : Optional[int] = 0 lowercase_ : Any = None lowercase_ : Optional[Any] = 0 for i in range(self.config.num_hidden_layers ): calculated_layer_num += 1 lowercase_ : Dict = self.encoder.adaptive_forward( __SCREAMING_SNAKE_CASE ,current_layer=__SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ,head_mask=__SCREAMING_SNAKE_CASE ) lowercase_ : List[Any] = self.pooler(__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = output_layers[i](__SCREAMING_SNAKE_CASE ) if regression: lowercase_ : Optional[Any] = logits.detach() if patient_result is not None: lowercase_ : Any = patient_result.detach() if (patient_result is not None) and torch.abs(patient_result - labels ) < self.regression_threshold: patient_counter += 1 else: lowercase_ : Any = 0 else: lowercase_ : List[Any] = logits.detach().argmax(dim=1 ) if patient_result is not None: lowercase_ : Tuple = patient_result.detach().argmax(dim=1 ) if (patient_result is not None) and torch.all(labels.eq(__SCREAMING_SNAKE_CASE ) ): patient_counter += 1 else: lowercase_ : Dict = 0 lowercase_ : int = logits if patient_counter == self.patience: break lowercase_ : List[Any] = [patient_result] self.inference_layers_num += calculated_layer_num self.inference_instances_num += 1 return res @add_start_docstrings( 'Bert Model transformer with PABEE and a sequence classification/regression head on top (a linear layer on top of\n the pooled output) e.g. for GLUE tasks. ' , lowercase_ , ) class UpperCamelCase ( lowercase_ ): def __init__( self ,__UpperCamelCase ) -> Dict: '''simple docstring''' super().__init__(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = config.num_labels lowercase_ : Tuple = BertModelWithPabee(__SCREAMING_SNAKE_CASE ) lowercase_ : str = nn.Dropout(config.hidden_dropout_prob ) lowercase_ : Optional[int] = nn.ModuleList( [nn.Linear(config.hidden_size ,self.config.num_labels ) for _ in range(config.num_hidden_layers )] ) self.init_weights() @add_start_docstrings_to_model_forward(__SCREAMING_SNAKE_CASE ) def _UpperCAmelCase ( self ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,__UpperCamelCase=None ,) -> List[Any]: '''simple docstring''' lowercase_ : Optional[Any] = self.bert( input_ids=__SCREAMING_SNAKE_CASE ,attention_mask=__SCREAMING_SNAKE_CASE ,token_type_ids=__SCREAMING_SNAKE_CASE ,position_ids=__SCREAMING_SNAKE_CASE ,head_mask=__SCREAMING_SNAKE_CASE ,inputs_embeds=__SCREAMING_SNAKE_CASE ,output_dropout=self.dropout ,output_layers=self.classifiers ,regression=self.num_labels == 1 ,) lowercase_ : Union[str, Any] = (logits[-1],) if labels is not None: lowercase_ : Optional[int] = None lowercase_ : str = 0 for ix, logits_item in enumerate(__SCREAMING_SNAKE_CASE ): if self.num_labels == 1: # We are doing regression lowercase_ : Union[str, Any] = MSELoss() lowercase_ : Optional[int] = loss_fct(logits_item.view(-1 ) ,labels.view(-1 ) ) else: lowercase_ : Optional[int] = CrossEntropyLoss() lowercase_ : Optional[Any] = loss_fct(logits_item.view(-1 ,self.num_labels ) ,labels.view(-1 ) ) if total_loss is None: lowercase_ : str = loss else: total_loss += loss * (ix + 1) total_weights += ix + 1 lowercase_ : Dict = (total_loss / total_weights,) + outputs return outputs
425
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A = logging.get_logger(__name__) __A = {"vocab_file": "spiece.model"} __A = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } __A = { "AI-Sweden/gpt-sw3-126m": 20_48, "AI-Sweden/gpt-sw3-350m": 20_48, "AI-Sweden/gpt-sw3-1.6b": 20_48, "AI-Sweden/gpt-sw3-6.7b": 20_48, "AI-Sweden/gpt-sw3-20b": 20_48, } class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : int = VOCAB_FILES_NAMES lowerCamelCase : int = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = ['input_ids', 'attention_mask'] def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : Union[str, Any]=False , __SCREAMING_SNAKE_CASE : List[str]=False , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Optional[Dict[str, Any]] = None , **__SCREAMING_SNAKE_CASE : Optional[Any] , ) -> None: __UpperCAmelCase ={} if sp_model_kwargs is None else sp_model_kwargs __UpperCAmelCase =kwargs.get("""name_or_path""" ) if name_or_path is None: logger.warning( """name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,""" """ you are testing the model, this can safely be ignored""" ) __UpperCAmelCase ="""None""" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __UpperCAmelCase ="""<|endoftext|>""" if eos_token is None else eos_token __UpperCAmelCase ="""<unk>""" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __UpperCAmelCase =unk_token if pad_token is None else pad_token __UpperCAmelCase =eos_token if bos_token is None else bos_token else: __UpperCAmelCase ="""<pad>""" if pad_token is None else pad_token __UpperCAmelCase ="""<s>""" if bos_token is None else bos_token super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) __UpperCAmelCase =do_lower_case __UpperCAmelCase =remove_space __UpperCAmelCase =keep_accents __UpperCAmelCase =vocab_file __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) # Used for whitespace normalization in input texts # fmt : off __UpperCAmelCase ={""" """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """""", """„"""} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __UpperCAmelCase =re.compile( f'''[{"".join(map(__SCREAMING_SNAKE_CASE , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]''' ) def __getstate__( self : Any ) -> str: __UpperCAmelCase =self.__dict__.copy() __UpperCAmelCase =None return state def __setstate__( self : str , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Union[str, Any]: __UpperCAmelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __UpperCAmelCase ={} __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def _a ( self : Union[str, Any] ) -> int: return len(self.sp_model ) def _a ( self : Dict , __SCREAMING_SNAKE_CASE : str ) -> str: __UpperCAmelCase =self.non_printing_characters_re.sub("""""" , __SCREAMING_SNAKE_CASE ) # Normalize whitespaces __UpperCAmelCase ="""""".join([char if char not in self.whitespaces else """ """ for char in text] ) # NFC Unicode normalization __UpperCAmelCase =unicodedata.normalize("""NFC""" , __SCREAMING_SNAKE_CASE ) return text def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> List[str]: __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) return self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str ) -> int: return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : int ) -> str: return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) @staticmethod def _a ( __SCREAMING_SNAKE_CASE : str ) -> str: return out_string def _a ( self : Any , __SCREAMING_SNAKE_CASE : List[str] ) -> str: __UpperCAmelCase =[] __UpperCAmelCase ="""""" __UpperCAmelCase =False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) + token __UpperCAmelCase =True __UpperCAmelCase =[] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =False out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) return out_string def _a ( self : Any ) -> Dict[str, int]: __UpperCAmelCase ={self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _a ( self : Any , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __UpperCAmelCase =os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE , """wb""" ) as fi: __UpperCAmelCase =self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, List[str]] , __SCREAMING_SNAKE_CASE : Union[str, bool] = False ) -> Union[List[int], List[List[int]], "torch.Tensor"]: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) else: __UpperCAmelCase =[self.preprocess_text(__SCREAMING_SNAKE_CASE ) for t in text] __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) if return_tensors is True or return_tensors == "pt": __UpperCAmelCase =torch.tensor(__SCREAMING_SNAKE_CASE ) return token_ids def _a ( self : str , __SCREAMING_SNAKE_CASE : Union[int, List[int]] ) -> str: return self.sp_model.decode(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : "Conversation" ) -> List[int]: __UpperCAmelCase =[f'''User: {text}''' if is_user else f'''Bot: {text}''' for is_user, text in conversation.iter_texts()] __UpperCAmelCase =( f'''{self.eos_token}{self.bos_token}''' + f'''{self.bos_token}'''.join(__SCREAMING_SNAKE_CASE ) + f'''{self.bos_token}Bot:''' ) return self.encode(text=__SCREAMING_SNAKE_CASE )
68
0
"""simple docstring""" from __future__ import annotations from collections.abc import Iterator class __lowerCAmelCase : """simple docstring""" def __init__( self : List[str] , _snake_case : int ) -> None: """simple docstring""" A_ = value A_ = None A_ = None class __lowerCAmelCase : """simple docstring""" def __init__( self : List[Any] , _snake_case : Node ) -> None: """simple docstring""" A_ = tree def lowerCamelCase__ ( self : Tuple , _snake_case : Node | None ) -> int: """simple docstring""" if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : int ) -> Iterator[int]: """simple docstring""" yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
115
import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow __A = logging.getLogger() @unittest.skip('Temporarily disable the doc tests.' ) @require_torch @require_tf @slow class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Path , __SCREAMING_SNAKE_CASE : Union[str, None] = None , __SCREAMING_SNAKE_CASE : Union[List[str], None] = None , __SCREAMING_SNAKE_CASE : Union[str, List[str], None] = None , __SCREAMING_SNAKE_CASE : bool = True , ) -> List[str]: __UpperCAmelCase =[file for file in os.listdir(__SCREAMING_SNAKE_CASE ) if os.path.isfile(os.path.join(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) )] if identifier is not None: __UpperCAmelCase =[file for file in files if identifier in file] if n_identifier is not None: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): for n_ in n_identifier: __UpperCAmelCase =[file for file in files if n_ not in file] else: __UpperCAmelCase =[file for file in files if n_identifier not in file] __UpperCAmelCase =ignore_files or [] ignore_files.append("""__init__.py""" ) __UpperCAmelCase =[file for file in files if file not in ignore_files] for file in files: # Open all files print("""Testing""" , __SCREAMING_SNAKE_CASE ) if only_modules: __UpperCAmelCase =file.split(""".""" )[0] try: __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =doctest.DocTestSuite(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =unittest.TextTestRunner().run(__SCREAMING_SNAKE_CASE ) self.assertIs(len(result.failures ) , 0 ) except AttributeError: logger.info(f'''{module_identifier} is not a module.''' ) else: __UpperCAmelCase =doctest.testfile(str("""..""" / directory / file ) , optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed , 0 ) def _a ( self : Optional[Any] ) -> List[str]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""modeling""" __UpperCAmelCase =[ """modeling_ctrl.py""", """modeling_tf_ctrl.py""", ] self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE , ignore_files=__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""tokenization""" self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] ) -> Optional[Any]: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase ="""configuration""" self.analyze_directory(__SCREAMING_SNAKE_CASE , identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : List[Any] ) -> Tuple: __UpperCAmelCase =Path("""src/transformers""" ) __UpperCAmelCase =["""configuration""", """modeling""", """tokenization"""] self.analyze_directory(__SCREAMING_SNAKE_CASE , n_identifier=__SCREAMING_SNAKE_CASE ) def _a ( self : Any ) -> Tuple: __UpperCAmelCase =Path("""docs/source""" ) __UpperCAmelCase =["""favicon.ico"""] self.analyze_directory(__SCREAMING_SNAKE_CASE , ignore_files=__SCREAMING_SNAKE_CASE , only_modules=__SCREAMING_SNAKE_CASE )
68
0
import json from typing import List, Optional, Tuple from tokenizers import pre_tokenizers, processors from ...tokenization_utils_base import AddedToken, BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roberta import RobertaTokenizer __snake_case :Optional[Any] =logging.get_logger(__name__) __snake_case :List[str] ={'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_file': 'tokenizer.json'} __snake_case :List[Any] ={ 'vocab_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/vocab.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/vocab.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/vocab.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/vocab.json', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/vocab.json', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/vocab.json' ), }, 'merges_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/merges.txt', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/merges.txt', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/merges.txt', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/merges.txt', 'roberta-base-openai-detector': 'https://huggingface.co/roberta-base-openai-detector/resolve/main/merges.txt', 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/merges.txt' ), }, 'tokenizer_file': { 'roberta-base': 'https://huggingface.co/roberta-base/resolve/main/tokenizer.json', 'roberta-large': 'https://huggingface.co/roberta-large/resolve/main/tokenizer.json', 'roberta-large-mnli': 'https://huggingface.co/roberta-large-mnli/resolve/main/tokenizer.json', 'distilroberta-base': 'https://huggingface.co/distilroberta-base/resolve/main/tokenizer.json', 'roberta-base-openai-detector': ( 'https://huggingface.co/roberta-base-openai-detector/resolve/main/tokenizer.json' ), 'roberta-large-openai-detector': ( 'https://huggingface.co/roberta-large-openai-detector/resolve/main/tokenizer.json' ), }, } __snake_case :str ={ 'roberta-base': 512, 'roberta-large': 512, 'roberta-large-mnli': 512, 'distilroberta-base': 512, 'roberta-base-openai-detector': 512, 'roberta-large-openai-detector': 512, } class lowerCAmelCase__ ( _lowerCamelCase ): A_ : Optional[Any] = VOCAB_FILES_NAMES A_ : Optional[int] = PRETRAINED_VOCAB_FILES_MAP A_ : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES A_ : Any = ['input_ids', 'attention_mask'] A_ : Optional[int] = RobertaTokenizer def __init__( self : int , __UpperCamelCase : Union[str, Any]=None , __UpperCamelCase : Any=None , __UpperCamelCase : List[Any]=None , __UpperCamelCase : Tuple="replace" , __UpperCamelCase : str="<s>" , __UpperCamelCase : List[Any]="</s>" , __UpperCamelCase : str="</s>" , __UpperCamelCase : Any="<s>" , __UpperCamelCase : int="<unk>" , __UpperCamelCase : Optional[Any]="<pad>" , __UpperCamelCase : int="<mask>" , __UpperCamelCase : Tuple=False , __UpperCamelCase : List[Any]=True , **__UpperCamelCase : List[Any] , ) -> int: super().__init__( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , errors=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE , trim_offsets=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) A = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get('add_prefix_space' , __SCREAMING_SNAKE_CASE ) != add_prefix_space: A = getattr(__SCREAMING_SNAKE_CASE , pre_tok_state.pop('type' ) ) A = add_prefix_space A = pre_tok_class(**__SCREAMING_SNAKE_CASE ) A = add_prefix_space A = 'post_processor' A = getattr(self.backend_tokenizer , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if tokenizer_component_instance: A = json.loads(tokenizer_component_instance.__getstate__() ) # The lists 'sep' and 'cls' must be cased in tuples for the object `post_processor_class` if "sep" in state: A = tuple(state['sep'] ) if "cls" in state: A = tuple(state['cls'] ) A = False if state.get('add_prefix_space' , __SCREAMING_SNAKE_CASE ) != add_prefix_space: A = add_prefix_space A = True if state.get('trim_offsets' , __SCREAMING_SNAKE_CASE ) != trim_offsets: A = trim_offsets A = True if changes_to_apply: A = getattr(__SCREAMING_SNAKE_CASE , state.pop('type' ) ) A = component_class(**__SCREAMING_SNAKE_CASE ) setattr(self.backend_tokenizer , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) @property def __UpperCamelCase ( self : List[str] ) -> str: if self._mask_token is None: if self.verbose: logger.error('Using mask_token, but it is not set yet.' ) return None return str(self._mask_token ) @mask_token.setter def __UpperCamelCase ( self : str , __UpperCamelCase : List[Any] ) -> List[Any]: A = AddedToken(__SCREAMING_SNAKE_CASE , lstrip=__SCREAMING_SNAKE_CASE , rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) else value A = value def __UpperCamelCase ( self : Any , *__UpperCamelCase : List[str] , **__UpperCamelCase : str ) -> BatchEncoding: A = kwargs.get('is_split_into_words' , __SCREAMING_SNAKE_CASE ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self : List[str] , *__UpperCamelCase : int , **__UpperCamelCase : Union[str, Any] ) -> BatchEncoding: A = kwargs.get('is_split_into_words' , __SCREAMING_SNAKE_CASE ) assert self.add_prefix_space or not is_split_into_words, ( f'''You need to instantiate {self.__class__.__name__} with add_prefix_space=True ''' "to use it with pretokenized inputs." ) return super()._encode_plus(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self : Any , __UpperCamelCase : str , __UpperCamelCase : Optional[str] = None ) -> Tuple[str]: A = self._tokenizer.model.save(__SCREAMING_SNAKE_CASE , name=__SCREAMING_SNAKE_CASE ) return tuple(__SCREAMING_SNAKE_CASE ) def __UpperCamelCase ( self : List[Any] , __UpperCamelCase : int , __UpperCamelCase : Optional[int]=None ) -> List[str]: A = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __UpperCamelCase ( self : str , __UpperCamelCase : List[int] , __UpperCamelCase : Optional[List[int]] = None ) -> List[int]: A = [self.sep_token_id] A = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
106
import random import unittest import numpy as np import transformers from transformers import is_flax_available, is_torch_available from transformers.testing_utils import is_pt_flax_cross_test, require_flax if is_flax_available(): import os import jax.numpy as jnp from jax import jit from transformers import AutoTokenizer, FlaxAutoModelForCausalLM from transformers.modeling_flax_pytorch_utils import load_flax_weights_in_pytorch_model __A = "0.12" # assumed parallelism: 8 if is_torch_available(): import torch def lowercase__ ( A_: int , A_: Optional[Any] , A_: List[str]=None ) -> List[str]: """simple docstring""" if rng is None: __UpperCAmelCase =random.Random() __UpperCAmelCase =1 for dim in shape: total_dims *= dim __UpperCAmelCase =[] for _ in range(A_ ): values.append(rng.randint(0 , vocab_size - 1 ) ) __UpperCAmelCase =np.array(A_ , dtype=jnp.intaa ).reshape(A_ ) return output def lowercase__ ( A_: List[str] , A_: List[str]=None ) -> Any: """simple docstring""" __UpperCAmelCase =ids_tensor(A_ , vocab_size=2 , rng=A_ ) # make sure that at least one token is attended to for each batch __UpperCAmelCase =1 return attn_mask @require_flax class _A : """simple docstring""" lowerCamelCase : Optional[Any] = None lowerCamelCase : int = () def _a ( self : str ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() # cut to half length & take max batch_size 3 __UpperCAmelCase =2 __UpperCAmelCase =inputs["""input_ids"""].shape[-1] // 2 __UpperCAmelCase =inputs["""input_ids"""][:max_batch_size, :sequence_length] __UpperCAmelCase =jnp.ones_like(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =attention_mask[:max_batch_size, :sequence_length] # generate max 5 tokens __UpperCAmelCase =input_ids.shape[-1] + 5 if config.eos_token_id is not None and config.pad_token_id is None: # hack to allow generate for models such as GPT2 as is done in `generate()` __UpperCAmelCase =config.eos_token_id return config, input_ids, attention_mask, max_length @is_pt_flax_cross_test def _a ( self : Union[str, Any] ) -> Optional[int]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =0 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model_class.__name__[4:] # Skip the "Flax" at the beginning __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =pt_model_class(__SCREAMING_SNAKE_CASE ).eval() __UpperCAmelCase =load_flax_weights_in_pytorch_model(__SCREAMING_SNAKE_CASE , flax_model.params ) __UpperCAmelCase =flax_model.generate(__SCREAMING_SNAKE_CASE ).sequences __UpperCAmelCase =pt_model.generate(torch.tensor(__SCREAMING_SNAKE_CASE , dtype=torch.long ) ) if flax_generation_outputs.shape[-1] > pt_generation_outputs.shape[-1]: __UpperCAmelCase =flax_generation_outputs[:, : pt_generation_outputs.shape[-1]] self.assertListEqual(pt_generation_outputs.numpy().tolist() , flax_generation_outputs.tolist() ) def _a ( self : Optional[int] ) -> Dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Union[str, Any] ) -> List[str]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =True __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : List[Any] ) -> Any: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =2 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Any ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =False __UpperCAmelCase =max_length __UpperCAmelCase =2 __UpperCAmelCase =2 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[0] , input_ids.shape[0] * config.num_return_sequences ) def _a ( self : Union[str, Any] ) -> List[Any]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =True __UpperCAmelCase =max_length __UpperCAmelCase =0.8 __UpperCAmelCase =10 __UpperCAmelCase =0.3 __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =max_length __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Optional[int] ) -> Any: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() __UpperCAmelCase =max_length __UpperCAmelCase =2 __UpperCAmelCase =1 __UpperCAmelCase =8 __UpperCAmelCase =9 for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : List[str] ) -> Dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =False __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Dict ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =True __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) def _a ( self : Dict ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._get_input_ids_and_config() # pad attention mask on the left __UpperCAmelCase =attention_mask.at[(0, 0)].set(0 ) __UpperCAmelCase =2 __UpperCAmelCase =max_length for model_class in self.all_generative_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model.generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertEqual(generation_outputs.shape[-1] , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =jit(model.generate ) __UpperCAmelCase =jit_generate(__SCREAMING_SNAKE_CASE , attention_mask=__SCREAMING_SNAKE_CASE ).sequences self.assertListEqual(generation_outputs.tolist() , jit_generation_outputs.tolist() ) @require_flax class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : int ) -> Any: __UpperCAmelCase =AutoTokenizer.from_pretrained("""hf-internal-testing/tiny-bert""" ) __UpperCAmelCase =FlaxAutoModelForCausalLM.from_pretrained("""hf-internal-testing/tiny-bert-flax-only""" ) __UpperCAmelCase ="""Hello world""" __UpperCAmelCase =tokenizer(__SCREAMING_SNAKE_CASE , return_tensors="""np""" ).input_ids # typos are quickly detected (the correct argument is `do_sample`) with self.assertRaisesRegex(__SCREAMING_SNAKE_CASE , """do_samples""" ): model.generate(__SCREAMING_SNAKE_CASE , do_samples=__SCREAMING_SNAKE_CASE ) # arbitrary arguments that will not be used anywhere are also not accepted with self.assertRaisesRegex(__SCREAMING_SNAKE_CASE , """foo""" ): __UpperCAmelCase ={"""foo""": """bar"""} model.generate(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE )
68
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) lowerCAmelCase_ = { 'configuration_speech_to_text': ['SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'Speech2TextConfig'], 'processing_speech_to_text': ['Speech2TextProcessor'], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['Speech2TextTokenizer'] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = ['Speech2TextFeatureExtractor'] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFSpeech2TextForConditionalGeneration', 'TFSpeech2TextModel', 'TFSpeech2TextPreTrainedModel', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase_ = [ 'SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST', 'Speech2TextForConditionalGeneration', 'Speech2TextModel', 'Speech2TextPreTrainedModel', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys lowerCAmelCase_ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
560
from __future__ import annotations from collections.abc import Iterator class _A : """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : int ) -> None: __UpperCAmelCase =value __UpperCAmelCase =None __UpperCAmelCase =None class _A : """simple docstring""" def __init__( self : List[Any] , __SCREAMING_SNAKE_CASE : Node ) -> None: __UpperCAmelCase =tree def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Node | None ) -> int: if node is None: return 0 return node.value + ( self.depth_first_search(node.left ) + self.depth_first_search(node.right ) ) def __iter__( self : int ) -> Iterator[int]: yield self.depth_first_search(self.tree ) if __name__ == "__main__": import doctest doctest.testmod()
68
0
'''simple docstring''' from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING a__ : int = logging.get_logger(__name__) a__ : List[str] = Dict[str, Any] a__ : List[Any] = List[Prediction] @add_end_docstrings(UpperCAmelCase_ ) class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' def __init__( self : int , *a__ : List[str] , **a__ : Any ): super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if self.framework == "tf": raise ValueError(f"The {self.__class__} is only available in PyTorch." ) requires_backends(self , '''vision''' ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def __snake_case ( self : str , **a__ : str ): UpperCAmelCase = {} if "threshold" in kwargs: UpperCAmelCase = kwargs['''threshold'''] return {}, {}, postprocess_kwargs def __call__( self : Union[str, Any] , *a__ : Dict , **a__ : Union[str, Any] ): return super().__call__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) def __snake_case ( self : List[str] , a__ : Optional[int] ): UpperCAmelCase = load_image(__SCREAMING_SNAKE_CASE ) UpperCAmelCase = torch.IntTensor([[image.height, image.width]] ) UpperCAmelCase = self.image_processor(images=[image] , return_tensors='''pt''' ) if self.tokenizer is not None: UpperCAmelCase = self.tokenizer(text=inputs['''words'''] , boxes=inputs['''boxes'''] , return_tensors='''pt''' ) UpperCAmelCase = target_size return inputs def __snake_case ( self : Tuple , a__ : Dict ): UpperCAmelCase = model_inputs.pop('''target_size''' ) UpperCAmelCase = self.model(**__SCREAMING_SNAKE_CASE ) UpperCAmelCase = outputs.__class__({'''target_size''': target_size, **outputs} ) if self.tokenizer is not None: UpperCAmelCase = model_inputs['''bbox'''] return model_outputs def __snake_case ( self : Union[str, Any] , a__ : List[str] , a__ : Any=0.9 ): UpperCAmelCase = model_outputs['''target_size'''] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. UpperCAmelCase, UpperCAmelCase = target_size[0].tolist() def unnormalize(a__ : Union[str, Any] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) UpperCAmelCase, UpperCAmelCase = model_outputs['''logits'''].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) UpperCAmelCase = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] UpperCAmelCase = [unnormalize(__SCREAMING_SNAKE_CASE ) for bbox in model_outputs['''bbox'''].squeeze(0 )] UpperCAmelCase = ['''score''', '''label''', '''box'''] UpperCAmelCase = [dict(zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) for vals in zip(scores.tolist() , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel UpperCAmelCase = self.image_processor.post_process_object_detection(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) UpperCAmelCase = raw_annotations[0] UpperCAmelCase = raw_annotation['''scores'''] UpperCAmelCase = raw_annotation['''labels'''] UpperCAmelCase = raw_annotation['''boxes'''] UpperCAmelCase = scores.tolist() UpperCAmelCase = [self.model.config.idalabel[label.item()] for label in labels] UpperCAmelCase = [self._get_bounding_box(__SCREAMING_SNAKE_CASE ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] UpperCAmelCase = ['''score''', '''label''', '''box'''] UpperCAmelCase = [ dict(zip(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) for vals in zip(raw_annotation['''scores'''] , raw_annotation['''labels'''] , raw_annotation['''boxes'''] ) ] return annotation def __snake_case ( self : Any , a__ : "torch.Tensor" ): if self.framework != "pt": raise ValueError('''The ObjectDetectionPipeline is only available in PyTorch.''' ) UpperCAmelCase, UpperCAmelCase, UpperCAmelCase, UpperCAmelCase = box.int().tolist() UpperCAmelCase = { '''xmin''': xmin, '''ymin''': ymin, '''xmax''': xmax, '''ymax''': ymax, } return bbox
51
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import os from ...utils.constants import SAGEMAKER_PARALLEL_EC2_INSTANCES, TORCH_DYNAMO_MODES from ...utils.dataclasses import ComputeEnvironment, SageMakerDistributedType from ...utils.imports import is_botoa_available from .config_args import SageMakerConfig from .config_utils import ( DYNAMO_BACKENDS, _ask_field, _ask_options, _convert_dynamo_backend, _convert_mixed_precision, _convert_sagemaker_distributed_mode, _convert_yes_no_to_bool, ) if is_botoa_available(): import botoa # noqa: F401 def lowercase__ ( A_: Union[str, Any] ) -> List[Any]: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ {"""Effect""": """Allow""", """Principal""": {"""Service""": """sagemaker.amazonaws.com"""}, """Action""": """sts:AssumeRole"""} ], } try: # create the role, associated with the chosen trust policy iam_client.create_role( RoleName=A_ , AssumeRolePolicyDocument=json.dumps(A_ , indent=2 ) ) __UpperCAmelCase ={ """Version""": """2012-10-17""", """Statement""": [ { """Effect""": """Allow""", """Action""": [ """sagemaker:*""", """ecr:GetDownloadUrlForLayer""", """ecr:BatchGetImage""", """ecr:BatchCheckLayerAvailability""", """ecr:GetAuthorizationToken""", """cloudwatch:PutMetricData""", """cloudwatch:GetMetricData""", """cloudwatch:GetMetricStatistics""", """cloudwatch:ListMetrics""", """logs:CreateLogGroup""", """logs:CreateLogStream""", """logs:DescribeLogStreams""", """logs:PutLogEvents""", """logs:GetLogEvents""", """s3:CreateBucket""", """s3:ListBucket""", """s3:GetBucketLocation""", """s3:GetObject""", """s3:PutObject""", ], """Resource""": """*""", } ], } # attach policy to role iam_client.put_role_policy( RoleName=A_ , PolicyName=F'''{role_name}_policy_permission''' , PolicyDocument=json.dumps(A_ , indent=2 ) , ) except iam_client.exceptions.EntityAlreadyExistsException: print(F'''role {role_name} already exists. Using existing one''' ) def lowercase__ ( A_: Dict ) -> Any: """simple docstring""" __UpperCAmelCase =botoa.client("""iam""" ) return iam_client.get_role(RoleName=A_ )["Role"]["Arn"] def lowercase__ ( ) -> Union[str, Any]: """simple docstring""" __UpperCAmelCase =_ask_options( """How do you want to authorize?""" , ["""AWS Profile""", """Credentials (AWS_ACCESS_KEY_ID, AWS_SECRET_ACCESS_KEY) """] , A_ , ) __UpperCAmelCase =None if credentials_configuration == 0: __UpperCAmelCase =_ask_field("""Enter your AWS Profile name: [default] """ , default="""default""" ) __UpperCAmelCase =aws_profile else: print( """Note you will need to provide AWS_ACCESS_KEY_ID and AWS_SECRET_ACCESS_KEY when you launch you training script with,""" """`accelerate launch --aws_access_key_id XXX --aws_secret_access_key YYY`""" ) __UpperCAmelCase =_ask_field("""AWS Access Key ID: """ ) __UpperCAmelCase =aws_access_key_id __UpperCAmelCase =_ask_field("""AWS Secret Access Key: """ ) __UpperCAmelCase =aws_secret_access_key __UpperCAmelCase =_ask_field("""Enter your AWS Region: [us-east-1]""" , default="""us-east-1""" ) __UpperCAmelCase =aws_region __UpperCAmelCase =_ask_options( """Do you already have an IAM Role for executing Amazon SageMaker Training Jobs?""" , ["""Provide IAM Role name""", """Create new IAM role using credentials"""] , A_ , ) if role_management == 0: __UpperCAmelCase =_ask_field("""Enter your IAM role name: """ ) else: __UpperCAmelCase ="""accelerate_sagemaker_execution_role""" print(F'''Accelerate will create an iam role "{iam_role_name}" using the provided credentials''' ) _create_iam_role_for_sagemaker(A_ ) __UpperCAmelCase =_ask_field( """Do you want to use custom Docker image? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_custom_docker_image: __UpperCAmelCase =_ask_field("""Enter your Docker image: """ , lambda A_ : str(A_ ).lower() ) __UpperCAmelCase =_ask_field( """Do you want to provide SageMaker input channels with data locations? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_inputs_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker inputs TSV file with columns (channel_name, data_location): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_field( """Do you want to enable SageMaker metrics? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =None if is_sagemaker_metrics_enabled: __UpperCAmelCase =_ask_field( """Enter the path to the SageMaker metrics TSV file with columns (metric_name, metric_regex): """ , lambda A_ : str(A_ ).lower() , ) __UpperCAmelCase =_ask_options( """What is the distributed mode?""" , ["""No distributed training""", """Data parallelism"""] , _convert_sagemaker_distributed_mode , ) __UpperCAmelCase ={} __UpperCAmelCase =_ask_field( """Do you wish to optimize your script with torch dynamo?[yes/NO]:""" , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_dynamo: __UpperCAmelCase ="""dynamo_""" __UpperCAmelCase =_ask_options( """Which dynamo backend would you like to use?""" , [x.lower() for x in DYNAMO_BACKENDS] , _convert_dynamo_backend , default=2 , ) __UpperCAmelCase =_ask_field( """Do you want to customize the defaults sent to torch.compile? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) if use_custom_options: __UpperCAmelCase =_ask_options( """Which mode do you want to use?""" , A_ , lambda A_ : TORCH_DYNAMO_MODES[int(A_ )] , default="""default""" , ) __UpperCAmelCase =_ask_field( """Do you want the fullgraph mode or it is ok to break model into several subgraphs? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase =_ask_field( """Do you want to enable dynamic shape tracing? [yes/NO]: """ , _convert_yes_no_to_bool , default=A_ , error_message="""Please enter yes or no.""" , ) __UpperCAmelCase ="""Which EC2 instance type you want to use for your training?""" if distributed_type != SageMakerDistributedType.NO: __UpperCAmelCase =_ask_options( A_ , A_ , lambda A_ : SAGEMAKER_PARALLEL_EC2_INSTANCES[int(A_ )] ) else: eca_instance_query += "? [ml.p3.2xlarge]:" __UpperCAmelCase =_ask_field(A_ , lambda A_ : str(A_ ).lower() , default="""ml.p3.2xlarge""" ) __UpperCAmelCase =1 if distributed_type in (SageMakerDistributedType.DATA_PARALLEL, SageMakerDistributedType.MODEL_PARALLEL): __UpperCAmelCase =_ask_field( """How many machines do you want use? [1]: """ , A_ , default=1 , ) __UpperCAmelCase =_ask_options( """Do you wish to use FP16 or BF16 (mixed precision)?""" , ["""no""", """fp16""", """bf16""", """fp8"""] , _convert_mixed_precision , ) if use_dynamo and mixed_precision == "no": print( """Torch dynamo used without mixed precision requires TF32 to be efficient. Accelerate will enable it by default when launching your scripts.""" ) return SageMakerConfig( image_uri=A_ , compute_environment=ComputeEnvironment.AMAZON_SAGEMAKER , distributed_type=A_ , use_cpu=A_ , dynamo_config=A_ , eca_instance_type=A_ , profile=A_ , region=A_ , iam_role_name=A_ , mixed_precision=A_ , num_machines=A_ , sagemaker_inputs_file=A_ , sagemaker_metrics_file=A_ , )
68
0
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os from accelerate.test_utils import execute_subprocess_async def __lowerCamelCase ( UpperCamelCase__=None ): """simple docstring""" if subparsers is not None: _UpperCAmelCase = subparsers.add_parser("test" ) else: _UpperCAmelCase = argparse.ArgumentParser("Accelerate test command" ) parser.add_argument( "--config_file" , default=A_ , help=( "The path to use to store the config file. Will default to a file named default_config.yaml in the cache " "location, which is the content of the environment `HF_HOME` suffixed with 'accelerate', or if you don't have " "such an environment variable, your cache directory ('~/.cache' or the content of `XDG_CACHE_HOME`) suffixed " "with 'huggingface'." ) , ) if subparsers is not None: parser.set_defaults(func=A_ ) return parser def __lowerCamelCase ( UpperCamelCase__ ): """simple docstring""" _UpperCAmelCase = os.path.sep.join(__file__.split(os.path.sep )[:-2] + ["test_utils", "scripts", "test_script.py"] ) if args.config_file is None: _UpperCAmelCase = script_name else: _UpperCAmelCase = f"--config_file={args.config_file} {script_name}" _UpperCAmelCase = ["accelerate-launch"] + test_args.split() _UpperCAmelCase = execute_subprocess_async(A_ , env=os.environ.copy() ) if result.returncode == 0: print("Test is a success! You are ready for your distributed training!" ) def __lowerCamelCase ( ): """simple docstring""" _UpperCAmelCase = test_command_parser() _UpperCAmelCase = parser.parse_args() test_command(A_ ) if __name__ == "__main__": main()
657
from ...configuration_utils import PretrainedConfig from ...utils import logging __A = logging.get_logger(__name__) __A = {"ctrl": "https://huggingface.co/ctrl/resolve/main/config.json"} class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Tuple = 'ctrl' lowerCamelCase : Any = ['past_key_values'] lowerCamelCase : Optional[int] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=246534 , __SCREAMING_SNAKE_CASE : int=256 , __SCREAMING_SNAKE_CASE : Optional[Any]=1280 , __SCREAMING_SNAKE_CASE : Optional[Any]=8192 , __SCREAMING_SNAKE_CASE : int=48 , __SCREAMING_SNAKE_CASE : Union[str, Any]=16 , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : Dict=0.1 , __SCREAMING_SNAKE_CASE : List[Any]=1e-6 , __SCREAMING_SNAKE_CASE : List[str]=0.02 , __SCREAMING_SNAKE_CASE : Union[str, Any]=True , **__SCREAMING_SNAKE_CASE : int , ) -> Any: __UpperCAmelCase =vocab_size __UpperCAmelCase =n_positions __UpperCAmelCase =n_embd __UpperCAmelCase =n_layer __UpperCAmelCase =n_head __UpperCAmelCase =dff __UpperCAmelCase =resid_pdrop __UpperCAmelCase =embd_pdrop __UpperCAmelCase =layer_norm_epsilon __UpperCAmelCase =initializer_range __UpperCAmelCase =use_cache super().__init__(**__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' def _a( UpperCamelCase__ : Optional[int], UpperCamelCase__ : Optional[Any], UpperCamelCase__ : Optional[int] ): '''simple docstring''' if n == 0: return 1 elif n % 2 == 1: return (binary_exponentiation(A_, n - 1, A_ ) * a) % mod else: SCREAMING_SNAKE_CASE__ : List[str] =binary_exponentiation(A_, n / 2, A_ ) return (b * b) % mod # a prime number a_ = 7_0_1 a_ = 1_0_0_0_0_0_0_0_0_0 a_ = 1_0 # using binary exponentiation function, O(log(p)): print((a / b) % p == (a * binary_exponentiation(b, p - 2, p)) % p) print((a / b) % p == (a * b ** (p - 2)) % p)
296
import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() __A = logging.get_logger(__name__) __A = [ ["attention", "attn"], ["encoder_attention", "encoder_attn"], ["q_lin", "q_proj"], ["k_lin", "k_proj"], ["v_lin", "v_proj"], ["out_lin", "out_proj"], ["norm_embeddings", "layernorm_embedding"], ["position_embeddings", "embed_positions"], ["embeddings", "embed_tokens"], ["ffn.lin", "fc"], ] def lowercase__ ( A_: Optional[Any] ) -> Union[str, Any]: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: __UpperCAmelCase =k.replace(A_ , A_ ) if k.startswith("""encoder""" ): __UpperCAmelCase =k.replace(""".attn""" , """.self_attn""" ) __UpperCAmelCase =k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm2""" , """final_layer_norm""" ) elif k.startswith("""decoder""" ): __UpperCAmelCase =k.replace("""norm1""" , """self_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm2""" , """encoder_attn_layer_norm""" ) __UpperCAmelCase =k.replace("""norm3""" , """final_layer_norm""" ) return k def lowercase__ ( A_: Tuple ) -> str: """simple docstring""" __UpperCAmelCase =[ """model.encoder.layernorm_embedding.weight""", """model.encoder.layernorm_embedding.bias""", """model.decoder.layernorm_embedding.weight""", """model.decoder.layernorm_embedding.bias""", ] for k in keys: __UpperCAmelCase =sd.pop(A_ ) __UpperCAmelCase =k.replace("""layernorm_embedding""" , """layer_norm""" ) assert new_k not in sd __UpperCAmelCase =v __A = ["START"] @torch.no_grad() def lowercase__ ( A_: List[Any] , A_: str , A_: int ) -> Optional[int]: """simple docstring""" __UpperCAmelCase =torch.load(A_ , map_location="""cpu""" ) __UpperCAmelCase =model["""model"""] __UpperCAmelCase =BlenderbotConfig.from_json_file(A_ ) __UpperCAmelCase =BlenderbotForConditionalGeneration(A_ ) __UpperCAmelCase =m.model.state_dict().keys() __UpperCAmelCase =[] __UpperCAmelCase ={} for k, v in sd.items(): if k in IGNORE_KEYS: continue __UpperCAmelCase =rename_state_dict_key(A_ ) if new_k not in valid_keys: failures.append([k, new_k] ) else: __UpperCAmelCase =v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(A_ ) m.model.load_state_dict(A_ , strict=A_ ) m.half() m.save_pretrained(A_ ) if __name__ == "__main__": __A = argparse.ArgumentParser() # Required parameters parser.add_argument("--src_path", type=str, help="like blenderbot-model.bin") parser.add_argument("--save_dir", default="hf_blenderbot", type=str, help="Where to save converted model.") parser.add_argument( "--hf_config_json", default="blenderbot-3b-config.json", type=str, help="Path to config to use" ) __A = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
68
0
"""simple docstring""" import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_funnel import FunnelTokenizer A = logging.get_logger(__name__) A = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} A = [ 'small', 'small-base', 'medium', 'medium-base', 'intermediate', 'intermediate-base', 'large', 'large-base', 'xlarge', 'xlarge-base', ] A = { 'vocab_file': { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/vocab.txt', 'funnel-transformer/small-base': 'https://huggingface.co/funnel-transformer/small-base/resolve/main/vocab.txt', 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/vocab.txt', 'funnel-transformer/medium-base': ( 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/vocab.txt' ), 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/vocab.txt' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/vocab.txt' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/vocab.txt', 'funnel-transformer/large-base': 'https://huggingface.co/funnel-transformer/large-base/resolve/main/vocab.txt', 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/vocab.txt', 'funnel-transformer/xlarge-base': ( 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/vocab.txt' ), }, 'tokenizer_file': { 'funnel-transformer/small': 'https://huggingface.co/funnel-transformer/small/resolve/main/tokenizer.json', 'funnel-transformer/small-base': ( 'https://huggingface.co/funnel-transformer/small-base/resolve/main/tokenizer.json' ), 'funnel-transformer/medium': 'https://huggingface.co/funnel-transformer/medium/resolve/main/tokenizer.json', 'funnel-transformer/medium-base': ( 'https://huggingface.co/funnel-transformer/medium-base/resolve/main/tokenizer.json' ), 'funnel-transformer/intermediate': ( 'https://huggingface.co/funnel-transformer/intermediate/resolve/main/tokenizer.json' ), 'funnel-transformer/intermediate-base': ( 'https://huggingface.co/funnel-transformer/intermediate-base/resolve/main/tokenizer.json' ), 'funnel-transformer/large': 'https://huggingface.co/funnel-transformer/large/resolve/main/tokenizer.json', 'funnel-transformer/large-base': ( 'https://huggingface.co/funnel-transformer/large-base/resolve/main/tokenizer.json' ), 'funnel-transformer/xlarge': 'https://huggingface.co/funnel-transformer/xlarge/resolve/main/tokenizer.json', 'funnel-transformer/xlarge-base': ( 'https://huggingface.co/funnel-transformer/xlarge-base/resolve/main/tokenizer.json' ), }, } A = {F'''funnel-transformer/{name}''': 5_1_2 for name in _model_names} A = {F'''funnel-transformer/{name}''': {'do_lower_case': True} for name in _model_names} class _a ( SCREAMING_SNAKE_CASE__): __magic_name__ = VOCAB_FILES_NAMES __magic_name__ = PRETRAINED_VOCAB_FILES_MAP __magic_name__ = PRETRAINED_INIT_CONFIGURATION __magic_name__ = FunnelTokenizer __magic_name__ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __magic_name__ = 2 def __init__( self : List[Any] , _lowercase : str=None , _lowercase : Any=None , _lowercase : int=True , _lowercase : Optional[int]="<unk>" , _lowercase : Any="<sep>" , _lowercase : Optional[int]="<pad>" , _lowercase : Optional[Any]="<cls>" , _lowercase : List[str]="<mask>" , _lowercase : Dict="<s>" , _lowercase : Any="</s>" , _lowercase : str=True , _lowercase : Dict=True , _lowercase : Optional[int]=None , _lowercase : Dict="##" , **_lowercase : Tuple , ) -> List[Any]: super().__init__( __SCREAMING_SNAKE_CASE , tokenizer_file=__SCREAMING_SNAKE_CASE , do_lower_case=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , sep_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , cls_token=__SCREAMING_SNAKE_CASE , mask_token=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , clean_text=__SCREAMING_SNAKE_CASE , tokenize_chinese_chars=__SCREAMING_SNAKE_CASE , strip_accents=__SCREAMING_SNAKE_CASE , wordpieces_prefix=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) snake_case : Tuple = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , __SCREAMING_SNAKE_CASE ) != do_lower_case or normalizer_state.get("strip_accents" , __SCREAMING_SNAKE_CASE ) != strip_accents or normalizer_state.get("handle_chinese_chars" , __SCREAMING_SNAKE_CASE ) != tokenize_chinese_chars ): snake_case : List[Any] = getattr(__SCREAMING_SNAKE_CASE , normalizer_state.pop("type" ) ) snake_case : int = do_lower_case snake_case : int = strip_accents snake_case : Optional[int] = tokenize_chinese_chars snake_case : Optional[Any] = normalizer_class(**__SCREAMING_SNAKE_CASE ) snake_case : int = do_lower_case def __lowercase ( self : str , _lowercase : List[Any] , _lowercase : Union[str, Any]=None ) -> Any: snake_case : int = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self : str , _lowercase : List[int] , _lowercase : Optional[List[int]] = None ) -> List[int]: snake_case : Any = [self.sep_token_id] snake_case : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] return len(cls ) * [self.cls_token_type_id] + len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self : Any , _lowercase : str , _lowercase : Optional[str] = None ) -> Tuple[str]: snake_case : Optional[Any] = self._tokenizer.model.save(__SCREAMING_SNAKE_CASE , name=__SCREAMING_SNAKE_CASE ) return tuple(__SCREAMING_SNAKE_CASE )
449
from itertools import permutations def lowercase__ ( A_: tuple ) -> bool: """simple docstring""" if num[3] % 2 != 0: return False if (num[2] + num[3] + num[4]) % 3 != 0: return False if num[5] % 5 != 0: return False __UpperCAmelCase =[7, 11, 13, 17] for i, test in enumerate(A_ ): if (num[i + 4] * 100 + num[i + 5] * 10 + num[i + 6]) % test != 0: return False return True def lowercase__ ( A_: int = 10 ) -> int: """simple docstring""" return sum( int("""""".join(map(A_ , A_ ) ) ) for num in permutations(range(A_ ) ) if is_substring_divisible(A_ ) ) if __name__ == "__main__": print(F"""{solution() = }""")
68
0
'''simple docstring''' import math def __UpperCamelCase ( a : int ) ->bool: assert isinstance(A_ , A_ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False snake_case = range(3 , int(math.sqrt(A_ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def __UpperCamelCase ( a : Any , a : Optional[int]=1 , **a : int ) ->int: snake_case = factor * value snake_case = value while not is_prime(A_ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **A_ ) return value
342
from __future__ import annotations from sys import maxsize from typing import Generic, TypeVar __A = TypeVar("T") def lowercase__ ( A_: int ) -> int: """simple docstring""" return (position - 1) // 2 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 1 def lowercase__ ( A_: int ) -> int: """simple docstring""" return (2 * position) + 2 class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[str] ) -> None: __UpperCAmelCase =[] __UpperCAmelCase ={} __UpperCAmelCase =0 def __len__( self : str ) -> int: return self.elements def __repr__( self : Dict ) -> str: return str(self.heap ) def _a ( self : Optional[int] ) -> bool: # Check if the priority queue is empty return self.elements == 0 def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an element with given priority to the queue self.heap.append((elem, weight) ) __UpperCAmelCase =self.elements self.elements += 1 self._bubble_up(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> T: # Remove and return the element with lowest weight (highest priority) if self.elements > 1: self._swap_nodes(0 , self.elements - 1 ) __UpperCAmelCase , __UpperCAmelCase =self.heap.pop() del self.position_map[elem] self.elements -= 1 if self.elements > 0: __UpperCAmelCase , __UpperCAmelCase =self.heap[0] self._bubble_down(__SCREAMING_SNAKE_CASE ) return elem def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Update the weight of the given key __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase =(elem, weight) if position > 0: __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._bubble_up(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) else: self._bubble_down(__SCREAMING_SNAKE_CASE ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (upward movement) [to be used internally # only] __UpperCAmelCase =self.position_map[elem] if curr_pos == 0: return None __UpperCAmelCase =get_parent_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase , __UpperCAmelCase =self.heap[parent_position] if parent_weight > weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_up(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : T ) -> None: # Place a node at the proper position (downward movement) [to be used # internally only] __UpperCAmelCase =self.position_map[elem] __UpperCAmelCase , __UpperCAmelCase =self.heap[curr_pos] __UpperCAmelCase =get_child_left_position(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =get_child_right_position(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements and child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < child_left_weight and child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) if child_left_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_left_position] if child_left_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) else: return None if child_right_position < self.elements: __UpperCAmelCase , __UpperCAmelCase =self.heap[child_right_position] if child_right_weight < weight: self._swap_nodes(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return self._bubble_down(__SCREAMING_SNAKE_CASE ) return None def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int ) -> None: # Swap the nodes at the given positions __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase =self.heap[nodea_pos][0] __UpperCAmelCase , __UpperCAmelCase =( self.heap[nodea_pos], self.heap[nodea_pos], ) __UpperCAmelCase =nodea_pos __UpperCAmelCase =nodea_pos class _A ( Generic[T] ): """simple docstring""" def __init__( self : List[Any] ) -> None: __UpperCAmelCase ={} __UpperCAmelCase =0 def __repr__( self : Tuple ) -> str: return str(self.connections ) def __len__( self : str ) -> int: return self.nodes def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : T ) -> None: # Add a node in the graph if it is not in the graph if node not in self.connections: __UpperCAmelCase ={} self.nodes += 1 def _a ( self : Union[str, Any] , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : T , __SCREAMING_SNAKE_CASE : int ) -> None: # Add an edge between 2 nodes in the graph self.add_node(__SCREAMING_SNAKE_CASE ) self.add_node(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =weight __UpperCAmelCase =weight def lowercase__ ( A_: GraphUndirectedWeighted[T] , ) -> tuple[dict[T, int], dict[T, T | None]]: """simple docstring""" __UpperCAmelCase ={node: maxsize for node in graph.connections} __UpperCAmelCase ={node: None for node in graph.connections} __UpperCAmelCase =MinPriorityQueue() for node, weight in dist.items(): priority_queue.push(A_ , A_ ) if priority_queue.is_empty(): return dist, parent # initialization __UpperCAmelCase =priority_queue.extract_min() __UpperCAmelCase =0 for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node # running prim's algorithm while not priority_queue.is_empty(): __UpperCAmelCase =priority_queue.extract_min() for neighbour in graph.connections[node]: if dist[neighbour] > dist[node] + graph.connections[node][neighbour]: __UpperCAmelCase =dist[node] + graph.connections[node][neighbour] priority_queue.update_key(A_ , dist[neighbour] ) __UpperCAmelCase =node return dist, parent
68
0
def lowerCamelCase_(lowerCamelCase_ ) -> list: UpperCAmelCase = [0] * len(A_ ) for i in range(1 , len(A_ ) ): # use last results for better performance - dynamic programming UpperCAmelCase = prefix_result[i - 1] while j > 0 and input_string[i] != input_string[j]: UpperCAmelCase = prefix_result[j - 1] if input_string[i] == input_string[j]: j += 1 UpperCAmelCase = j return prefix_result def lowerCamelCase_(lowerCamelCase_ ) -> int: return max(prefix_function(A_ ) ) if __name__ == "__main__": import doctest doctest.testmod()
323
from dataclasses import dataclass, field from typing import Tuple from ..utils import cached_property, is_tf_available, logging, requires_backends from .benchmark_args_utils import BenchmarkArguments if is_tf_available(): import tensorflow as tf __A = logging.get_logger(__name__) @dataclass class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[int] = [ 'no_inference', 'no_cuda', 'no_tpu', 'no_speed', 'no_memory', 'no_env_print', 'no_multi_process', ] def __init__( self : Any , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Dict: for deprecated_arg in self.deprecated_args: if deprecated_arg in kwargs: __UpperCAmelCase =deprecated_arg[3:] __UpperCAmelCase =not kwargs.pop(__SCREAMING_SNAKE_CASE ) logger.warning( f'''{deprecated_arg} is depreciated. Please use --no-{positive_arg} or''' f''' {positive_arg}={kwargs[positive_arg]}''' ) __UpperCAmelCase =kwargs.pop("""tpu_name""" , self.tpu_name ) __UpperCAmelCase =kwargs.pop("""device_idx""" , self.device_idx ) __UpperCAmelCase =kwargs.pop("""eager_mode""" , self.eager_mode ) __UpperCAmelCase =kwargs.pop("""use_xla""" , self.use_xla ) super().__init__(**__SCREAMING_SNAKE_CASE ) lowerCamelCase : str = field( default=UpperCamelCase , metadata={'help': 'Name of TPU'} , ) lowerCamelCase : int = field( default=0 , metadata={'help': 'CPU / GPU device index. Defaults to 0.'} , ) lowerCamelCase : bool = field(default=UpperCamelCase , metadata={'help': 'Benchmark models in eager model.'} ) lowerCamelCase : bool = field( default=UpperCamelCase , metadata={ 'help': 'Benchmark models using XLA JIT compilation. Note that `eager_model` has to be set to `False`.' } , ) @cached_property def _a ( self : List[str] ) -> Tuple["tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) __UpperCAmelCase =None if self.tpu: try: if self.tpu_name: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver(self.tpu_name ) else: __UpperCAmelCase =tf.distribute.cluster_resolver.TPUClusterResolver() except ValueError: __UpperCAmelCase =None return tpu @cached_property def _a ( self : Tuple ) -> Tuple["tf.distribute.Strategy", "tf.distribute.cluster_resolver.TPUClusterResolver"]: requires_backends(self , ["""tf"""] ) if self.is_tpu: tf.config.experimental_connect_to_cluster(self._setup_tpu ) tf.tpu.experimental.initialize_tpu_system(self._setup_tpu ) __UpperCAmelCase =tf.distribute.TPUStrategy(self._setup_tpu ) else: # currently no multi gpu is allowed if self.is_gpu: # TODO: Currently only single GPU is supported tf.config.set_visible_devices(self.gpu_list[self.device_idx] , """GPU""" ) __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/gpu:{self.device_idx}''' ) else: tf.config.set_visible_devices([] , """GPU""" ) # disable GPU __UpperCAmelCase =tf.distribute.OneDeviceStrategy(device=f'''/cpu:{self.device_idx}''' ) return strategy @property def _a ( self : Optional[Any] ) -> bool: requires_backends(self , ["""tf"""] ) return self._setup_tpu is not None @property def _a ( self : str ) -> "tf.distribute.Strategy": requires_backends(self , ["""tf"""] ) return self._setup_strategy @property def _a ( self : Dict ) -> Optional[int]: requires_backends(self , ["""tf"""] ) return tf.config.list_physical_devices("""GPU""" ) @property def _a ( self : List[str] ) -> int: requires_backends(self , ["""tf"""] ) if self.cuda: return len(self.gpu_list ) return 0 @property def _a ( self : List[str] ) -> bool: return self.n_gpu > 0
68
0
'''simple docstring''' from __future__ import annotations from PIL import Image # Define glider example __lowerCAmelCase : int = [ [0, 1, 0, 0, 0, 0, 0, 0], [0, 0, 1, 0, 0, 0, 0, 0], [1, 1, 1, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0], ] # Define blinker example __lowerCAmelCase : Optional[int] = [[0, 1, 0], [0, 1, 0], [0, 1, 0]] def lowerCAmelCase ( UpperCamelCase__ : list[list[int]] ): """simple docstring""" __UpperCAmelCase = [] for i in range(len(A_ ) ): __UpperCAmelCase = [] for j in range(len(cells[i] ) ): # Get the number of live neighbours __UpperCAmelCase = 0 if i > 0 and j > 0: neighbour_count += cells[i - 1][j - 1] if i > 0: neighbour_count += cells[i - 1][j] if i > 0 and j < len(cells[i] ) - 1: neighbour_count += cells[i - 1][j + 1] if j > 0: neighbour_count += cells[i][j - 1] if j < len(cells[i] ) - 1: neighbour_count += cells[i][j + 1] if i < len(A_ ) - 1 and j > 0: neighbour_count += cells[i + 1][j - 1] if i < len(A_ ) - 1: neighbour_count += cells[i + 1][j] if i < len(A_ ) - 1 and j < len(cells[i] ) - 1: neighbour_count += cells[i + 1][j + 1] # Rules of the game of life (excerpt from Wikipedia): # 1. Any live cell with two or three live neighbours survives. # 2. Any dead cell with three live neighbours becomes a live cell. # 3. All other live cells die in the next generation. # Similarly, all other dead cells stay dead. __UpperCAmelCase = cells[i][j] == 1 if ( (alive and 2 <= neighbour_count <= 3) or not alive and neighbour_count == 3 ): next_generation_row.append(1 ) else: next_generation_row.append(0 ) next_generation.append(A_ ) return next_generation def lowerCAmelCase ( UpperCamelCase__ : list[list[int]] , UpperCamelCase__ : int ): """simple docstring""" __UpperCAmelCase = [] for _ in range(A_ ): # Create output image __UpperCAmelCase = Image.new('''RGB''' , (len(cells[0] ), len(A_ )) ) __UpperCAmelCase = img.load() # Save cells to image for x in range(len(A_ ) ): for y in range(len(cells[0] ) ): __UpperCAmelCase = 2_5_5 - cells[y][x] * 2_5_5 __UpperCAmelCase = (colour, colour, colour) # Save image images.append(A_ ) __UpperCAmelCase = new_generation(A_ ) return images if __name__ == "__main__": __lowerCAmelCase : Tuple = generate_images(GLIDER, 16) images[0].save("out.gif", save_all=True, append_images=images[1:])
262
import unittest import numpy as np import torch from diffusers import ScoreSdeVePipeline, ScoreSdeVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Dict: torch.manual_seed(0 ) __UpperCAmelCase =UNetaDModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=3 , out_channels=3 , down_block_types=("""DownBlock2D""", """AttnDownBlock2D""") , up_block_types=("""AttnUpBlock2D""", """UpBlock2D""") , ) return model def _a ( self : int ) -> Union[str, Any]: __UpperCAmelCase =self.dummy_uncond_unet __UpperCAmelCase =ScoreSdeVeScheduler() __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=2 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE , return_dict=__SCREAMING_SNAKE_CASE )[ 0 ] __UpperCAmelCase =image[0, -3:, -3:, -1] __UpperCAmelCase =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) __UpperCAmelCase =np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1e-2 @slow @require_torch class _A ( unittest.TestCase ): """simple docstring""" def _a ( self : Optional[int] ) -> int: __UpperCAmelCase ="""google/ncsnpp-church-256""" __UpperCAmelCase =UNetaDModel.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVeScheduler.from_pretrained(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ScoreSdeVePipeline(unet=__SCREAMING_SNAKE_CASE , scheduler=__SCREAMING_SNAKE_CASE ) sde_ve.to(__SCREAMING_SNAKE_CASE ) sde_ve.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.manual_seed(0 ) __UpperCAmelCase =sde_ve(num_inference_steps=10 , output_type="""numpy""" , generator=__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) __UpperCAmelCase =np.array([0.0, 1.0, 1.0, 1.0, 1.0, 1.0, 1.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2
68
0
"""simple docstring""" import torch from transformers import AutoModel class UpperCamelCase ( torch.nn.Module ): def __init__( self ,__UpperCamelCase="sayef/fsner-bert-base-uncased" ) -> Optional[int]: '''simple docstring''' super(__SCREAMING_SNAKE_CASE ,self ).__init__() lowercase_ : Any = AutoModel.from_pretrained(__SCREAMING_SNAKE_CASE ,return_dict=__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = torch.nn.CosineSimilarity(3 ,1e-08 ) lowercase_ : Optional[Any] = torch.nn.Softmax(dim=1 ) def _UpperCAmelCase ( self ,**__UpperCamelCase ) -> List[Any]: '''simple docstring''' return self.bert(**__SCREAMING_SNAKE_CASE ).last_hidden_state def _UpperCAmelCase ( self ,__UpperCamelCase ) -> str: '''simple docstring''' return token_embeddings.sum(2 ,keepdim=__SCREAMING_SNAKE_CASE ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase=1 ) -> Optional[int]: '''simple docstring''' return self.softmax(T * self.cos(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) ) def _UpperCAmelCase ( self ,__UpperCamelCase ,__UpperCamelCase ) -> Any: '''simple docstring''' lowercase_ : List[Any] = W_supports['sizes'].tolist() lowercase_ : int = W_supports['start_token_id'].item() lowercase_ : Optional[int] = W_supports['end_token_id'].item() del W_supports["sizes"] del W_supports["start_token_id"] del W_supports["end_token_id"] lowercase_ : List[Any] = self.BERT(**__SCREAMING_SNAKE_CASE ) lowercase_ : List[str] = self.BERT(**__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = None lowercase_ : Dict = None lowercase_ : Union[str, Any] = W_supports['input_ids'] == start_token_id lowercase_ : Dict = W_supports['input_ids'] == end_token_id for i, size in enumerate(__SCREAMING_SNAKE_CASE ): if i == 0: lowercase_ : str = 0 else: lowercase_ : Optional[Any] = support_sizes[i - 1] lowercase_ : int = S[s : s + size][start_token_masks[s : s + size]] lowercase_ : Optional[int] = S[s : s + size][end_token_masks[s : s + size]] lowercase_ : int = torch.matmul(q[i] ,s_start.T ).sum(1 ).softmax(0 ) lowercase_ : Dict = torch.matmul(q[i] ,s_end.T ).sum(1 ).softmax(0 ) if p_starts is not None: lowercase_ : int = torch.vstack((p_starts, p_start) ) lowercase_ : Any = torch.vstack((p_ends, p_end) ) else: lowercase_ : Tuple = p_start lowercase_ : Union[str, Any] = p_end return p_starts, p_ends
425
from typing import Any, Dict, List, Optional, Tuple, Union import torch from torch import nn from torch.utils.data import DistributedSampler, RandomSampler from transformers import PreTrainedModel, Trainer, logging from transformers.integrations import is_fairscale_available from transformers.models.fsmt.configuration_fsmt import FSMTConfig from transformers.optimization import ( Adafactor, AdamW, get_constant_schedule, get_constant_schedule_with_warmup, get_cosine_schedule_with_warmup, get_cosine_with_hard_restarts_schedule_with_warmup, get_linear_schedule_with_warmup, get_polynomial_decay_schedule_with_warmup, ) from transformers.trainer_pt_utils import get_tpu_sampler from transformers.training_args import ParallelMode from transformers.utils import is_torch_tpu_available if is_fairscale_available(): from fairscale.optim import OSS __A = logging.get_logger(__name__) __A = { "linear": get_linear_schedule_with_warmup, "cosine": get_cosine_schedule_with_warmup, "cosine_w_restarts": get_cosine_with_hard_restarts_schedule_with_warmup, "polynomial": get_polynomial_decay_schedule_with_warmup, "constant": get_constant_schedule, "constant_w_warmup": get_constant_schedule_with_warmup, } class _A ( UpperCamelCase ): """simple docstring""" def __init__( self : List[str] , __SCREAMING_SNAKE_CASE : Union[str, Any]=None , __SCREAMING_SNAKE_CASE : str=None , *__SCREAMING_SNAKE_CASE : Union[str, Any] , **__SCREAMING_SNAKE_CASE : List[Any] ) -> Any: super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if config is None: assert isinstance(self.model , __SCREAMING_SNAKE_CASE ), ( "If no `config` is passed the model to be trained has to be of type `PreTrainedModel`, but is" f''' {self.model.__class__}''' ) __UpperCAmelCase =self.model.config else: __UpperCAmelCase =config __UpperCAmelCase =data_args __UpperCAmelCase =self.config.tgt_vocab_size if isinstance(self.config , __SCREAMING_SNAKE_CASE ) else self.config.vocab_size if self.args.label_smoothing != 0 or (self.data_args is not None and self.data_args.ignore_pad_token_for_loss): assert self.config.pad_token_id is not None, ( "Make sure that `config.pad_token_id` is correcly defined when ignoring `pad_token` for loss" " calculation or doing label smoothing." ) if self.config.pad_token_id is None and self.config.eos_token_id is not None: logger.warning( f'''The `config.pad_token_id` is `None`. Using `config.eos_token_id` = {self.config.eos_token_id} for''' """ padding..""" ) if self.args.label_smoothing == 0: __UpperCAmelCase =torch.nn.CrossEntropyLoss(ignore_index=self.config.pad_token_id ) else: # dynamically import label_smoothed_nll_loss from utils import label_smoothed_nll_loss __UpperCAmelCase =label_smoothed_nll_loss def _a ( self : Any , __SCREAMING_SNAKE_CASE : int ) -> Any: if self.optimizer is None: __UpperCAmelCase =["""bias""", """LayerNorm.weight"""] __UpperCAmelCase =[ { """params""": [p for n, p in self.model.named_parameters() if not any(nd in n for nd in no_decay )], """weight_decay""": self.args.weight_decay, }, { """params""": [p for n, p in self.model.named_parameters() if any(nd in n for nd in no_decay )], """weight_decay""": 0.0, }, ] __UpperCAmelCase =Adafactor if self.args.adafactor else AdamW if self.args.adafactor: __UpperCAmelCase =Adafactor __UpperCAmelCase ={"""scale_parameter""": False, """relative_step""": False} else: __UpperCAmelCase =AdamW __UpperCAmelCase ={ """betas""": (self.args.adam_betaa, self.args.adam_betaa), """eps""": self.args.adam_epsilon, } __UpperCAmelCase =self.args.learning_rate if self.sharded_ddp: __UpperCAmelCase =OSS( params=__SCREAMING_SNAKE_CASE , optim=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) else: __UpperCAmelCase =optimizer_cls(__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) if self.lr_scheduler is None: __UpperCAmelCase =self._get_lr_scheduler(__SCREAMING_SNAKE_CASE ) else: # ignoring --lr_scheduler logger.warning("""scheduler is passed to `Seq2SeqTrainer`, `--lr_scheduler` arg is ignored.""" ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[int] ) -> Any: __UpperCAmelCase =arg_to_scheduler[self.args.lr_scheduler] if self.args.lr_scheduler == "constant": __UpperCAmelCase =schedule_func(self.optimizer ) elif self.args.lr_scheduler == "constant_w_warmup": __UpperCAmelCase =schedule_func(self.optimizer , num_warmup_steps=self.args.warmup_steps ) else: __UpperCAmelCase =schedule_func( self.optimizer , num_warmup_steps=self.args.warmup_steps , num_training_steps=__SCREAMING_SNAKE_CASE ) return scheduler def _a ( self : Optional[Any] ) -> Optional[torch.utils.data.Sampler]: if isinstance(self.train_dataset , torch.utils.data.IterableDataset ): return None elif is_torch_tpu_available(): return get_tpu_sampler(self.train_dataset ) else: if self.args.sortish_sampler: self.train_dataset.make_sortish_sampler( self.args.per_device_train_batch_size , distributed=(self.args.parallel_mode == ParallelMode.DISTRIBUTED) , ) return ( RandomSampler(self.train_dataset ) if self.args.local_rank == -1 else DistributedSampler(self.train_dataset ) ) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Tuple: if self.args.label_smoothing == 0: if self.data_args is not None and self.data_args.ignore_pad_token_for_loss: # force training to ignore pad token __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =self.loss_fn(logits.view(-1 , logits.shape[-1] ) , labels.view(-1 ) ) else: # compute usual loss via models __UpperCAmelCase , __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[:2] else: # compute label smoothed loss __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE , use_cache=__SCREAMING_SNAKE_CASE )[0] __UpperCAmelCase =torch.nn.functional.log_softmax(__SCREAMING_SNAKE_CASE , dim=-1 ) __UpperCAmelCase , __UpperCAmelCase =self.loss_fn(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.args.label_smoothing , ignore_index=self.config.pad_token_id ) return loss, logits def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] ) -> Dict: __UpperCAmelCase =inputs.pop("""labels""" ) __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) return loss def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : nn.Module , __SCREAMING_SNAKE_CASE : Dict[str, Union[torch.Tensor, Any]] , __SCREAMING_SNAKE_CASE : bool , __SCREAMING_SNAKE_CASE : Optional[List[str]] = None , ) -> Tuple[Optional[float], Optional[torch.Tensor], Optional[torch.Tensor]]: __UpperCAmelCase =self._prepare_inputs(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """max_length""": self.data_args.val_max_target_length if self.data_args is not None else self.config.max_length, """num_beams""": self.data_args.eval_beams if self.data_args is not None else self.config.num_beams, } if self.args.predict_with_generate and not self.args.prediction_loss_only: __UpperCAmelCase =self.model.generate( inputs["""input_ids"""] , attention_mask=inputs["""attention_mask"""] , **__SCREAMING_SNAKE_CASE , ) # in case the batch is shorter than max length, the output should be padded if generated_tokens.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) __UpperCAmelCase =inputs.pop("""labels""" ) with torch.no_grad(): # compute loss on predict data __UpperCAmelCase , __UpperCAmelCase =self._compute_loss(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =loss.mean().detach() if self.args.prediction_loss_only: return (loss, None, None) __UpperCAmelCase =generated_tokens if self.args.predict_with_generate else logits if labels.shape[-1] < gen_kwargs["max_length"]: __UpperCAmelCase =self._pad_tensors_to_max_len(__SCREAMING_SNAKE_CASE , gen_kwargs["""max_length"""] ) return (loss, logits, labels) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : int ) -> List[Any]: # If PAD token is not defined at least EOS token has to be defined __UpperCAmelCase =self.config.pad_token_id if self.config.pad_token_id is not None else self.config.eos_token_id if pad_token_id is None: raise ValueError( """Make sure that either `config.pad_token_id` or `config.eos_token_id` is defined if tensor has to be""" f''' padded to `max_length`={max_length}''' ) __UpperCAmelCase =pad_token_id * torch.ones( (tensor.shape[0], max_length) , dtype=tensor.dtype , device=tensor.device ) __UpperCAmelCase =tensor return padded_tensor
68
0
"""simple docstring""" import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation UpperCamelCase_ : Tuple = logging.get_logger(__name__) UpperCamelCase_ : int = {'''vocab_file''': '''spiece.model'''} UpperCamelCase_ : List[str] = { '''vocab_file''': { '''AI-Sweden/gpt-sw3-126m''': '''https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-350m''': '''https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-1.6b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-6.7b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model''', '''AI-Sweden/gpt-sw3-20b''': '''https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model''', } } UpperCamelCase_ : Union[str, Any] = { '''AI-Sweden/gpt-sw3-126m''': 2048, '''AI-Sweden/gpt-sw3-350m''': 2048, '''AI-Sweden/gpt-sw3-1.6b''': 2048, '''AI-Sweden/gpt-sw3-6.7b''': 2048, '''AI-Sweden/gpt-sw3-20b''': 2048, } class __lowerCAmelCase ( _lowercase ): """simple docstring""" snake_case = VOCAB_FILES_NAMES snake_case = PRETRAINED_VOCAB_FILES_MAP snake_case = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case = ['input_ids', 'attention_mask'] def __init__( self : Dict , _snake_case : Union[str, Any] , _snake_case : Optional[Any]=False , _snake_case : Union[str, Any]=False , _snake_case : List[str]=False , _snake_case : Dict=None , _snake_case : List[Any]=None , _snake_case : Dict=None , _snake_case : Any=None , _snake_case : Optional[Dict[str, Any]] = None , **_snake_case : Optional[Any] , ) -> None: """simple docstring""" A_ = {} if sp_model_kwargs is None else sp_model_kwargs A_ = kwargs.get("name_or_path" ) if name_or_path is None: logger.warning( "name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b," " you are testing the model, this can safely be ignored" ) A_ = "None" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing A_ = "<|endoftext|>" if eos_token is None else eos_token A_ = "<unk>" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: A_ = unk_token if pad_token is None else pad_token A_ = eos_token if bos_token is None else bos_token else: A_ = "<pad>" if pad_token is None else pad_token A_ = "<s>" if bos_token is None else bos_token super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) A_ = do_lower_case A_ = remove_space A_ = keep_accents A_ = vocab_file A_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) # Used for whitespace normalization in input texts # fmt : off A_ = {" ", " ", " ", " ", " ", " ", " ", " ", " ", " ", "", "„"} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing A_ = re.compile( F'[{"".join(map(__SCREAMING_SNAKE_CASE , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8_203] ) )}]' ) def __getstate__( self : Any ) -> str: """simple docstring""" A_ = self.__dict__.copy() A_ = None return state def __setstate__( self : str , _snake_case : Optional[Any] ) -> Union[str, Any]: """simple docstring""" A_ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): A_ = {} A_ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def lowerCamelCase__ ( self : Union[str, Any] ) -> int: """simple docstring""" return len(self.sp_model ) def lowerCamelCase__ ( self : Dict , _snake_case : str ) -> str: """simple docstring""" A_ = self.non_printing_characters_re.sub("" , __SCREAMING_SNAKE_CASE ) # Normalize whitespaces A_ = "".join([char if char not in self.whitespaces else " " for char in text] ) # NFC Unicode normalization A_ = unicodedata.normalize("NFC" , __SCREAMING_SNAKE_CASE ) return text def lowerCamelCase__ ( self : List[str] , _snake_case : str , **_snake_case : Union[str, Any] ) -> List[str]: """simple docstring""" A_ = self.preprocess_text(__SCREAMING_SNAKE_CASE ) return self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : List[str] , _snake_case : str ) -> int: """simple docstring""" return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : Tuple , _snake_case : int ) -> str: """simple docstring""" return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) @staticmethod def lowerCamelCase__ ( _snake_case : str ) -> str: """simple docstring""" return out_string def lowerCamelCase__ ( self : Any , _snake_case : List[str] ) -> str: """simple docstring""" A_ = [] A_ = "" A_ = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) + token A_ = True A_ = [] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE ) A_ = False out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) return out_string def lowerCamelCase__ ( self : Any ) -> Dict[str, int]: """simple docstring""" A_ = {self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase__ ( self : Any , _snake_case : str , _snake_case : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A_ = os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE , "wb" ) as fi: A_ = self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def lowerCamelCase__ ( self : List[Any] , _snake_case : Union[str, List[str]] , _snake_case : Union[str, bool] = False ) -> Union[List[int], List[List[int]], "torch.Tensor"]: """simple docstring""" if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): A_ = self.preprocess_text(__SCREAMING_SNAKE_CASE ) A_ = self.sp_model.encode(__SCREAMING_SNAKE_CASE ) else: A_ = [self.preprocess_text(__SCREAMING_SNAKE_CASE ) for t in text] A_ = self.sp_model.encode(__SCREAMING_SNAKE_CASE ) if return_tensors is True or return_tensors == "pt": A_ = torch.tensor(__SCREAMING_SNAKE_CASE ) return token_ids def lowerCamelCase__ ( self : str , _snake_case : Union[int, List[int]] ) -> str: """simple docstring""" return self.sp_model.decode(__SCREAMING_SNAKE_CASE ) def lowerCamelCase__ ( self : Optional[Any] , _snake_case : "Conversation" ) -> List[int]: """simple docstring""" A_ = [F'User: {text}' if is_user else F'Bot: {text}' for is_user, text in conversation.iter_texts()] A_ = ( F'{self.eos_token}{self.bos_token}' + F'{self.bos_token}'.join(__SCREAMING_SNAKE_CASE ) + F'{self.bos_token}Bot:' ) return self.encode(text=__SCREAMING_SNAKE_CASE )
115
import random import unittest import numpy as np from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionImgaImgPipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : List[Any] = 'hf-internal-testing/tiny-random-OnnxStableDiffusionPipeline' def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str=0 ) -> Any: __UpperCAmelCase =floats_tensor((1, 3, 128, 128) , rng=random.Random(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =np.random.RandomState(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ={ """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """generator""": generator, """num_inference_steps""": 3, """strength""": 0.75, """guidance_scale""": 7.5, """output_type""": """numpy""", } return inputs def _a ( self : Optional[Any] ) -> int: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1].flatten() assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.69_643, 0.58_484, 0.50_314, 0.58_760, 0.55_368, 0.59_643, 0.51_529, 0.41_217, 0.49_087] ) assert np.abs(image_slice - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Union[str, Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=__SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.61_737, 0.54_642, 0.53_183, 0.54_465, 0.52_742, 0.60_525, 0.49_969, 0.40_655, 0.48_154] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Optional[Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =LMSDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) # warmup pass to apply optimizations __UpperCAmelCase =pipe(**self.get_dummy_inputs() ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_761, 0.59_977, 0.49_033, 0.49_619, 0.54_282, 0.50_311, 0.47_600, 0.40_918, 0.45_203] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Optional[Any]: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.52_911, 0.60_004, 0.49_229, 0.49_805, 0.54_502, 0.50_680, 0.47_777, 0.41_028, 0.45_304] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 def _a ( self : Union[str, Any] ) -> Dict: __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained(self.hub_checkpoint , provider="""CPUExecutionProvider""" ) __UpperCAmelCase =DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.get_dummy_inputs() __UpperCAmelCase =pipe(**__SCREAMING_SNAKE_CASE ).images __UpperCAmelCase =image[0, -3:, -3:, -1] assert image.shape == (1, 128, 128, 3) __UpperCAmelCase =np.array([0.65_331, 0.58_277, 0.48_204, 0.56_059, 0.53_665, 0.56_235, 0.50_969, 0.40_009, 0.46_552] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-1 @nightly @require_onnxruntime @require_torch_gpu class _A ( unittest.TestCase ): """simple docstring""" @property def _a ( self : List[str] ) -> Optional[int]: return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _a ( self : Dict ) -> int: __UpperCAmelCase =ort.SessionOptions() __UpperCAmelCase =False return options def _a ( self : Dict ) -> Any: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) # using the PNDM scheduler by default __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """CompVis/stable-diffusion-v1-4""" , revision="""onnx""" , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=10 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.4_909, 0.5_059, 0.5_372, 0.4_623, 0.4_876, 0.5_049, 0.4_820, 0.4_956, 0.5_019] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2 def _a ( self : List[str] ) -> str: __UpperCAmelCase =load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/img2img/sketch-mountains-input.jpg""" ) __UpperCAmelCase =init_image.resize((768, 512) ) __UpperCAmelCase =LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-v1-5""" , subfolder="""scheduler""" , revision="""onnx""" ) __UpperCAmelCase =OnnxStableDiffusionImgaImgPipeline.from_pretrained( """runwayml/stable-diffusion-v1-5""" , revision="""onnx""" , scheduler=__SCREAMING_SNAKE_CASE , safety_checker=__SCREAMING_SNAKE_CASE , feature_extractor=__SCREAMING_SNAKE_CASE , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase ="""A fantasy landscape, trending on artstation""" __UpperCAmelCase =np.random.RandomState(0 ) __UpperCAmelCase =pipe( prompt=__SCREAMING_SNAKE_CASE , image=__SCREAMING_SNAKE_CASE , strength=0.75 , guidance_scale=7.5 , num_inference_steps=20 , generator=__SCREAMING_SNAKE_CASE , output_type="""np""" , ) __UpperCAmelCase =output.images __UpperCAmelCase =images[0, 255:258, 383:386, -1] assert images.shape == (1, 512, 768, 3) __UpperCAmelCase =np.array([0.8_043, 0.926, 0.9_581, 0.8_119, 0.8_954, 0.913, 0.7_209, 0.7_463, 0.7_431] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2e-2
68
0
from ...configuration_utils import PretrainedConfig from ...utils import logging __snake_case :Tuple =logging.get_logger(__name__) __snake_case :str ={ 'google/realm-cc-news-pretrained-embedder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-embedder/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-encoder': ( 'https://huggingface.co/google/realm-cc-news-pretrained-encoder/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-scorer': ( 'https://huggingface.co/google/realm-cc-news-pretrained-scorer/resolve/main/config.json' ), 'google/realm-cc-news-pretrained-openqa': ( 'https://huggingface.co/google/realm-cc-news-pretrained-openqa/aresolve/main/config.json' ), 'google/realm-orqa-nq-openqa': 'https://huggingface.co/google/realm-orqa-nq-openqa/resolve/main/config.json', 'google/realm-orqa-nq-reader': 'https://huggingface.co/google/realm-orqa-nq-reader/resolve/main/config.json', 'google/realm-orqa-wq-openqa': 'https://huggingface.co/google/realm-orqa-wq-openqa/resolve/main/config.json', 'google/realm-orqa-wq-reader': 'https://huggingface.co/google/realm-orqa-wq-reader/resolve/main/config.json', # See all REALM models at https://huggingface.co/models?filter=realm } class lowerCAmelCase__ ( _lowerCamelCase ): A_ : Optional[Any] = 'realm' def __init__( self : str , __UpperCamelCase : Union[str, Any]=30_522 , __UpperCamelCase : List[str]=768 , __UpperCamelCase : Any=128 , __UpperCamelCase : Any=12 , __UpperCamelCase : str=12 , __UpperCamelCase : Optional[int]=8 , __UpperCamelCase : int=3_072 , __UpperCamelCase : Optional[Any]="gelu_new" , __UpperCamelCase : str=0.1 , __UpperCamelCase : Union[str, Any]=0.1 , __UpperCamelCase : Dict=512 , __UpperCamelCase : Tuple=2 , __UpperCamelCase : List[str]=0.0_2 , __UpperCamelCase : Any=1e-12 , __UpperCamelCase : List[str]=256 , __UpperCamelCase : Any=10 , __UpperCamelCase : Any=1e-3 , __UpperCamelCase : Dict=5 , __UpperCamelCase : Any=320 , __UpperCamelCase : List[Any]=13_353_718 , __UpperCamelCase : List[str]=5_000 , __UpperCamelCase : List[str]=1 , __UpperCamelCase : Optional[Any]=0 , __UpperCamelCase : Union[str, Any]=2 , **__UpperCamelCase : int , ) -> Optional[Any]: super().__init__(pad_token_id=__SCREAMING_SNAKE_CASE , bos_token_id=__SCREAMING_SNAKE_CASE , eos_token_id=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) # Common config A = vocab_size A = max_position_embeddings A = hidden_size A = retriever_proj_size A = num_hidden_layers A = num_attention_heads A = num_candidates A = intermediate_size A = hidden_act A = hidden_dropout_prob A = attention_probs_dropout_prob A = initializer_range A = type_vocab_size A = layer_norm_eps # Reader config A = span_hidden_size A = max_span_width A = reader_layer_norm_eps A = reader_beam_size A = reader_seq_len # Retrieval config A = num_block_records A = searcher_beam_size
106
import argparse import glob import logging import os import time from argparse import Namespace import numpy as np import torch from lightning_base import BaseTransformer, add_generic_args, generic_train from torch.utils.data import DataLoader, TensorDataset from transformers import glue_compute_metrics as compute_metrics from transformers import glue_convert_examples_to_features as convert_examples_to_features from transformers import glue_output_modes, glue_tasks_num_labels from transformers import glue_processors as processors __A = logging.getLogger(__name__) class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : Optional[Any] = 'sequence-classification' def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Tuple ) -> Optional[Any]: if type(__SCREAMING_SNAKE_CASE ) == dict: __UpperCAmelCase =Namespace(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =glue_output_modes[hparams.task] __UpperCAmelCase =glue_tasks_num_labels[hparams.task] super().__init__(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , self.mode ) def _a ( self : str , **__SCREAMING_SNAKE_CASE : Dict ) -> List[str]: return self.model(**__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Dict ) -> List[Any]: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =outputs[0] __UpperCAmelCase =self.trainer.lr_schedulers[0]["""scheduler"""] __UpperCAmelCase ={"""loss""": loss, """rate""": lr_scheduler.get_last_lr()[-1]} return {"loss": loss, "log": tensorboard_logs} def _a ( self : Tuple ) -> List[Any]: __UpperCAmelCase =self.hparams __UpperCAmelCase =processors[args.task]() __UpperCAmelCase =processor.get_labels() for mode in ["train", "dev"]: __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) if os.path.exists(__SCREAMING_SNAKE_CASE ) and not args.overwrite_cache: logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) else: logger.info("""Creating features from dataset file at %s""" , args.data_dir ) __UpperCAmelCase =( processor.get_dev_examples(args.data_dir ) if mode == """dev""" else processor.get_train_examples(args.data_dir ) ) __UpperCAmelCase =convert_examples_to_features( __SCREAMING_SNAKE_CASE , self.tokenizer , max_length=args.max_seq_length , label_list=self.labels , output_mode=args.glue_output_mode , ) logger.info("""Saving features into cached file %s""" , __SCREAMING_SNAKE_CASE ) torch.save(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : bool = False ) -> DataLoader: __UpperCAmelCase ="""dev""" if mode == """test""" else mode __UpperCAmelCase =self._feature_file(__SCREAMING_SNAKE_CASE ) logger.info("""Loading features from cached file %s""" , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.load(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([f.input_ids for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.attention_mask for f in features] , dtype=torch.long ) __UpperCAmelCase =torch.tensor([f.token_type_ids for f in features] , dtype=torch.long ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.long ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =torch.tensor([f.label for f in features] , dtype=torch.float ) return DataLoader( TensorDataset(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , batch_size=__SCREAMING_SNAKE_CASE , shuffle=__SCREAMING_SNAKE_CASE , ) def _a ( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : int ) -> str: __UpperCAmelCase ={"""input_ids""": batch[0], """attention_mask""": batch[1], """labels""": batch[3]} if self.config.model_type not in ["distilbert", "bart"]: __UpperCAmelCase =batch[2] if self.config.model_type in ["""bert""", """xlnet""", """albert"""] else None __UpperCAmelCase =self(**__SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase =outputs[:2] __UpperCAmelCase =logits.detach().cpu().numpy() __UpperCAmelCase =inputs["""labels"""].detach().cpu().numpy() return {"val_loss": tmp_eval_loss.detach().cpu(), "pred": preds, "target": out_label_ids} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : Any ) -> tuple: __UpperCAmelCase =torch.stack([x["""val_loss"""] for x in outputs] ).mean().detach().cpu().item() __UpperCAmelCase =np.concatenate([x["""pred"""] for x in outputs] , axis=0 ) if self.hparams.glue_output_mode == "classification": __UpperCAmelCase =np.argmax(__SCREAMING_SNAKE_CASE , axis=1 ) elif self.hparams.glue_output_mode == "regression": __UpperCAmelCase =np.squeeze(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =np.concatenate([x["""target"""] for x in outputs] , axis=0 ) __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase =[[] for _ in range(out_label_ids.shape[0] )] __UpperCAmelCase ={**{"""val_loss""": val_loss_mean}, **compute_metrics(self.hparams.task , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )} __UpperCAmelCase =dict(results.items() ) __UpperCAmelCase =results return ret, preds_list, out_label_list def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : list ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] return {"val_loss": logs["val_loss"], "log": logs, "progress_bar": logs} def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> dict: __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =self._eval_end(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =ret["""log"""] # `val_loss` is the key returned by `self._eval_end()` but actually refers to `test_loss` return {"avg_test_loss": logs["val_loss"], "log": logs, "progress_bar": logs} @staticmethod def _a ( __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[Any]: BaseTransformer.add_model_specific_args(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) parser.add_argument( """--max_seq_length""" , default=128 , type=__SCREAMING_SNAKE_CASE , help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ) , ) parser.add_argument( """--task""" , default="""""" , type=__SCREAMING_SNAKE_CASE , required=__SCREAMING_SNAKE_CASE , help="""The GLUE task to run""" , ) parser.add_argument( """--gpus""" , default=0 , type=__SCREAMING_SNAKE_CASE , help="""The number of GPUs allocated for this, it is by default 0 meaning none""" , ) parser.add_argument( """--overwrite_cache""" , action="""store_true""" , help="""Overwrite the cached training and evaluation sets""" ) return parser def lowercase__ ( ) -> str: """simple docstring""" __UpperCAmelCase =argparse.ArgumentParser() add_generic_args(A_ , os.getcwd() ) __UpperCAmelCase =GLUETransformer.add_model_specific_args(A_ , os.getcwd() ) __UpperCAmelCase =parser.parse_args() # If output_dir not provided, a folder will be generated in pwd if args.output_dir is None: __UpperCAmelCase =os.path.join( """./results""" , F'''{args.task}_{time.strftime("%Y%m%d_%H%M%S" )}''' , ) os.makedirs(args.output_dir ) __UpperCAmelCase =GLUETransformer(A_ ) __UpperCAmelCase =generic_train(A_ , A_ ) # Optionally, predict on dev set and write to output_dir if args.do_predict: __UpperCAmelCase =sorted(glob.glob(os.path.join(args.output_dir , """checkpoint-epoch=*.ckpt""" ) , recursive=A_ ) ) __UpperCAmelCase =model.load_from_checkpoint(checkpoints[-1] ) return trainer.test(A_ ) if __name__ == "__main__": main()
68
0
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging lowerCAmelCase_ = logging.get_logger(__name__) lowerCAmelCase_ = {'vocab_file': 'spiece.model'} lowerCAmelCase_ = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class __A ( A_ ): '''simple docstring''' def __init__( self : List[str] ,_snake_case : Optional[Any] ,_snake_case : Dict=False ,_snake_case : List[str]=True ,_snake_case : Union[str, Any]=False ,_snake_case : Dict="<s>" ,_snake_case : Optional[Any]="</s>" ,_snake_case : str="<unk>" ,_snake_case : str="<sep>" ,_snake_case : Any="<pad>" ,_snake_case : Tuple="<cls>" ,_snake_case : Optional[int]="<mask>" ,_snake_case : Tuple=["<eop>", "<eod>"] ,_snake_case : Optional[Dict[str, Any]] = None ,**_snake_case : Optional[int] ,) -> None: """simple docstring""" lowercase__ : Tuple = AddedToken(__SCREAMING_SNAKE_CASE ,lstrip=__SCREAMING_SNAKE_CASE ,rstrip=__SCREAMING_SNAKE_CASE ) if isinstance(__SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE ) else mask_token lowercase__ : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE ,remove_space=__SCREAMING_SNAKE_CASE ,keep_accents=__SCREAMING_SNAKE_CASE ,bos_token=__SCREAMING_SNAKE_CASE ,eos_token=__SCREAMING_SNAKE_CASE ,unk_token=__SCREAMING_SNAKE_CASE ,sep_token=__SCREAMING_SNAKE_CASE ,pad_token=__SCREAMING_SNAKE_CASE ,cls_token=__SCREAMING_SNAKE_CASE ,mask_token=__SCREAMING_SNAKE_CASE ,additional_special_tokens=__SCREAMING_SNAKE_CASE ,sp_model_kwargs=self.sp_model_kwargs ,**__SCREAMING_SNAKE_CASE ,) lowercase__ : List[Any] = 3 lowercase__ : Optional[Any] = do_lower_case lowercase__ : Optional[Any] = remove_space lowercase__ : str = keep_accents lowercase__ : Union[str, Any] = vocab_file lowercase__ : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( '''You need to install jieba to use CpmTokenizer or CpmTokenizerFast. ''' '''See https://pypi.org/project/jieba/ for installation.''' ) lowercase__ : List[Any] = jieba lowercase__ : Optional[Any] = str.maketrans(''' \n''' ,'''\u2582\u2583''' ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def UpperCAmelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" return len(self.sp_model ) def UpperCAmelCase ( self : Optional[int] ) -> List[Any]: """simple docstring""" lowercase__ : Union[str, Any] = {self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ) -> List[str]: """simple docstring""" lowercase__ : Any = self.__dict__.copy() lowercase__ : Union[str, Any] = None return state def __setstate__( self : str ,_snake_case : Tuple ) -> str: """simple docstring""" lowercase__ : List[str] = d # for backward compatibility if not hasattr(self ,'''sp_model_kwargs''' ): lowercase__ : List[str] = {} lowercase__ : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCAmelCase ( self : Optional[int] ,_snake_case : Any ) -> Tuple: """simple docstring""" if self.remove_space: lowercase__ : Any = ''' '''.join(inputs.strip().split() ) else: lowercase__ : Dict = inputs lowercase__ : Optional[int] = outputs.replace('''``''' ,'''\"''' ).replace('''\'\'''' ,'''\"''' ) if not self.keep_accents: lowercase__ : Any = unicodedata.normalize('''NFKD''' ,__SCREAMING_SNAKE_CASE ) lowercase__ : int = ''''''.join([c for c in outputs if not unicodedata.combining(__SCREAMING_SNAKE_CASE )] ) if self.do_lower_case: lowercase__ : Union[str, Any] = outputs.lower() return outputs def UpperCAmelCase ( self : List[str] ,_snake_case : str ) -> List[str]: """simple docstring""" lowercase__ : str = self.preprocess_text(__SCREAMING_SNAKE_CASE ) lowercase__ : int = self.sp_model.encode(__SCREAMING_SNAKE_CASE ,out_type=__SCREAMING_SNAKE_CASE ) lowercase__ : str = [] for piece in pieces: if len(__SCREAMING_SNAKE_CASE ) > 1 and piece[-1] == str(''',''' ) and piece[-2].isdigit(): lowercase__ : Union[str, Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(__SCREAMING_SNAKE_CASE ,'''''' ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: lowercase__ : List[str] = cur_pieces[1:] else: lowercase__ : Optional[int] = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__SCREAMING_SNAKE_CASE ) else: new_pieces.append(__SCREAMING_SNAKE_CASE ) return new_pieces def UpperCAmelCase ( self : Optional[Any] ,_snake_case : Optional[Any] ) -> str: """simple docstring""" return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self : str ,_snake_case : int ) -> Any: """simple docstring""" return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) def UpperCAmelCase ( self : Optional[Any] ,_snake_case : Tuple ) -> List[Any]: """simple docstring""" lowercase__ : Any = ''''''.join(__SCREAMING_SNAKE_CASE ).replace(__SCREAMING_SNAKE_CASE ,''' ''' ).strip() return out_string def UpperCAmelCase ( self : Union[str, Any] ,_snake_case : List[int] ,_snake_case : Optional[List[int]] = None ) -> List[int]: """simple docstring""" lowercase__ : Tuple = [self.sep_token_id] lowercase__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def UpperCAmelCase ( self : str ,_snake_case : List[int] ,_snake_case : Optional[List[int]] = None ,_snake_case : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__SCREAMING_SNAKE_CASE ,token_ids_a=__SCREAMING_SNAKE_CASE ,already_has_special_tokens=__SCREAMING_SNAKE_CASE ) if token_ids_a is not None: return ([0] * len(__SCREAMING_SNAKE_CASE )) + [1] + ([0] * len(__SCREAMING_SNAKE_CASE )) + [1, 1] return ([0] * len(__SCREAMING_SNAKE_CASE )) + [1, 1] def UpperCAmelCase ( self : Optional[int] ,_snake_case : List[int] ,_snake_case : Optional[List[int]] = None ) -> List[int]: """simple docstring""" lowercase__ : List[str] = [self.sep_token_id] lowercase__ : Dict = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def UpperCAmelCase ( self : List[str] ,_snake_case : str ,_snake_case : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : List[str] = os.path.join( __SCREAMING_SNAKE_CASE ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file ,__SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE ,'''wb''' ) as fi: lowercase__ : Any = self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def UpperCAmelCase ( self : List[Any] ,*_snake_case : Any ,**_snake_case : Dict ) -> Dict: """simple docstring""" lowercase__ : List[Any] = super()._decode(*__SCREAMING_SNAKE_CASE ,**__SCREAMING_SNAKE_CASE ) lowercase__ : Dict = text.replace(''' ''' ,'''''' ).replace('''\u2582''' ,''' ''' ).replace('''\u2583''' ,'''\n''' ) return text
560
def lowercase__ ( A_: int , A_: int ) -> int: """simple docstring""" return 1 if input_a == input_a else 0 def lowercase__ ( ) -> None: """simple docstring""" assert xnor_gate(0 , 0 ) == 1 assert xnor_gate(0 , 1 ) == 0 assert xnor_gate(1 , 0 ) == 0 assert xnor_gate(1 , 1 ) == 1 if __name__ == "__main__": print(xnor_gate(0, 0)) print(xnor_gate(0, 1)) print(xnor_gate(1, 0)) print(xnor_gate(1, 1))
68
0
'''simple docstring''' from typing import List, Optional from tokenizers import ByteLevelBPETokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_blenderbot_small import BlenderbotSmallTokenizer a__ : Optional[Any] = logging.get_logger(__name__) a__ : List[str] = { 'vocab_file': 'vocab.json', 'merges_file': 'merges.txt', 'tokenizer_config_file': 'tokenizer_config.json', } a__ : List[str] = { 'vocab_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/vocab.json' }, 'merges_file': { 'facebook/blenderbot_small-90M': 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/merges.txt' }, 'tokenizer_config_file': { 'facebook/blenderbot_small-90M': ( 'https://huggingface.co/facebook/blenderbot_small-90M/resolve/main/tokenizer_config.json' ) }, } a__ : int = { 'facebook/blenderbot_small-90M': 512, } class lowerCAmelCase__ ( UpperCAmelCase_ ): '''simple docstring''' _lowerCamelCase =VOCAB_FILES_NAMES _lowerCamelCase =PRETRAINED_VOCAB_FILES_MAP _lowerCamelCase =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _lowerCamelCase =BlenderbotSmallTokenizer def __init__( self : Union[str, Any] , a__ : Any=None , a__ : Any=None , a__ : Optional[Any]="<|endoftext|>" , a__ : str="<|endoftext|>" , a__ : Any="<|endoftext|>" , a__ : Tuple=False , a__ : Optional[int]=True , **a__ : Tuple , ): super().__init__( ByteLevelBPETokenizer( vocab=__SCREAMING_SNAKE_CASE , merges=__SCREAMING_SNAKE_CASE , add_prefix_space=__SCREAMING_SNAKE_CASE , trim_offsets=__SCREAMING_SNAKE_CASE , ) , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE , ) UpperCAmelCase = add_prefix_space def __snake_case ( self : Union[str, Any] , a__ : Dict , a__ : Optional[Any]=None ): UpperCAmelCase = [self.bos_token_id] + token_ids_a + [self.eos_token_id] if token_ids_a is None: return output return output + [self.eos_token_id] + token_ids_a + [self.eos_token_id] def __snake_case ( self : Dict , a__ : List[int] , a__ : Optional[List[int]] = None ): UpperCAmelCase = [self.sep_token_id] UpperCAmelCase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0]
51
from __future__ import annotations import bisect def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] < item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> int: """simple docstring""" if hi < 0: __UpperCAmelCase =len(A_ ) while lo < hi: __UpperCAmelCase =lo + (hi - lo) // 2 if sorted_collection[mid] <= item: __UpperCAmelCase =mid + 1 else: __UpperCAmelCase =mid return lo def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_left(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int , A_: int = 0 , A_: int = -1 ) -> None: """simple docstring""" sorted_collection.insert(bisect_right(A_ , A_ , A_ , A_ ) , A_ ) def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =0 __UpperCAmelCase =len(A_ ) - 1 while left <= right: __UpperCAmelCase =left + (right - left) // 2 __UpperCAmelCase =sorted_collection[midpoint] if current_item == item: return midpoint elif item < current_item: __UpperCAmelCase =midpoint - 1 else: __UpperCAmelCase =midpoint + 1 return None def lowercase__ ( A_: list[int] , A_: int ) -> int | None: """simple docstring""" __UpperCAmelCase =bisect.bisect_left(A_ , A_ ) if index != len(A_ ) and sorted_collection[index] == item: return index return None def lowercase__ ( A_: list[int] , A_: int , A_: int , A_: int ) -> int | None: """simple docstring""" if right < left: return None __UpperCAmelCase =left + (right - left) // 2 if sorted_collection[midpoint] == item: return midpoint elif sorted_collection[midpoint] > item: return binary_search_by_recursion(A_ , A_ , A_ , midpoint - 1 ) else: return binary_search_by_recursion(A_ , A_ , midpoint + 1 , A_ ) if __name__ == "__main__": __A = input("Enter numbers separated by comma:\n").strip() __A = sorted(int(item) for item in user_input.split(",")) __A = int(input("Enter a single number to be found in the list:\n")) __A = binary_search(collection, target) if result is None: print(F"""{target} was not found in {collection}.""") else: print(F"""{target} was found at position {result} in {collection}.""")
68
0
"""simple docstring""" from ....configuration_utils import PretrainedConfig from ....utils import logging __magic_name__ = logging.get_logger(__name__) # TODO: upload to AWS __magic_name__ = { '''yjernite/retribert-base-uncased''': ( '''https://huggingface.co/yjernite/retribert-base-uncased/resolve/main/config.json''' ), } class _lowerCAmelCase ( lowerCamelCase ): lowercase_ : Any = 'retribert' def __init__( self , a_=30522 , a_=768 , a_=8 , a_=12 , a_=3072 , a_="gelu" , a_=0.1 , a_=0.1 , a_=512 , a_=2 , a_=0.02 , a_=1e-12 , a_=True , a_=128 , a_=0 , **a_ , ) -> Optional[Any]: super().__init__(pad_token_id=__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) _UpperCAmelCase = vocab_size _UpperCAmelCase = hidden_size _UpperCAmelCase = num_hidden_layers _UpperCAmelCase = num_attention_heads _UpperCAmelCase = hidden_act _UpperCAmelCase = intermediate_size _UpperCAmelCase = hidden_dropout_prob _UpperCAmelCase = attention_probs_dropout_prob _UpperCAmelCase = max_position_embeddings _UpperCAmelCase = type_vocab_size _UpperCAmelCase = initializer_range _UpperCAmelCase = layer_norm_eps _UpperCAmelCase = share_encoders _UpperCAmelCase = projection_dim
657
from typing import List from .keymap import KEYMAP, get_character def lowercase__ ( A_: str ) -> str: """simple docstring""" def decorator(A_: int ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += [key] setattr(A_ , """handle_key""" , A_ ) return func return decorator def lowercase__ ( *A_: List[str] ) -> Optional[int]: """simple docstring""" def decorator(A_: Tuple ): __UpperCAmelCase =getattr(A_ , """handle_key""" , [] ) handle += keys setattr(A_ , """handle_key""" , A_ ) return func return decorator class _A ( UpperCamelCase ): """simple docstring""" def __new__( cls : str , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : List[str] ) -> int: __UpperCAmelCase =super().__new__(cls , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) if not hasattr(__SCREAMING_SNAKE_CASE , """key_handler""" ): setattr(__SCREAMING_SNAKE_CASE , """key_handler""" , {} ) setattr(__SCREAMING_SNAKE_CASE , """handle_input""" , KeyHandler.handle_input ) for value in attrs.values(): __UpperCAmelCase =getattr(__SCREAMING_SNAKE_CASE , """handle_key""" , [] ) for key in handled_keys: __UpperCAmelCase =value return new_cls @staticmethod def _a ( cls : Dict ) -> List[Any]: __UpperCAmelCase =get_character() if char != KEYMAP["undefined"]: __UpperCAmelCase =ord(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =cls.key_handler.get(__SCREAMING_SNAKE_CASE ) if handler: __UpperCAmelCase =char return handler(cls ) else: return None def lowercase__ ( cls: str ) -> int: """simple docstring""" return KeyHandler(cls.__name__ , cls.__bases__ , cls.__dict__.copy() )
68
0
'''simple docstring''' def _a( UpperCamelCase__ : int = 1, UpperCamelCase__ : int = 1_0_0_0 ): '''simple docstring''' SCREAMING_SNAKE_CASE__ : List[Any] =1 SCREAMING_SNAKE_CASE__ : List[str] =0 for divide_by_number in range(A_, digit + 1 ): SCREAMING_SNAKE_CASE__ : int =[] SCREAMING_SNAKE_CASE__ : Optional[int] =numerator for _ in range(1, digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(A_ ): SCREAMING_SNAKE_CASE__ : Optional[Any] =len(A_ ) SCREAMING_SNAKE_CASE__ : List[str] =divide_by_number else: has_been_divided.append(A_ ) SCREAMING_SNAKE_CASE__ : Dict =now_divide * 1_0 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
296
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
68
0
"""simple docstring""" import math from collections import defaultdict from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin, SchedulerOutput def __SCREAMING_SNAKE_CASE ( lowerCamelCase_: str , lowerCamelCase_: int=0.9_9_9 , lowerCamelCase_: List[str]="cosine" , ): """simple docstring""" if alpha_transform_type == "cosine": def alpha_bar_fn(lowerCamelCase_: Tuple ): return math.cos((t + 0.0_0_8) / 1.0_0_8 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(lowerCamelCase_: int ): return math.exp(t * -1_2.0 ) else: raise ValueError(f'''Unsupported alpha_tranform_type: {alpha_transform_type}''' ) snake_case : Union[str, Any] = [] for i in range(A_ ): snake_case : Any = i / num_diffusion_timesteps snake_case : Dict = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(A_ ) / alpha_bar_fn(A_ ) , A_ ) ) return torch.tensor(A_ , dtype=torch.floataa ) class _a ( SCREAMING_SNAKE_CASE__ , SCREAMING_SNAKE_CASE__): __magic_name__ = [e.name for e in KarrasDiffusionSchedulers] __magic_name__ = 2 @register_to_config def __init__( self : List[str] , _lowercase : int = 1000 , _lowercase : float = 0.00085 , _lowercase : float = 0.012 , _lowercase : str = "linear" , _lowercase : Optional[Union[np.ndarray, List[float]]] = None , _lowercase : str = "epsilon" , _lowercase : str = "linspace" , _lowercase : int = 0 , ) -> Optional[Any]: if trained_betas is not None: snake_case : str = torch.tensor(__SCREAMING_SNAKE_CASE , dtype=torch.floataa ) elif beta_schedule == "linear": snake_case : Optional[Any] = torch.linspace(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , dtype=torch.floataa ) elif beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. snake_case : Any = ( torch.linspace(beta_start**0.5 , beta_end**0.5 , __SCREAMING_SNAKE_CASE , dtype=torch.floataa ) ** 2 ) elif beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule snake_case : str = betas_for_alpha_bar(__SCREAMING_SNAKE_CASE ) else: raise NotImplementedError(F'''{beta_schedule} does is not implemented for {self.__class__}''' ) snake_case : List[str] = 1.0 - self.betas snake_case : Optional[int] = torch.cumprod(self.alphas , dim=0 ) # set all values self.set_timesteps(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def __lowercase ( self : Any , _lowercase : Optional[int] , _lowercase : str=None ) -> Optional[int]: if schedule_timesteps is None: snake_case : List[str] = self.timesteps snake_case : str = (schedule_timesteps == timestep).nonzero() # The sigma index that is taken for the **very** first `step` # is always the second index (or the last index if there is only 1) # This way we can ensure we don't accidentally skip a sigma in # case we start in the middle of the denoising schedule (e.g. for image-to-image) if len(self._index_counter ) == 0: snake_case : List[Any] = 1 if len(__SCREAMING_SNAKE_CASE ) > 1 else 0 else: snake_case : Optional[Any] = timestep.cpu().item() if torch.is_tensor(__SCREAMING_SNAKE_CASE ) else timestep snake_case : Tuple = self._index_counter[timestep_int] return indices[pos].item() @property def __lowercase ( self : List[str] ) -> Any: # standard deviation of the initial noise distribution if self.config.timestep_spacing in ["linspace", "trailing"]: return self.sigmas.max() return (self.sigmas.max() ** 2 + 1) ** 0.5 def __lowercase ( self : int , _lowercase : torch.FloatTensor , _lowercase : Union[float, torch.FloatTensor] , ) -> torch.FloatTensor: snake_case : List[str] = self.index_for_timestep(__SCREAMING_SNAKE_CASE ) if self.state_in_first_order: snake_case : Dict = self.sigmas[step_index] else: snake_case : Union[str, Any] = self.sigmas_interpol[step_index] snake_case : List[Any] = sample / ((sigma**2 + 1) ** 0.5) return sample def __lowercase ( self : Union[str, Any] , _lowercase : int , _lowercase : Union[str, torch.device] = None , _lowercase : Optional[int] = None , ) -> Optional[Any]: snake_case : List[Any] = num_inference_steps snake_case : List[str] = num_train_timesteps or self.config.num_train_timesteps # "linspace", "leading", "trailing" corresponds to annotation of Table 2. of https://arxiv.org/abs/2305.08891 if self.config.timestep_spacing == "linspace": snake_case : Optional[Any] = np.linspace(0 , num_train_timesteps - 1 , __SCREAMING_SNAKE_CASE , dtype=__SCREAMING_SNAKE_CASE )[::-1].copy() elif self.config.timestep_spacing == "leading": snake_case : Union[str, Any] = num_train_timesteps // self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case : List[str] = (np.arange(0 , __SCREAMING_SNAKE_CASE ) * step_ratio).round()[::-1].copy().astype(__SCREAMING_SNAKE_CASE ) timesteps += self.config.steps_offset elif self.config.timestep_spacing == "trailing": snake_case : Optional[Any] = num_train_timesteps / self.num_inference_steps # creates integer timesteps by multiplying by ratio # casting to int to avoid issues when num_inference_step is power of 3 snake_case : Tuple = (np.arange(__SCREAMING_SNAKE_CASE , 0 , -step_ratio )).round().copy().astype(__SCREAMING_SNAKE_CASE ) timesteps -= 1 else: raise ValueError( F'''{self.config.timestep_spacing} is not supported. Please make sure to choose one of \'linspace\', \'leading\' or \'trailing\'.''' ) snake_case : List[Any] = np.array(((1 - self.alphas_cumprod) / self.alphas_cumprod) ** 0.5 ) snake_case : List[str] = torch.from_numpy(np.log(__SCREAMING_SNAKE_CASE ) ).to(__SCREAMING_SNAKE_CASE ) snake_case : Union[str, Any] = np.interp(__SCREAMING_SNAKE_CASE , np.arange(0 , len(__SCREAMING_SNAKE_CASE ) ) , __SCREAMING_SNAKE_CASE ) snake_case : Optional[int] = np.concatenate([sigmas, [0.0]] ).astype(np.floataa ) snake_case : Tuple = torch.from_numpy(__SCREAMING_SNAKE_CASE ).to(device=__SCREAMING_SNAKE_CASE ) # interpolate sigmas snake_case : Union[str, Any] = sigmas.log().lerp(sigmas.roll(1 ).log() , 0.5 ).exp() snake_case : Optional[Any] = torch.cat([sigmas[:1], sigmas[1:].repeat_interleave(2 ), sigmas[-1:]] ) snake_case : Dict = torch.cat( [sigmas_interpol[:1], sigmas_interpol[1:].repeat_interleave(2 ), sigmas_interpol[-1:]] ) if str(__SCREAMING_SNAKE_CASE ).startswith("mps" ): # mps does not support float64 snake_case : Dict = torch.from_numpy(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE , dtype=torch.floataa ) else: snake_case : Any = torch.from_numpy(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE ) # interpolate timesteps snake_case : Union[str, Any] = self.sigma_to_t(__SCREAMING_SNAKE_CASE ).to(__SCREAMING_SNAKE_CASE , dtype=timesteps.dtype ) snake_case : int = torch.stack((timesteps_interpol[1:-1, None], timesteps[1:, None]) , dim=-1 ).flatten() snake_case : List[Any] = torch.cat([timesteps[:1], interleaved_timesteps] ) snake_case : List[str] = None # for exp beta schedules, such as the one for `pipeline_shap_e.py` # we need an index counter snake_case : List[Any] = defaultdict(__SCREAMING_SNAKE_CASE ) def __lowercase ( self : Optional[Any] , _lowercase : Optional[int] ) -> List[str]: # get log sigma snake_case : List[str] = sigma.log() # get distribution snake_case : List[Any] = log_sigma - self.log_sigmas[:, None] # get sigmas range snake_case : Any = dists.ge(0 ).cumsum(dim=0 ).argmax(dim=0 ).clamp(max=self.log_sigmas.shape[0] - 2 ) snake_case : Tuple = low_idx + 1 snake_case : Optional[Any] = self.log_sigmas[low_idx] snake_case : Optional[Any] = self.log_sigmas[high_idx] # interpolate sigmas snake_case : str = (low - log_sigma) / (low - high) snake_case : Tuple = w.clamp(0 , 1 ) # transform interpolation to time range snake_case : Union[str, Any] = (1 - w) * low_idx + w * high_idx snake_case : Optional[Any] = t.view(sigma.shape ) return t @property def __lowercase ( self : str ) -> Optional[Any]: return self.sample is None def __lowercase ( self : Dict , _lowercase : Union[torch.FloatTensor, np.ndarray] , _lowercase : Union[float, torch.FloatTensor] , _lowercase : Union[torch.FloatTensor, np.ndarray] , _lowercase : bool = True , ) -> Union[SchedulerOutput, Tuple]: snake_case : Tuple = self.index_for_timestep(__SCREAMING_SNAKE_CASE ) # advance index counter by 1 snake_case : Dict = timestep.cpu().item() if torch.is_tensor(__SCREAMING_SNAKE_CASE ) else timestep self._index_counter[timestep_int] += 1 if self.state_in_first_order: snake_case : Optional[Any] = self.sigmas[step_index] snake_case : Optional[Any] = self.sigmas_interpol[step_index + 1] snake_case : Any = self.sigmas[step_index + 1] else: # 2nd order / KDPM2's method snake_case : str = self.sigmas[step_index - 1] snake_case : Tuple = self.sigmas_interpol[step_index] snake_case : Any = self.sigmas[step_index] # currently only gamma=0 is supported. This usually works best anyways. # We can support gamma in the future but then need to scale the timestep before # passing it to the model which requires a change in API snake_case : List[Any] = 0 snake_case : List[str] = sigma * (gamma + 1) # Note: sigma_hat == sigma for now # 1. compute predicted original sample (x_0) from sigma-scaled predicted noise if self.config.prediction_type == "epsilon": snake_case : Any = sigma_hat if self.state_in_first_order else sigma_interpol snake_case : Dict = sample - sigma_input * model_output elif self.config.prediction_type == "v_prediction": snake_case : Optional[Any] = sigma_hat if self.state_in_first_order else sigma_interpol snake_case : Union[str, Any] = model_output * (-sigma_input / (sigma_input**2 + 1) ** 0.5) + ( sample / (sigma_input**2 + 1) ) elif self.config.prediction_type == "sample": raise NotImplementedError("prediction_type not implemented yet: sample" ) else: raise ValueError( F'''prediction_type given as {self.config.prediction_type} must be one of `epsilon`, or `v_prediction`''' ) if self.state_in_first_order: # 2. Convert to an ODE derivative for 1st order snake_case : Optional[Any] = (sample - pred_original_sample) / sigma_hat # 3. delta timestep snake_case : Union[str, Any] = sigma_interpol - sigma_hat # store for 2nd order step snake_case : str = sample else: # DPM-Solver-2 # 2. Convert to an ODE derivative for 2nd order snake_case : Union[str, Any] = (sample - pred_original_sample) / sigma_interpol # 3. delta timestep snake_case : Optional[int] = sigma_next - sigma_hat snake_case : Optional[int] = self.sample snake_case : List[str] = None snake_case : List[str] = sample + derivative * dt if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__SCREAMING_SNAKE_CASE ) def __lowercase ( self : Any , _lowercase : torch.FloatTensor , _lowercase : torch.FloatTensor , _lowercase : torch.FloatTensor , ) -> torch.FloatTensor: # Make sure sigmas and timesteps have the same device and dtype as original_samples snake_case : Any = self.sigmas.to(device=original_samples.device , dtype=original_samples.dtype ) if original_samples.device.type == "mps" and torch.is_floating_point(__SCREAMING_SNAKE_CASE ): # mps does not support float64 snake_case : List[Any] = self.timesteps.to(original_samples.device , dtype=torch.floataa ) snake_case : Any = timesteps.to(original_samples.device , dtype=torch.floataa ) else: snake_case : Union[str, Any] = self.timesteps.to(original_samples.device ) snake_case : Optional[int] = timesteps.to(original_samples.device ) snake_case : List[str] = [self.index_for_timestep(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) for t in timesteps] snake_case : str = sigmas[step_indices].flatten() while len(sigma.shape ) < len(original_samples.shape ): snake_case : Any = sigma.unsqueeze(-1 ) snake_case : Union[str, Any] = original_samples + noise * sigma return noisy_samples def __len__( self : int ) -> List[str]: return self.config.num_train_timesteps
449
import unittest import numpy as np from transformers import RobertaConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_flax_available(): from transformers.models.roberta.modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, ) class _A ( unittest.TestCase ): """simple docstring""" def __init__( self : Any , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=7 , __SCREAMING_SNAKE_CASE : Tuple=True , __SCREAMING_SNAKE_CASE : List[str]=True , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : int=True , __SCREAMING_SNAKE_CASE : Optional[Any]=99 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : int=5 , __SCREAMING_SNAKE_CASE : Dict=4 , __SCREAMING_SNAKE_CASE : str=37 , __SCREAMING_SNAKE_CASE : Union[str, Any]="gelu" , __SCREAMING_SNAKE_CASE : int=0.1 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : str=512 , __SCREAMING_SNAKE_CASE : Dict=16 , __SCREAMING_SNAKE_CASE : Any=2 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.02 , __SCREAMING_SNAKE_CASE : List[str]=4 , ) -> Optional[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =seq_length __UpperCAmelCase =is_training __UpperCAmelCase =use_attention_mask __UpperCAmelCase =use_token_type_ids __UpperCAmelCase =use_labels __UpperCAmelCase =vocab_size __UpperCAmelCase =hidden_size __UpperCAmelCase =num_hidden_layers __UpperCAmelCase =num_attention_heads __UpperCAmelCase =intermediate_size __UpperCAmelCase =hidden_act __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =max_position_embeddings __UpperCAmelCase =type_vocab_size __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =initializer_range __UpperCAmelCase =num_choices def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCAmelCase =None if self.use_attention_mask: __UpperCAmelCase =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCAmelCase =None if self.use_token_type_ids: __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) __UpperCAmelCase =RobertaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__SCREAMING_SNAKE_CASE , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": attention_mask} return config, inputs_dict def _a ( self : List[str] ) -> Dict: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase =True __UpperCAmelCase =floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) __UpperCAmelCase =ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, encoder_hidden_states, encoder_attention_mask, ) @require_flax class _A ( UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Union[str, Any] = True lowerCamelCase : Union[str, Any] = ( ( FlaxRobertaModel, FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, ) if is_flax_available() else () ) def _a ( self : List[Any] ) -> List[str]: __UpperCAmelCase =FlaxRobertaModelTester(self ) @slow def _a ( self : Optional[Any] ) -> List[Any]: for model_class_name in self.all_model_classes: __UpperCAmelCase =model_class_name.from_pretrained("""roberta-base""" , from_pt=__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =model(np.ones((1, 1) ) ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' def __UpperCamelCase ( a : int , a : Tuple ) ->Union[str, Any]: snake_case = [1] for i in range(2 , A_ ): factorials.append(factorials[-1] * i ) assert 0 <= k < factorials[-1] * n, "k out of bounds" snake_case = [] snake_case = list(range(A_ ) ) # Find permutation while factorials: snake_case = factorials.pop() snake_case , snake_case = divmod(A_ , A_ ) permutation.append(elements[number] ) elements.remove(elements[number] ) permutation.append(elements[0] ) return permutation if __name__ == "__main__": import doctest doctest.testmod()
342
from __future__ import annotations def lowercase__ ( A_: list[list[int]] ) -> int: """simple docstring""" for i in range(1 , len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1 , len(A_ ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1 , len(A_ ) ): for j in range(1 , len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j] , matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
68
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available __lowerCamelCase : int = { "configuration_tapas": ["TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP", "TapasConfig"], "tokenization_tapas": ["TapasTokenizer"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : List[Any] = [ "TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TapasForMaskedLM", "TapasForQuestionAnswering", "TapasForSequenceClassification", "TapasModel", "TapasPreTrainedModel", "load_tf_weights_in_tapas", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCamelCase : int = [ "TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST", "TFTapasForMaskedLM", "TFTapasForQuestionAnswering", "TFTapasForSequenceClassification", "TFTapasModel", "TFTapasPreTrainedModel", ] if TYPE_CHECKING: from .configuration_tapas import TAPAS_PRETRAINED_CONFIG_ARCHIVE_MAP, TapasConfig from .tokenization_tapas import TapasTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tapas import ( TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TapasForMaskedLM, TapasForQuestionAnswering, TapasForSequenceClassification, TapasModel, TapasPreTrainedModel, load_tf_weights_in_tapas, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_tapas import ( TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST, TFTapasForMaskedLM, TFTapasForQuestionAnswering, TFTapasForSequenceClassification, TFTapasModel, TFTapasPreTrainedModel, ) else: import sys __lowerCamelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
323
import numpy as np from cva import COLOR_BGR2GRAY, CV_8UC3, cvtColor, filteraD, imread, imshow, waitKey def lowercase__ ( A_: int , A_: int , A_: int , A_: int , A_: int , A_: int ) -> np.ndarray: """simple docstring""" if (ksize % 2) == 0: __UpperCAmelCase =ksize + 1 __UpperCAmelCase =np.zeros((ksize, ksize) , dtype=np.floataa ) # each value for y in range(A_ ): for x in range(A_ ): # distance from center __UpperCAmelCase =x - ksize // 2 __UpperCAmelCase =y - ksize // 2 # degree to radiant __UpperCAmelCase =theta / 180 * np.pi __UpperCAmelCase =np.cos(_theta ) __UpperCAmelCase =np.sin(_theta ) # get kernel x __UpperCAmelCase =cos_theta * px + sin_theta * py # get kernel y __UpperCAmelCase =-sin_theta * px + cos_theta * py # fill kernel __UpperCAmelCase =np.exp( -(_x**2 + gamma**2 * _y**2) / (2 * sigma**2) ) * np.cos(2 * np.pi * _x / lambd + psi ) return gabor if __name__ == "__main__": import doctest doctest.testmod() # read original image __A = imread("../image_data/lena.jpg") # turn image in gray scale value __A = cvtColor(img, COLOR_BGR2GRAY) # Apply multiple Kernel to detect edges __A = np.zeros(gray.shape[:2]) for theta in [0, 30, 60, 90, 1_20, 1_50]: __A = gabor_filter_kernel(10, 8, theta, 10, 0, 0) out += filteraD(gray, CV_8UC3, kernel_aa) __A = out / out.max() * 2_55 __A = out.astype(np.uinta) imshow("Original", gray) imshow("Gabor filter with 20x20 mask and 6 directions", out) waitKey(0)
68
0
'''simple docstring''' import os import unittest from huggingface_hub.utils import are_progress_bars_disabled import transformers.models.bart.tokenization_bart from transformers import logging from transformers.testing_utils import CaptureLogger, mockenv, mockenv_context from transformers.utils.logging import disable_progress_bar, enable_progress_bar class A ( unittest.TestCase ): def snake_case__ ( self : Any ) -> int: __UpperCAmelCase = logging.get_logger() # the current default level is logging.WARNING __UpperCAmelCase = logging.get_verbosity() logging.set_verbosity_error() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_warning() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_info() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) logging.set_verbosity_debug() self.assertEqual(logger.getEffectiveLevel() , logging.get_verbosity() ) # restore to the original level logging.set_verbosity(__SCREAMING_SNAKE_CASE ) def snake_case__ ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase = logging.get_verbosity() __UpperCAmelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __UpperCAmelCase = '''Testing 1, 2, 3''' # should be able to log warnings (if default settings weren't overridden by `pytest --log-level-all`) if level_origin <= logging.WARNING: with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: logger.warning(__SCREAMING_SNAKE_CASE ) self.assertEqual(cl.out , msg + '''\n''' ) # this is setting the level for all of `transformers.*` loggers logging.set_verbosity_error() # should not be able to log warnings with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: logger.warning(__SCREAMING_SNAKE_CASE ) self.assertEqual(cl.out , '''''' ) # should be able to log warnings again logging.set_verbosity_warning() with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: logger.warning(__SCREAMING_SNAKE_CASE ) self.assertEqual(cl.out , msg + '''\n''' ) # restore to the original level logging.set_verbosity(__SCREAMING_SNAKE_CASE ) @mockenv(TRANSFORMERS_VERBOSITY='''error''' ) def snake_case__ ( self : Dict ) -> List[str]: # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() # this action activates the env var __UpperCAmelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __UpperCAmelCase = os.getenv('''TRANSFORMERS_VERBOSITY''' , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase = logging.log_levels[env_level_str] __UpperCAmelCase = logging.get_verbosity() self.assertEqual( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , f"""TRANSFORMERS_VERBOSITY={env_level_str}/{env_level}, but internal verbosity is {current_level}""" , ) # restore to the original level __UpperCAmelCase = '''''' transformers.utils.logging._reset_library_root_logger() @mockenv(TRANSFORMERS_VERBOSITY='''super-error''' ) def snake_case__ ( self : Union[str, Any] ) -> str: # reset for the env var to take effect, next time some logger call is made transformers.utils.logging._reset_library_root_logger() __UpperCAmelCase = logging.logging.getLogger() with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: # this action activates the env var logging.get_logger('''transformers.models.bart.tokenization_bart''' ) self.assertIn('''Unknown option TRANSFORMERS_VERBOSITY=super-error''' , cl.out ) # no need to restore as nothing was changed def snake_case__ ( self : Dict ) -> Optional[int]: # testing `logger.warning_advice()` transformers.utils.logging._reset_library_root_logger() __UpperCAmelCase = logging.get_logger('''transformers.models.bart.tokenization_bart''' ) __UpperCAmelCase = '''Testing 1, 2, 3''' with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''1''' ): # nothing should be logged as env var disables this method with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: logger.warning_advice(__SCREAMING_SNAKE_CASE ) self.assertEqual(cl.out , '''''' ) with mockenv_context(TRANSFORMERS_NO_ADVISORY_WARNINGS='''''' ): # should log normally as TRANSFORMERS_NO_ADVISORY_WARNINGS is unset with CaptureLogger(__SCREAMING_SNAKE_CASE ) as cl: logger.warning_advice(__SCREAMING_SNAKE_CASE ) self.assertEqual(cl.out , msg + '''\n''' ) def lowerCAmelCase ( ): """simple docstring""" disable_progress_bar() assert are_progress_bars_disabled() enable_progress_bar() assert not are_progress_bars_disabled()
262
import collections import inspect import unittest from transformers import SwinvaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel from transformers.models.swinva.modeling_swinva import SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _A : """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : List[Any]=13 , __SCREAMING_SNAKE_CASE : Dict=32 , __SCREAMING_SNAKE_CASE : Optional[int]=2 , __SCREAMING_SNAKE_CASE : Optional[int]=3 , __SCREAMING_SNAKE_CASE : List[str]=16 , __SCREAMING_SNAKE_CASE : Union[str, Any]=[1, 2, 1] , __SCREAMING_SNAKE_CASE : List[Any]=[2, 2, 4] , __SCREAMING_SNAKE_CASE : str=2 , __SCREAMING_SNAKE_CASE : Any=2.0 , __SCREAMING_SNAKE_CASE : Optional[int]=True , __SCREAMING_SNAKE_CASE : int=0.0 , __SCREAMING_SNAKE_CASE : Dict=0.0 , __SCREAMING_SNAKE_CASE : Union[str, Any]=0.1 , __SCREAMING_SNAKE_CASE : Any="gelu" , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : List[Any]=0.02 , __SCREAMING_SNAKE_CASE : Tuple=1e-5 , __SCREAMING_SNAKE_CASE : Optional[Any]=True , __SCREAMING_SNAKE_CASE : Optional[int]=None , __SCREAMING_SNAKE_CASE : str=True , __SCREAMING_SNAKE_CASE : Any=10 , __SCREAMING_SNAKE_CASE : Dict=8 , ) -> List[Any]: __UpperCAmelCase =parent __UpperCAmelCase =batch_size __UpperCAmelCase =image_size __UpperCAmelCase =patch_size __UpperCAmelCase =num_channels __UpperCAmelCase =embed_dim __UpperCAmelCase =depths __UpperCAmelCase =num_heads __UpperCAmelCase =window_size __UpperCAmelCase =mlp_ratio __UpperCAmelCase =qkv_bias __UpperCAmelCase =hidden_dropout_prob __UpperCAmelCase =attention_probs_dropout_prob __UpperCAmelCase =drop_path_rate __UpperCAmelCase =hidden_act __UpperCAmelCase =use_absolute_embeddings __UpperCAmelCase =patch_norm __UpperCAmelCase =layer_norm_eps __UpperCAmelCase =initializer_range __UpperCAmelCase =is_training __UpperCAmelCase =scope __UpperCAmelCase =use_labels __UpperCAmelCase =type_sequence_label_size __UpperCAmelCase =encoder_stride def _a ( self : Tuple ) -> Optional[int]: __UpperCAmelCase =floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __UpperCAmelCase =None if self.use_labels: __UpperCAmelCase =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCAmelCase =self.get_config() return config, pixel_values, labels def _a ( self : List[Any] ) -> Optional[Any]: return SwinvaConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , embed_dim=self.embed_dim , depths=self.depths , num_heads=self.num_heads , window_size=self.window_size , mlp_ratio=self.mlp_ratio , qkv_bias=self.qkv_bias , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , drop_path_rate=self.drop_path_rate , hidden_act=self.hidden_act , use_absolute_embeddings=self.use_absolute_embeddings , path_norm=self.patch_norm , layer_norm_eps=self.layer_norm_eps , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Any , __SCREAMING_SNAKE_CASE : Tuple , __SCREAMING_SNAKE_CASE : Dict ) -> Optional[int]: __UpperCAmelCase =SwinvaModel(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =((config.image_size // config.patch_size) ** 2) // (4 ** (len(config.depths ) - 1)) __UpperCAmelCase =int(config.embed_dim * 2 ** (len(config.depths ) - 1) ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, expected_seq_len, expected_dim) ) def _a ( self : int , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : List[Any] , __SCREAMING_SNAKE_CASE : int ) -> Tuple: __UpperCAmelCase =SwinvaForMaskedImageModeling(config=__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_channels, self.image_size, self.image_size) ) # test greyscale images __UpperCAmelCase =1 __UpperCAmelCase =SwinvaForMaskedImageModeling(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 1, self.image_size, self.image_size) ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[int] , __SCREAMING_SNAKE_CASE : Union[str, Any] ) -> Tuple: __UpperCAmelCase =self.type_sequence_label_size __UpperCAmelCase =SwinvaForImageClassification(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() __UpperCAmelCase =model(__SCREAMING_SNAKE_CASE , labels=__SCREAMING_SNAKE_CASE ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def _a ( self : List[str] ) -> Tuple: __UpperCAmelCase =self.prepare_config_and_inputs() __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =config_and_inputs __UpperCAmelCase ={"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class _A ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): """simple docstring""" lowerCamelCase : Optional[int] = ( (SwinvaModel, SwinvaForImageClassification, SwinvaForMaskedImageModeling) if is_torch_available() else () ) lowerCamelCase : Tuple = ( {'feature-extraction': SwinvaModel, 'image-classification': SwinvaForImageClassification} if is_torch_available() else {} ) lowerCamelCase : Dict = False lowerCamelCase : Tuple = False lowerCamelCase : List[str] = False lowerCamelCase : Tuple = False def _a ( self : str ) -> str: __UpperCAmelCase =SwinvaModelTester(self ) __UpperCAmelCase =ConfigTester(self , config_class=__SCREAMING_SNAKE_CASE , embed_dim=37 ) def _a ( self : List[Any] ) -> Optional[int]: self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _a ( self : str ) -> str: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__SCREAMING_SNAKE_CASE ) @unittest.skip(reason="""Got `CUDA error: misaligned address` with PyTorch 2.0.0.""" ) def _a ( self : Tuple ) -> Tuple: pass @unittest.skip(reason="""Swinv2 does not use inputs_embeds""" ) def _a ( self : Optional[Any] ) -> int: pass def _a ( self : Tuple ) -> int: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __UpperCAmelCase =model.get_output_embeddings() self.assertTrue(x is None or isinstance(__SCREAMING_SNAKE_CASE , nn.Linear ) ) def _a ( self : str ) -> List[str]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __UpperCAmelCase =[*signature.parameters.keys()] __UpperCAmelCase =["""pixel_values"""] self.assertListEqual(arg_names[:1] , __SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =True for model_class in self.all_model_classes: __UpperCAmelCase =True __UpperCAmelCase =False __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions __UpperCAmelCase =len(self.model_tester.depths ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # check that output_attentions also work using config del inputs_dict["output_attentions"] __UpperCAmelCase =True __UpperCAmelCase =config.window_size**2 __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) __UpperCAmelCase =len(__SCREAMING_SNAKE_CASE ) # Check attention is always last and order is fine __UpperCAmelCase =True __UpperCAmelCase =True __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) if hasattr(self.model_tester , """num_hidden_states_types""" ): __UpperCAmelCase =self.model_tester.num_hidden_states_types else: # also another +1 for reshaped_hidden_states __UpperCAmelCase =2 self.assertEqual(out_len + added_hidden_states , len(__SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.attentions self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) self.assertListEqual( list(self_attentions[0].shape[-3:] ) , [self.model_tester.num_heads[0], window_size_squared, window_size_squared] , ) def _a ( self : Optional[int] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : Optional[Any] , __SCREAMING_SNAKE_CASE : List[str] , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> int: __UpperCAmelCase =model_class(__SCREAMING_SNAKE_CASE ) model.to(__SCREAMING_SNAKE_CASE ) model.eval() with torch.no_grad(): __UpperCAmelCase =model(**self._prepare_for_class(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) ) __UpperCAmelCase =outputs.hidden_states __UpperCAmelCase =getattr( self.model_tester , """expected_num_hidden_layers""" , len(self.model_tester.depths ) + 1 ) self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) # Swinv2 has a different seq_length __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =(image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) __UpperCAmelCase =outputs.reshaped_hidden_states self.assertEqual(len(__SCREAMING_SNAKE_CASE ) , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase =reshaped_hidden_states[0].shape __UpperCAmelCase =( reshaped_hidden_states[0].view(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , height * width ).permute(0 , 2 , 1 ) ) self.assertListEqual( list(reshaped_hidden_states.shape[-2:] ) , [num_patches, self.model_tester.embed_dim] , ) def _a ( self : str ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =3 __UpperCAmelCase =( self.model_tester.image_size if isinstance(self.model_tester.image_size , collections.abc.Iterable ) else (self.model_tester.image_size, self.model_tester.image_size) ) __UpperCAmelCase =( config.patch_size if isinstance(config.patch_size , collections.abc.Iterable ) else (config.patch_size, config.patch_size) ) __UpperCAmelCase =image_size[0] + patch_size[0] - (image_size[0] % patch_size[0]) __UpperCAmelCase =image_size[1] + patch_size[1] - (image_size[1] % patch_size[1]) for model_class in self.all_model_classes: __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __UpperCAmelCase =True self.check_hidden_states_output(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , (padded_height, padded_width) ) def _a ( self : Optional[int] ) -> Tuple: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple ) -> Dict: __UpperCAmelCase =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__SCREAMING_SNAKE_CASE ) @slow def _a ( self : int ) -> Dict: for model_name in SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCAmelCase =SwinvaModel.from_pretrained(__SCREAMING_SNAKE_CASE ) self.assertIsNotNone(__SCREAMING_SNAKE_CASE ) def _a ( self : Dict ) -> Union[str, Any]: __UpperCAmelCase , __UpperCAmelCase =self.model_tester.prepare_config_and_inputs_for_common() __UpperCAmelCase =_config_zero_init(__SCREAMING_SNAKE_CASE ) for model_class in self.all_model_classes: __UpperCAmelCase =model_class(config=__SCREAMING_SNAKE_CASE ) for name, param in model.named_parameters(): if "embeddings" not in name and "logit_scale" not in name and param.requires_grad: self.assertIn( ((param.data.mean() * 1e9).round() / 1e9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @require_vision @require_torch class _A ( unittest.TestCase ): """simple docstring""" @cached_property def _a ( self : Tuple ) -> Dict: return ( AutoImageProcessor.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ) if is_vision_available() else None ) @slow def _a ( self : int ) -> Optional[int]: __UpperCAmelCase =SwinvaForImageClassification.from_pretrained("""microsoft/swinv2-tiny-patch4-window8-256""" ).to( __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.default_image_processor __UpperCAmelCase =Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) __UpperCAmelCase =image_processor(images=__SCREAMING_SNAKE_CASE , return_tensors="""pt""" ).to(__SCREAMING_SNAKE_CASE ) # forward pass with torch.no_grad(): __UpperCAmelCase =model(**__SCREAMING_SNAKE_CASE ) # verify the logits __UpperCAmelCase =torch.Size((1, 1000) ) self.assertEqual(outputs.logits.shape , __SCREAMING_SNAKE_CASE ) __UpperCAmelCase =torch.tensor([-0.3_947, -0.4_306, 0.0_026] ).to(__SCREAMING_SNAKE_CASE ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __SCREAMING_SNAKE_CASE , atol=1e-4 ) )
68
0
"""simple docstring""" import functools import gc import inspect import torch from .imports import is_npu_available, is_xpu_available def lowercase__( *__SCREAMING_SNAKE_CASE : Optional[int] ): if not isinstance(A_ , A_ ): lowercase_ : List[Any] = list(A_ ) for i in range(len(A_ ) ): lowercase_ : Tuple = None gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() return objects def lowercase__( __SCREAMING_SNAKE_CASE : Exception ): lowercase_ : int = [ 'CUDA out of memory.', # CUDA OOM 'cuDNN error: CUDNN_STATUS_NOT_SUPPORTED.', # CUDNN SNAFU 'DefaultCPUAllocator: can\'t allocate memory', # CPU OOM ] if isinstance(A_ , A_ ) and len(exception.args ) == 1: return any(err in exception.args[0] for err in _statements ) return False def lowercase__( __SCREAMING_SNAKE_CASE : callable = None , __SCREAMING_SNAKE_CASE : int = 1_28 ): if function is None: return functools.partial(A_ , starting_batch_size=A_ ) lowercase_ : Union[str, Any] = starting_batch_size def decorator(*__SCREAMING_SNAKE_CASE : int , **__SCREAMING_SNAKE_CASE : Any ): nonlocal batch_size gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() lowercase_ : Any = list(inspect.signature(A_ ).parameters.keys() ) # Guard against user error if len(A_ ) < (len(A_ ) + 1): lowercase_ : Dict = ', '.join([F'''{arg}={value}''' for arg, value in zip(params[1:] , args[1:] )] ) raise TypeError( F'''Batch size was passed into `{function.__name__}` as the first argument when called.''' F'''Remove this as the decorator already does so: `{function.__name__}({arg_str})`''' ) while True: if batch_size == 0: raise RuntimeError('No executable batch size found, reached zero.' ) try: return function(A_ , *A_ , **A_ ) except Exception as e: if should_reduce_batch_size(A_ ): gc.collect() if is_xpu_available(): torch.xpu.empty_cache() elif is_npu_available(): torch.npu.empty_cache() else: torch.cuda.empty_cache() batch_size //= 2 else: raise return decorator
425
import os import re import unicodedata from shutil import copyfile from typing import TYPE_CHECKING, Any, Dict, List, Optional, Tuple, Union import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import is_torch_available, logging if is_torch_available(): import torch if TYPE_CHECKING: from transformers.pipelines.conversational import Conversation __A = logging.get_logger(__name__) __A = {"vocab_file": "spiece.model"} __A = { "vocab_file": { "AI-Sweden/gpt-sw3-126m": "https://huggingface.co/AI-Sweden/gpt-sw3-126m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-350m": "https://huggingface.co/AI-Sweden/gpt-sw3-350m/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-1.6b": "https://huggingface.co/AI-Sweden/gpt-sw3-1.6b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-6.7b": "https://huggingface.co/AI-Sweden/gpt-sw3-6.7b/resolve/main/spiece.model", "AI-Sweden/gpt-sw3-20b": "https://huggingface.co/AI-Sweden/gpt-sw3-20b/resolve/main/spiece.model", } } __A = { "AI-Sweden/gpt-sw3-126m": 20_48, "AI-Sweden/gpt-sw3-350m": 20_48, "AI-Sweden/gpt-sw3-1.6b": 20_48, "AI-Sweden/gpt-sw3-6.7b": 20_48, "AI-Sweden/gpt-sw3-20b": 20_48, } class _A ( UpperCamelCase ): """simple docstring""" lowerCamelCase : int = VOCAB_FILES_NAMES lowerCamelCase : int = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase : Optional[Any] = ['input_ids', 'attention_mask'] def __init__( self : Dict , __SCREAMING_SNAKE_CASE : Union[str, Any] , __SCREAMING_SNAKE_CASE : Optional[Any]=False , __SCREAMING_SNAKE_CASE : Union[str, Any]=False , __SCREAMING_SNAKE_CASE : List[str]=False , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : List[Any]=None , __SCREAMING_SNAKE_CASE : Dict=None , __SCREAMING_SNAKE_CASE : Any=None , __SCREAMING_SNAKE_CASE : Optional[Dict[str, Any]] = None , **__SCREAMING_SNAKE_CASE : Optional[Any] , ) -> None: __UpperCAmelCase ={} if sp_model_kwargs is None else sp_model_kwargs __UpperCAmelCase =kwargs.get("""name_or_path""" ) if name_or_path is None: logger.warning( """name_or_path not provided, will work for all GPTSw3 models except gpt-sw3-7b,""" """ you are testing the model, this can safely be ignored""" ) __UpperCAmelCase ="""None""" # Default definitions for our 2 tokenizer versions, with None-checks to enable proper testing __UpperCAmelCase ="""<|endoftext|>""" if eos_token is None else eos_token __UpperCAmelCase ="""<unk>""" if unk_token is None else unk_token if "gpt-sw3-7b" in name_or_path: __UpperCAmelCase =unk_token if pad_token is None else pad_token __UpperCAmelCase =eos_token if bos_token is None else bos_token else: __UpperCAmelCase ="""<pad>""" if pad_token is None else pad_token __UpperCAmelCase ="""<s>""" if bos_token is None else bos_token super().__init__( do_lower_case=__SCREAMING_SNAKE_CASE , remove_space=__SCREAMING_SNAKE_CASE , keep_accents=__SCREAMING_SNAKE_CASE , bos_token=__SCREAMING_SNAKE_CASE , eos_token=__SCREAMING_SNAKE_CASE , unk_token=__SCREAMING_SNAKE_CASE , pad_token=__SCREAMING_SNAKE_CASE , sp_model_kwargs=self.sp_model_kwargs , **__SCREAMING_SNAKE_CASE , ) __UpperCAmelCase =do_lower_case __UpperCAmelCase =remove_space __UpperCAmelCase =keep_accents __UpperCAmelCase =vocab_file __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__SCREAMING_SNAKE_CASE ) # Used for whitespace normalization in input texts # fmt : off __UpperCAmelCase ={""" """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """ """, """""", """„"""} # fmt : on # Regular expression to remove non-printing characters (e.g. some unicode control chars) in preprocessing __UpperCAmelCase =re.compile( f'''[{"".join(map(__SCREAMING_SNAKE_CASE , list(range(0 , 9 ) ) + list(range(11 , 32 ) ) + list(range(127 , 160 ) ) + [160, 173, 8203] ) )}]''' ) def __getstate__( self : Any ) -> str: __UpperCAmelCase =self.__dict__.copy() __UpperCAmelCase =None return state def __setstate__( self : str , __SCREAMING_SNAKE_CASE : Optional[Any] ) -> Union[str, Any]: __UpperCAmelCase =d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): __UpperCAmelCase ={} __UpperCAmelCase =spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) @property # Copied from transformers.models.albert.tokenization_albert.AlbertTokenizer.vocab_size def _a ( self : Union[str, Any] ) -> int: return len(self.sp_model ) def _a ( self : Dict , __SCREAMING_SNAKE_CASE : str ) -> str: __UpperCAmelCase =self.non_printing_characters_re.sub("""""" , __SCREAMING_SNAKE_CASE ) # Normalize whitespaces __UpperCAmelCase ="""""".join([char if char not in self.whitespaces else """ """ for char in text] ) # NFC Unicode normalization __UpperCAmelCase =unicodedata.normalize("""NFC""" , __SCREAMING_SNAKE_CASE ) return text def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str , **__SCREAMING_SNAKE_CASE : Union[str, Any] ) -> List[str]: __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) return self.sp_model.encode(__SCREAMING_SNAKE_CASE , out_type=__SCREAMING_SNAKE_CASE ) def _a ( self : List[str] , __SCREAMING_SNAKE_CASE : str ) -> int: return self.sp_model.PieceToId(__SCREAMING_SNAKE_CASE ) def _a ( self : Tuple , __SCREAMING_SNAKE_CASE : int ) -> str: return self.sp_model.IdToPiece(__SCREAMING_SNAKE_CASE ) @staticmethod def _a ( __SCREAMING_SNAKE_CASE : str ) -> str: return out_string def _a ( self : Any , __SCREAMING_SNAKE_CASE : List[str] ) -> str: __UpperCAmelCase =[] __UpperCAmelCase ="""""" __UpperCAmelCase =False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: # TODO: Check if this is needed, as it ensures that decode(encode(doc)) != doc by adding extra whitespace in the decoded document if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) + token __UpperCAmelCase =True __UpperCAmelCase =[] else: current_sub_tokens.append(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =False out_string += self.sp_model.decode(__SCREAMING_SNAKE_CASE ) return out_string def _a ( self : Any ) -> Dict[str, int]: __UpperCAmelCase ={self.convert_ids_to_tokens(__SCREAMING_SNAKE_CASE ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _a ( self : Any , __SCREAMING_SNAKE_CASE : str , __SCREAMING_SNAKE_CASE : Optional[str] = None ) -> Tuple[str]: if not os.path.isdir(__SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return __UpperCAmelCase =os.path.join( __SCREAMING_SNAKE_CASE , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__SCREAMING_SNAKE_CASE ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __SCREAMING_SNAKE_CASE ) elif not os.path.isfile(self.vocab_file ): with open(__SCREAMING_SNAKE_CASE , """wb""" ) as fi: __UpperCAmelCase =self.sp_model.serialized_model_proto() fi.write(__SCREAMING_SNAKE_CASE ) return (out_vocab_file,) def _a ( self : List[Any] , __SCREAMING_SNAKE_CASE : Union[str, List[str]] , __SCREAMING_SNAKE_CASE : Union[str, bool] = False ) -> Union[List[int], List[List[int]], "torch.Tensor"]: if isinstance(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ): __UpperCAmelCase =self.preprocess_text(__SCREAMING_SNAKE_CASE ) __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) else: __UpperCAmelCase =[self.preprocess_text(__SCREAMING_SNAKE_CASE ) for t in text] __UpperCAmelCase =self.sp_model.encode(__SCREAMING_SNAKE_CASE ) if return_tensors is True or return_tensors == "pt": __UpperCAmelCase =torch.tensor(__SCREAMING_SNAKE_CASE ) return token_ids def _a ( self : str , __SCREAMING_SNAKE_CASE : Union[int, List[int]] ) -> str: return self.sp_model.decode(__SCREAMING_SNAKE_CASE ) def _a ( self : Optional[Any] , __SCREAMING_SNAKE_CASE : "Conversation" ) -> List[int]: __UpperCAmelCase =[f'''User: {text}''' if is_user else f'''Bot: {text}''' for is_user, text in conversation.iter_texts()] __UpperCAmelCase =( f'''{self.eos_token}{self.bos_token}''' + f'''{self.bos_token}'''.join(__SCREAMING_SNAKE_CASE ) + f'''{self.bos_token}Bot:''' ) return self.encode(text=__SCREAMING_SNAKE_CASE )
68
0
'''simple docstring''' from collections.abc import Callable def __UpperCAmelCase ( _UpperCAmelCase : Callable[[float], float] , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: __snake_case = a __snake_case = b if function(_UpperCAmelCase ) == 0: # one of the a or b is a root for the function return a elif function(_UpperCAmelCase ) == 0: return b elif ( function(_UpperCAmelCase ) * function(_UpperCAmelCase ) > 0 ): # if none of these are root and they are both positive or negative, # then this algorithm can't find the root raise ValueError("could not find root in given interval." ) else: __snake_case = start + (end - start) / 2.0 while abs(start - mid ) > 10**-7: # until precisely equals to 10^-7 if function(_UpperCAmelCase ) == 0: return mid elif function(_UpperCAmelCase ) * function(_UpperCAmelCase ) < 0: __snake_case = mid else: __snake_case = mid __snake_case = start + (end - start) / 2.0 return mid def __UpperCAmelCase ( _UpperCAmelCase : float ) -> float: return x**3 - 2 * x - 5 if __name__ == "__main__": print(bisection(f, 1, 1_000)) import doctest doctest.testmod()
69
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
1
'''simple docstring''' from argparse import ArgumentParser from .env import EnvironmentCommand def __UpperCAmelCase ( ) -> Optional[Any]: __snake_case = ArgumentParser("Diffusers CLI tool" , usage="diffusers-cli <command> [<args>]" ) __snake_case = parser.add_subparsers(help="diffusers-cli command helpers" ) # Register commands EnvironmentCommand.register_subcommand(_UpperCAmelCase ) # Let's go __snake_case = parser.parse_args() if not hasattr(_UpperCAmelCase , "func" ): parser.print_help() exit(1 ) # Run __snake_case = args.func(_UpperCAmelCase ) service.run() if __name__ == "__main__": main()
69
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
1
'''simple docstring''' from __future__ import annotations class SCREAMING_SNAKE_CASE__ : def __init__( self : Dict , a_ : int = 0 ): """simple docstring""" __snake_case = key def A ( self : Union[str, Any] , a_ : str , a_ : int ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) __snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(a_ ) ^ key ) for ch in content] def A ( self : int , a_ : str , a_ : int ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) __snake_case = key or self.__key or 1 # make sure key is an appropriate size key %= 255 return [chr(ord(a_ ) ^ key ) for ch in content] def A ( self : int , a_ : str , a_ : int = 0 ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) __snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned __snake_case = "" for ch in content: ans += chr(ord(a_ ) ^ key ) return ans def A ( self : Union[str, Any] , a_ : str , a_ : int = 0 ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) __snake_case = key or self.__key or 1 # make sure key can be any size while key > 255: key -= 255 # This will be returned __snake_case = "" for ch in content: ans += chr(ord(a_ ) ^ key ) return ans def A ( self : str , a_ : str , a_ : int = 0 ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) try: with open(a_ ) as fin, open("encrypt.out" , "w+" ) as fout: # actual encrypt-process for line in fin: fout.write(self.encrypt_string(a_ , a_ ) ) except OSError: return False return True def A ( self : str , a_ : str , a_ : int ): """simple docstring""" assert isinstance(a_ , a_ ) and isinstance(a_ , a_ ) try: with open(a_ ) as fin, open("decrypt.out" , "w+" ) as fout: # actual encrypt-process for line in fin: fout.write(self.decrypt_string(a_ , a_ ) ) except OSError: return False return True # Tests # crypt = XORCipher() # key = 67 # # test encrypt # print(crypt.encrypt("hallo welt",key)) # # test decrypt # print(crypt.decrypt(crypt.encrypt("hallo welt",key), key)) # # test encrypt_string # print(crypt.encrypt_string("hallo welt",key)) # # test decrypt_string # print(crypt.decrypt_string(crypt.encrypt_string("hallo welt",key),key)) # if (crypt.encrypt_file("test.txt",key)): # print("encrypt successful") # else: # print("encrypt unsuccessful") # if (crypt.decrypt_file("encrypt.out",key)): # print("decrypt successful") # else: # print("decrypt unsuccessful")
69
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
1
'''simple docstring''' import itertools import os from collections import Counter, defaultdict from concurrent.futures import ThreadPoolExecutor, as_completed import numpy as np import datasets from .execute import check_correctness a : Tuple = '''\ @misc{chen2021evaluating, title={Evaluating Large Language Models Trained on Code}, author={Mark Chen and Jerry Tworek and Heewoo Jun and Qiming Yuan \ and Henrique Ponde de Oliveira Pinto and Jared Kaplan and Harri Edwards \ and Yuri Burda and Nicholas Joseph and Greg Brockman and Alex Ray \ and Raul Puri and Gretchen Krueger and Michael Petrov and Heidy Khlaaf \ and Girish Sastry and Pamela Mishkin and Brooke Chan and Scott Gray \ and Nick Ryder and Mikhail Pavlov and Alethea Power and Lukasz Kaiser \ and Mohammad Bavarian and Clemens Winter and Philippe Tillet \ and Felipe Petroski Such and Dave Cummings and Matthias Plappert \ and Fotios Chantzis and Elizabeth Barnes and Ariel Herbert-Voss \ and William Hebgen Guss and Alex Nichol and Alex Paino and Nikolas Tezak \ and Jie Tang and Igor Babuschkin and Suchir Balaji and Shantanu Jain \ and William Saunders and Christopher Hesse and Andrew N. Carr \ and Jan Leike and Josh Achiam and Vedant Misra and Evan Morikawa \ and Alec Radford and Matthew Knight and Miles Brundage and Mira Murati \ and Katie Mayer and Peter Welinder and Bob McGrew and Dario Amodei \ and Sam McCandlish and Ilya Sutskever and Wojciech Zaremba}, year={2021}, eprint={2107.03374}, archivePrefix={arXiv}, primaryClass={cs.LG} } ''' a : Optional[int] = '''\ This metric implements the evaluation harness for the HumanEval problem solving dataset described in the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). ''' a : Tuple = ''' Calculates how good are predictions given some references, using certain scores Args: predictions: list of candidates to evaluate. Each candidates should be a list of strings with several code candidates to solve the problem. references: a list with a test for each prediction. Each test should evaluate the correctness of a code candidate. k: number of code candidates to consider in the evaluation (Default: [1, 10, 100]) num_workers: number of workers used to evaluate the canidate programs (Default: 4). timeout: Returns: pass_at_k: dict with pass rates for each k results: dict with granular results of each unittest Examples: >>> code_eval = datasets.load_metric("code_eval") >>> test_cases = ["assert add(2,3)==5"] >>> candidates = [["def add(a,b): return a*b", "def add(a, b): return a+b"]] >>> pass_at_k, results = code_eval.compute(references=test_cases, predictions=candidates, k=[1, 2]) >>> print(pass_at_k) {\'pass@1\': 0.5, \'pass@2\': 1.0} ''' a : Optional[int] = ''' ################################################################################ !!!WARNING!!! ################################################################################ The "code_eval" metric executes untrusted model-generated code in Python. Although it is highly unlikely that model-generated code will do something overtly malicious in response to this test suite, model-generated code may act destructively due to a lack of model capability or alignment. Users are strongly encouraged to sandbox this evaluation suite so that it does not perform destructive actions on their host or network. For more information on how OpenAI sandboxes its code, see the paper "Evaluating Large Language Models Trained on Code" (https://arxiv.org/abs/2107.03374). Once you have read this disclaimer and taken appropriate precautions, set the environment variable HF_ALLOW_CODE_EVAL="1". Within Python you can to this with: >>> import os >>> os.environ["HF_ALLOW_CODE_EVAL"] = "1" ################################################################################\ ''' a : Optional[int] = '''The MIT License Copyright (c) OpenAI (https://openai.com) Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restriction, including without limitation the rights to use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions: The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software. THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def A ( self : Dict ): """simple docstring""" return datasets.MetricInfo( # This is the description that will appear on the metrics page. description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" ) ), "references": datasets.Value("string" ), } ) , homepage="https://github.com/openai/human-eval" , codebase_urls=["https://github.com/openai/human-eval"] , reference_urls=["https://github.com/openai/human-eval"] , license=_LICENSE , ) def A ( self : str , a_ : List[Any] , a_ : Optional[int] , a_ : Any=[1, 10, 100] , a_ : Optional[int]=4 , a_ : Optional[int]=3.0 ): """simple docstring""" if os.getenv("HF_ALLOW_CODE_EVAL" , 0 ) != "1": raise ValueError(_WARNING ) if os.name == "nt": raise NotImplementedError("This metric is currently not supported on Windows." ) with ThreadPoolExecutor(max_workers=a_ ) as executor: __snake_case = [] __snake_case = Counter() __snake_case = 0 __snake_case = defaultdict(a_ ) for task_id, (candidates, test_case) in enumerate(zip(a_ , a_ ) ): for candidate in candidates: __snake_case = candidate + "\n" + test_case __snake_case = (test_program, timeout, task_id, completion_id[task_id]) __snake_case = executor.submit(a_ , *a_ ) futures.append(a_ ) completion_id[task_id] += 1 n_samples += 1 for future in as_completed(a_ ): __snake_case = future.result() results[result["task_id"]].append((result["completion_id"], result) ) __snake_case , __snake_case = [], [] for result in results.values(): result.sort() __snake_case = [r[1]["passed"] for r in result] total.append(len(a_ ) ) correct.append(sum(a_ ) ) __snake_case = np.array(a_ ) __snake_case = np.array(a_ ) __snake_case = k __snake_case = {f'''pass@{k}''': estimate_pass_at_k(a_ , a_ , a_ ).mean() for k in ks if (total >= k).all()} return pass_at_k, results def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any ) -> List[Any]: def estimator(_UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> float: if n - c < k: return 1.0 return 1.0 - np.prod(1.0 - k / np.arange(n - c + 1 , n + 1 ) ) if isinstance(_UpperCAmelCase , _UpperCAmelCase ): __snake_case = itertools.repeat(_UpperCAmelCase , len(_UpperCAmelCase ) ) else: assert len(_UpperCAmelCase ) == len(_UpperCAmelCase ) __snake_case = iter(_UpperCAmelCase ) return np.array([estimator(int(_UpperCAmelCase ) , int(_UpperCAmelCase ) , _UpperCAmelCase ) for n, c in zip(_UpperCAmelCase , _UpperCAmelCase )] )
69
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
1
'''simple docstring''' from typing import List import jiwer import jiwer.transforms as tr from packaging import version import datasets from datasets.config import PY_VERSION if PY_VERSION < version.parse('''3.8'''): import importlib_metadata else: import importlib.metadata as importlib_metadata a : Optional[int] = '''''' if version.parse(importlib_metadata.version('''jiwer''')) < version.parse('''2.3.0'''): class SCREAMING_SNAKE_CASE__ ( tr.AbstractTransform ): def __init__( self : int , a_ : str = " " ): """simple docstring""" __snake_case = sentence_delimiter def A ( self : Tuple , a_ : str ): """simple docstring""" return list(a_ ) def A ( self : Tuple , a_ : List[str] ): """simple docstring""" __snake_case = [] for sent_idx, sentence in enumerate(a_ ): chars.extend(self.process_string(a_ ) ) if self.sentence_delimiter is not None and self.sentence_delimiter != "" and sent_idx < len(a_ ) - 1: chars.append(self.sentence_delimiter ) return chars a : Optional[int] = tr.Compose( [tr.RemoveMultipleSpaces(), tr.Strip(), SentencesToListOfCharacters(SENTENCE_DELIMITER)] ) else: a : Dict = tr.Compose( [ tr.RemoveMultipleSpaces(), tr.Strip(), tr.ReduceToSingleSentence(SENTENCE_DELIMITER), tr.ReduceToListOfListOfChars(), ] ) a : str = '''\ @inproceedings{inproceedings, author = {Morris, Andrew and Maier, Viktoria and Green, Phil}, year = {2004}, month = {01}, pages = {}, title = {From WER and RIL to MER and WIL: improved evaluation measures for connected speech recognition.} } ''' a : str = '''\ Character error rate (CER) is a common metric of the performance of an automatic speech recognition system. CER is similar to Word Error Rate (WER), but operates on character instead of word. Please refer to docs of WER for further information. Character error rate can be computed as: CER = (S + D + I) / N = (S + D + I) / (S + D + C) where S is the number of substitutions, D is the number of deletions, I is the number of insertions, C is the number of correct characters, N is the number of characters in the reference (N=S+D+C). CER\'s output is not always a number between 0 and 1, in particular when there is a high number of insertions. This value is often associated to the percentage of characters that were incorrectly predicted. The lower the value, the better the performance of the ASR system with a CER of 0 being a perfect score. ''' a : Any = ''' Computes CER score of transcribed segments against references. Args: references: list of references for each speech input. predictions: list of transcribtions to score. concatenate_texts: Whether or not to concatenate sentences before evaluation, set to True for more accurate result. Returns: (float): the character error rate Examples: >>> predictions = ["this is the prediction", "there is an other sample"] >>> references = ["this is the reference", "there is another one"] >>> cer = datasets.load_metric("cer") >>> cer_score = cer.compute(predictions=predictions, references=references) >>> print(cer_score) 0.34146341463414637 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class SCREAMING_SNAKE_CASE__ ( datasets.Metric ): def A ( self : Union[str, Any] ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/jitsi/jiwer/"] , reference_urls=[ "https://en.wikipedia.org/wiki/Word_error_rate", "https://sites.google.com/site/textdigitisation/qualitymeasures/computingerrorrates", ] , ) def A ( self : Tuple , a_ : Any , a_ : List[str] , a_ : Optional[Any]=False ): """simple docstring""" if concatenate_texts: return jiwer.compute_measures( a_ , a_ , truth_transform=a_ , hypothesis_transform=a_ , )["wer"] __snake_case = 0 __snake_case = 0 for prediction, reference in zip(a_ , a_ ): __snake_case = jiwer.compute_measures( a_ , a_ , truth_transform=a_ , hypothesis_transform=a_ , ) incorrect += measures["substitutions"] + measures["deletions"] + measures["insertions"] total += measures["substitutions"] + measures["deletions"] + measures["hits"] return incorrect / total
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' a : Union[str, Any] = {'''a''': ['''c''', '''b'''], '''b''': ['''d''', '''e'''], '''c''': [], '''d''': [], '''e''': []} a : Tuple = ['''a''', '''b''', '''c''', '''d''', '''e'''] def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ) -> Optional[Any]: __snake_case = start # add current to visited visited.append(_UpperCAmelCase ) __snake_case = edges[current] for neighbor in neighbors: # if neighbor not in visited, visit if neighbor not in visited: __snake_case = topological_sort(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # if all neighbors visited add current to sort sort.append(_UpperCAmelCase ) # if all vertices haven't been visited select a new one to visit if len(_UpperCAmelCase ) != len(_UpperCAmelCase ): for vertice in vertices: if vertice not in visited: __snake_case = topological_sort(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # return sort return sort if __name__ == "__main__": a : List[str] = topological_sort('''a''', [], []) print(sort)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging a : str = logging.get_logger(__name__) a : Dict = { '''SCUT-DLVCLab/lilt-roberta-en-base''': ( '''https://huggingface.co/SCUT-DLVCLab/lilt-roberta-en-base/resolve/main/config.json''' ), } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """lilt""" def __init__( self : List[Any] , a_ : List[Any]=30_522 , a_ : Optional[int]=768 , a_ : Optional[Any]=12 , a_ : Union[str, Any]=12 , a_ : Optional[int]=3_072 , a_ : Dict="gelu" , a_ : Union[str, Any]=0.1 , a_ : str=0.1 , a_ : Optional[int]=512 , a_ : Tuple=2 , a_ : Dict=0.02 , a_ : Tuple=1e-12 , a_ : str=0 , a_ : Union[str, Any]="absolute" , a_ : Dict=None , a_ : List[str]=4 , a_ : Optional[Any]=1_024 , **a_ : str , ): """simple docstring""" super().__init__(pad_token_id=a_ , **a_ ) __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = hidden_act __snake_case = intermediate_size __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = initializer_range __snake_case = layer_norm_eps __snake_case = position_embedding_type __snake_case = classifier_dropout __snake_case = channel_shrink_ratio __snake_case = max_ad_position_embeddings
69
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
1
'''simple docstring''' import inspect import unittest from transformers import DecisionTransformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import DecisionTransformerModel from transformers.models.decision_transformer.modeling_decision_transformer import ( DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[Any] , a_ : Dict=13 , a_ : Optional[Any]=7 , a_ : int=6 , a_ : str=17 , a_ : Union[str, Any]=23 , a_ : Dict=11 , a_ : Optional[Any]=True , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = act_dim __snake_case = state_dim __snake_case = hidden_size __snake_case = max_length __snake_case = is_training def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = floats_tensor((self.batch_size, self.seq_length, self.state_dim) ) __snake_case = floats_tensor((self.batch_size, self.seq_length, self.act_dim) ) __snake_case = floats_tensor((self.batch_size, self.seq_length, 1) ) __snake_case = floats_tensor((self.batch_size, self.seq_length, 1) ) __snake_case = ids_tensor((self.batch_size, self.seq_length) , vocab_size=1_000 ) __snake_case = random_attention_mask((self.batch_size, self.seq_length) ) __snake_case = self.get_config() return ( config, states, actions, rewards, returns_to_go, timesteps, attention_mask, ) def A ( self : List[str] ): """simple docstring""" return DecisionTransformerConfig( batch_size=self.batch_size , seq_length=self.seq_length , act_dim=self.act_dim , state_dim=self.state_dim , hidden_size=self.hidden_size , max_length=self.max_length , ) def A ( self : List[str] , a_ : Dict , a_ : Any , a_ : Optional[Any] , a_ : Optional[int] , a_ : Tuple , a_ : List[str] , a_ : Any , ): """simple docstring""" __snake_case = DecisionTransformerModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ , a_ , a_ , a_ , a_ ) self.parent.assertEqual(result.state_preds.shape , states.shape ) self.parent.assertEqual(result.action_preds.shape , actions.shape ) self.parent.assertEqual(result.return_preds.shape , returns_to_go.shape ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length * 3, self.hidden_size) ) # seq length *3 as there are 3 modelities: states, returns and actions def A ( self : Optional[int] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = { "states": states, "actions": actions, "rewards": rewards, "returns_to_go": returns_to_go, "timesteps": timesteps, "attention_mask": attention_mask, } return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = (DecisionTransformerModel,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = () __SCREAMING_SNAKE_CASE = {"""feature-extraction""": DecisionTransformerModel} if is_torch_available() else {} # Ignoring of a failing test from GenerationTesterMixin, as the model does not use inputs_ids __SCREAMING_SNAKE_CASE = False # Ignoring of a failing tests from ModelTesterMixin, as the model does not implement these features __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[int] ): """simple docstring""" __snake_case = DecisionTransformerModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) @slow def A ( self : int ): """simple docstring""" for model_name in DECISION_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case = DecisionTransformerModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) def A ( self : int ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case = model_class(a_ ) __snake_case = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case = [*signature.parameters.keys()] __snake_case = [ "states", "actions", "rewards", "returns_to_go", "timesteps", "attention_mask", ] self.assertListEqual(arg_names[: len(a_ )] , a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : List[Any] ): """simple docstring""" __snake_case = 2 # number of steps of autoregressive prediction we will perform __snake_case = 10 # defined by the RL environment, may be normalized __snake_case = DecisionTransformerModel.from_pretrained("edbeeching/decision-transformer-gym-hopper-expert" ) __snake_case = model.to(a_ ) __snake_case = model.config torch.manual_seed(0 ) __snake_case = torch.randn(1 , 1 , config.state_dim ).to(device=a_ , dtype=torch.floataa ) # env.reset() __snake_case = torch.tensor( [[0.242793, -0.28693074, 0.8742613], [0.67815274, -0.08101085, -0.12952147]] , device=a_ ) __snake_case = torch.tensor(a_ , device=a_ , dtype=torch.floataa ).reshape(1 , 1 , 1 ) __snake_case = state __snake_case = torch.zeros(1 , 0 , config.act_dim , device=a_ , dtype=torch.floataa ) __snake_case = torch.zeros(1 , 0 , device=a_ , dtype=torch.floataa ) __snake_case = torch.tensor(0 , device=a_ , dtype=torch.long ).reshape(1 , 1 ) for step in range(a_ ): __snake_case = torch.cat([actions, torch.zeros(1 , 1 , config.act_dim , device=a_ )] , dim=1 ) __snake_case = torch.cat([rewards, torch.zeros(1 , 1 , device=a_ )] , dim=1 ) __snake_case = torch.ones(1 , states.shape[1] ).to(dtype=torch.long , device=states.device ) with torch.no_grad(): __snake_case , __snake_case , __snake_case = model( states=a_ , actions=a_ , rewards=a_ , returns_to_go=a_ , timesteps=a_ , attention_mask=a_ , return_dict=a_ , ) self.assertEqual(action_pred.shape , actions.shape ) self.assertTrue(torch.allclose(action_pred[0, -1] , expected_outputs[step] , atol=1e-4 ) ) __snake_case , __snake_case , __snake_case , __snake_case = ( # env.step(action) torch.randn(1 , 1 , config.state_dim ).to(device=a_ , dtype=torch.floataa ), 1.0, False, {}, ) __snake_case = action_pred[0, -1] __snake_case = torch.cat([states, state] , dim=1 ) __snake_case = returns_to_go[0, -1] - reward __snake_case = torch.cat([returns_to_go, pred_return.reshape(1 , 1 , 1 )] , dim=1 ) __snake_case = torch.cat( [timesteps, torch.ones((1, 1) , device=a_ , dtype=torch.long ) * (step + 1)] , dim=1 )
69
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float(moles / volume ) * nfactor ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((moles * 0.0821 * temperature) / (volume) ) ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((moles * 0.0821 * temperature) / (pressure) ) ) def __UpperCAmelCase ( _UpperCAmelCase : float , _UpperCAmelCase : float , _UpperCAmelCase : float ) -> float: return round(float((pressure * volume) / (0.0821 * moles) ) ) if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> float: __snake_case = (num_of_terms / 2) * (2 * first_term + (num_of_terms - 1) * common_diff) # formula for sum of series return total def __UpperCAmelCase ( ) -> List[Any]: print(sum_of_series(1 , 1 , 10 ) ) if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' import math def __UpperCAmelCase ( _UpperCAmelCase : list , _UpperCAmelCase : int ) -> int: __snake_case = len(_UpperCAmelCase ) __snake_case = int(math.floor(math.sqrt(_UpperCAmelCase ) ) ) __snake_case = 0 while arr[min(_UpperCAmelCase , _UpperCAmelCase ) - 1] < x: __snake_case = step step += int(math.floor(math.sqrt(_UpperCAmelCase ) ) ) if prev >= n: return -1 while arr[prev] < x: __snake_case = prev + 1 if prev == min(_UpperCAmelCase , _UpperCAmelCase ): return -1 if arr[prev] == x: return prev return -1 if __name__ == "__main__": a : int = input('''Enter numbers separated by a comma:\n''').strip() a : int = [int(item) for item in user_input.split(''',''')] a : Optional[Any] = int(input('''Enter the number to be searched:\n''')) a : Tuple = jump_search(arr, x) if res == -1: print('''Number not found!''') else: print(F'''Number {x} is at index {res}''')
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' from math import factorial def __UpperCAmelCase ( _UpperCAmelCase : int = 20 ) -> int: __snake_case = 2 * n # middle entry of odd rows starting at row 3 is the solution for n = 1, # 2, 3,... __snake_case = n // 2 return int(factorial(_UpperCAmelCase ) / (factorial(_UpperCAmelCase ) * factorial(n - k )) ) if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution(20)) else: try: a : Optional[Any] = int(sys.argv[1]) print(solution(n)) except ValueError: print('''Invalid entry - please enter a number.''')
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' from multiprocessing import Lock, Pipe, Process # lock used to ensure that two processes do not access a pipe at the same time a : Optional[Any] = Lock() def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[int] ) -> Tuple: global process_lock # we perform n swaps since after n swaps we know we are sorted # we *could* stop early if we are sorted already, but it takes as long to # find out we are sorted as it does to sort the list with this algorithm for i in range(0 , 10 ): if (i + position) % 2 == 0 and r_send is not None: # send your value to your right neighbor process_lock.acquire() r_send[1].send(_UpperCAmelCase ) process_lock.release() # receive your right neighbor's value process_lock.acquire() __snake_case = rr_cv[0].recv() process_lock.release() # take the lower value since you are on the left __snake_case = min(_UpperCAmelCase , _UpperCAmelCase ) elif (i + position) % 2 != 0 and l_send is not None: # send your value to your left neighbor process_lock.acquire() l_send[1].send(_UpperCAmelCase ) process_lock.release() # receive your left neighbor's value process_lock.acquire() __snake_case = lr_cv[0].recv() process_lock.release() # take the higher value since you are on the right __snake_case = max(_UpperCAmelCase , _UpperCAmelCase ) # after all swaps are performed, send the values back to main result_pipe[1].send(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> str: __snake_case = [] __snake_case = [] # initialize the list of pipes where the values will be retrieved for _ in arr: result_pipe.append(Pipe() ) # creates the processes # the first and last process only have one neighbor so they are made outside # of the loop __snake_case = Pipe() __snake_case = Pipe() process_array_.append( Process( target=_UpperCAmelCase , args=(0, arr[0], None, temp_rs, None, temp_rr, result_pipe[0]) , ) ) __snake_case = temp_rs __snake_case = temp_rr for i in range(1 , len(_UpperCAmelCase ) - 1 ): __snake_case = Pipe() __snake_case = Pipe() process_array_.append( Process( target=_UpperCAmelCase , args=(i, arr[i], temp_ls, temp_rs, temp_lr, temp_rr, result_pipe[i]) , ) ) __snake_case = temp_rs __snake_case = temp_rr process_array_.append( Process( target=_UpperCAmelCase , args=( len(_UpperCAmelCase ) - 1, arr[len(_UpperCAmelCase ) - 1], temp_ls, None, temp_lr, None, result_pipe[len(_UpperCAmelCase ) - 1], ) , ) ) # start the processes for p in process_array_: p.start() # wait for the processes to end and write their values to the list for p in range(0 , len(_UpperCAmelCase ) ): __snake_case = result_pipe[p][0].recv() process_array_[p].join() return arr def __UpperCAmelCase ( ) -> List[Any]: __snake_case = list(range(10 , 0 , -1 ) ) print("Initial List" ) print(*_UpperCAmelCase ) __snake_case = odd_even_transposition(_UpperCAmelCase ) print("Sorted List\n" ) print(*_UpperCAmelCase ) if __name__ == "__main__": main()
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> Tuple: __snake_case = 1 __snake_case = 2 while i * i <= n: __snake_case = 0 while n % i == 0: n //= i multiplicity += 1 n_divisors *= multiplicity + 1 i += 1 if n > 1: n_divisors *= 2 return n_divisors def __UpperCAmelCase ( ) -> Tuple: __snake_case = 1 __snake_case = 1 while True: i += 1 t_num += i if count_divisors(_UpperCAmelCase ) > 5_00: break return t_num if __name__ == "__main__": print(solution())
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' from typing import Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature from ...image_transforms import get_image_size, pad, rescale, to_channel_dimension_format from ...image_utils import ChannelDimension, ImageInput, make_list_of_images, to_numpy_array, valid_images from ...utils import TensorType, logging a : Any = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""pixel_values"""] def __init__( self : str , a_ : bool = True , a_ : Union[int, float] = 1 / 255 , a_ : bool = True , a_ : int = 8 , **a_ : Any , ): """simple docstring""" super().__init__(**a_ ) __snake_case = do_rescale __snake_case = rescale_factor __snake_case = do_pad __snake_case = pad_size def A ( self : Optional[Any] , a_ : np.ndarray , a_ : float , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Tuple ): """simple docstring""" return rescale(a_ , scale=a_ , data_format=a_ , **a_ ) def A ( self : Optional[int] , a_ : np.ndarray , a_ : int , a_ : Optional[Union[str, ChannelDimension]] = None ): """simple docstring""" __snake_case , __snake_case = get_image_size(a_ ) __snake_case = (old_height // size + 1) * size - old_height __snake_case = (old_width // size + 1) * size - old_width return pad(a_ , ((0, pad_height), (0, pad_width)) , mode="symmetric" , data_format=a_ ) def A ( self : int , a_ : ImageInput , a_ : Optional[bool] = None , a_ : Optional[float] = None , a_ : Optional[bool] = None , a_ : Optional[int] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : Union[str, ChannelDimension] = ChannelDimension.FIRST , **a_ : str , ): """simple docstring""" __snake_case = do_rescale if do_rescale is not None else self.do_rescale __snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case = do_pad if do_pad is not None else self.do_pad __snake_case = pad_size if pad_size is not None else self.pad_size __snake_case = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) # All transformations expect numpy arrays. __snake_case = [to_numpy_array(a_ ) for image in images] if do_rescale: __snake_case = [self.rescale(image=a_ , scale=a_ ) for image in images] if do_pad: __snake_case = [self.pad(a_ , size=a_ ) for image in images] __snake_case = [to_channel_dimension_format(a_ , a_ ) for image in images] __snake_case = {"pixel_values": images} return BatchFeature(data=a_ , tensor_type=a_ )
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' import json import os import re import sys import urllib.request import requests from bsa import BeautifulSoup a : Dict = { '''User-Agent''': '''Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36''' ''' (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582''' } def __UpperCAmelCase ( _UpperCAmelCase : str = "dhaka" , _UpperCAmelCase : int = 5 ) -> int: __snake_case = min(_UpperCAmelCase , 50 ) # Prevent abuse! __snake_case = { "q": query, "tbm": "isch", "hl": "en", "ijn": "0", } __snake_case = requests.get("https://www.google.com/search" , params=_UpperCAmelCase , headers=_UpperCAmelCase ) __snake_case = BeautifulSoup(html.text , "html.parser" ) __snake_case = "".join( re.findall(R"AF_initDataCallback\(([^<]+)\);" , str(soup.select("script" ) ) ) ) __snake_case = json.dumps(_UpperCAmelCase ) __snake_case = json.loads(_UpperCAmelCase ) __snake_case = re.findall( R"\[\"GRID_STATE0\",null,\[\[1,\[0,\".*?\",(.*),\"All\"," , _UpperCAmelCase , ) if not matched_google_image_data: return 0 __snake_case = re.sub( R"\[\"(https\:\/\/encrypted-tbn0\.gstatic\.com\/images\?.*?)\",\d+,\d+\]" , "" , str(_UpperCAmelCase ) , ) __snake_case = re.findall( R"(?:'|,),\[\"(https:|http.*?)\",\d+,\d+\]" , _UpperCAmelCase , ) for index, fixed_full_res_image in enumerate(_UpperCAmelCase ): if index >= max_images: return index __snake_case = bytes(_UpperCAmelCase , "ascii" ).decode( "unicode-escape" ) __snake_case = bytes(_UpperCAmelCase , "ascii" ).decode( "unicode-escape" ) __snake_case = urllib.request.build_opener() __snake_case = [ ( "User-Agent", "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36" " (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36 Edge/18.19582", ) ] urllib.request.install_opener(_UpperCAmelCase ) __snake_case = F'''query_{query.replace(" " , "_" )}''' if not os.path.exists(_UpperCAmelCase ): os.makedirs(_UpperCAmelCase ) urllib.request.urlretrieve( # noqa: S310 _UpperCAmelCase , F'''{path_name}/original_size_img_{index}.jpg''' ) return index if __name__ == "__main__": try: a : Any = download_images_from_google_query(sys.argv[1]) print(F'''{image_count} images were downloaded to disk.''') except IndexError: print('''Please provide a search term.''') raise
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1
'''simple docstring''' import math def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return math.sqrt(_UpperCAmelCase ) * math.sqrt(_UpperCAmelCase ) == num def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: __snake_case = 0 __snake_case = n while left <= right: __snake_case = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: __snake_case = mid - 1 else: __snake_case = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import mpmath # for roots of unity import numpy as np class SCREAMING_SNAKE_CASE__ : def __init__( self : Tuple , a_ : Optional[int]=None , a_ : int=None ): """simple docstring""" __snake_case = list(poly_a or [0] )[:] __snake_case = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() __snake_case = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() __snake_case = len(self.polyB ) # Add 0 to make lengths equal a power of 2 __snake_case = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform __snake_case = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product __snake_case = self.__multiply() def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" __snake_case = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(a_ ) <= 1: return dft[0] # __snake_case = self.c_max_length // 2 while next_ncol > 0: __snake_case = [[] for i in range(a_ )] __snake_case = self.root**next_ncol # First half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step __snake_case = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(a_ ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update __snake_case = new_dft __snake_case = next_ncol // 2 return dft[0] def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.__dft("A" ) __snake_case = self.__dft("B" ) __snake_case = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT __snake_case = 2 while next_ncol <= self.c_max_length: __snake_case = [[] for i in range(a_ )] __snake_case = self.root ** (next_ncol // 2) __snake_case = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update __snake_case = new_inverse_c next_ncol *= 2 # Unpack __snake_case = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__( self : Optional[int] ): """simple docstring""" __snake_case = "A = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyA[: self.len_A] ) ) __snake_case = "B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.polyB[: self.len_B] ) ) __snake_case = "A*B = " + " + ".join( f'''{coef}*x^{i}''' for coef, i in enumerate(self.product ) ) return f'''{a}\n{b}\n{c}''' # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import inspect from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch import torch.utils.checkpoint from ...models import UNetaDModel, VQModel from ...schedulers import ( DDIMScheduler, DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, PNDMScheduler, ) from ...utils import PIL_INTERPOLATION, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> int: __snake_case , __snake_case = image.size __snake_case , __snake_case = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __snake_case = image.resize((w, h) , resample=PIL_INTERPOLATION["lanczos"] ) __snake_case = np.array(_UpperCAmelCase ).astype(np.floataa ) / 255.0 __snake_case = image[None].transpose(0 , 3 , 1 , 2 ) __snake_case = torch.from_numpy(_UpperCAmelCase ) return 2.0 * image - 1.0 class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Dict , a_ : VQModel , a_ : UNetaDModel , a_ : Union[ DDIMScheduler, PNDMScheduler, LMSDiscreteScheduler, EulerDiscreteScheduler, EulerAncestralDiscreteScheduler, DPMSolverMultistepScheduler, ] , ): """simple docstring""" super().__init__() self.register_modules(vqvae=a_ , unet=a_ , scheduler=a_ ) @torch.no_grad() def __call__( self : Any , a_ : Union[torch.Tensor, PIL.Image.Image] = None , a_ : Optional[int] = 1 , a_ : Optional[int] = 100 , a_ : Optional[float] = 0.0 , a_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , a_ : Optional[str] = "pil" , a_ : bool = True , ): """simple docstring""" if isinstance(a_ , PIL.Image.Image ): __snake_case = 1 elif isinstance(a_ , torch.Tensor ): __snake_case = image.shape[0] else: raise ValueError(f'''`image` has to be of type `PIL.Image.Image` or `torch.Tensor` but is {type(a_ )}''' ) if isinstance(a_ , PIL.Image.Image ): __snake_case = preprocess(a_ ) __snake_case , __snake_case = image.shape[-2:] # in_channels should be 6: 3 for latents, 3 for low resolution image __snake_case = (batch_size, self.unet.config.in_channels // 2, height, width) __snake_case = next(self.unet.parameters() ).dtype __snake_case = randn_tensor(a_ , generator=a_ , device=self.device , dtype=a_ ) __snake_case = image.to(device=self.device , dtype=a_ ) # set timesteps and move to the correct device self.scheduler.set_timesteps(a_ , device=self.device ) __snake_case = self.scheduler.timesteps # scale the initial noise by the standard deviation required by the scheduler __snake_case = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature. # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] __snake_case = "eta" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) __snake_case = {} if accepts_eta: __snake_case = eta for t in self.progress_bar(a_ ): # concat latents and low resolution image in the channel dimension. __snake_case = torch.cat([latents, image] , dim=1 ) __snake_case = self.scheduler.scale_model_input(a_ , a_ ) # predict the noise residual __snake_case = self.unet(a_ , a_ ).sample # compute the previous noisy sample x_t -> x_t-1 __snake_case = self.scheduler.step(a_ , a_ , a_ , **a_ ).prev_sample # decode the image latents with the VQVAE __snake_case = self.vqvae.decode(a_ ).sample __snake_case = torch.clamp(a_ , -1.0 , 1.0 ) __snake_case = image / 2 + 0.5 __snake_case = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __snake_case = self.numpy_to_pil(a_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=a_ )
69
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : List[Any] = { '''configuration_table_transformer''': [ '''TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''TableTransformerConfig''', '''TableTransformerOnnxConfig''', ] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Tuple = [ '''TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TableTransformerForObjectDetection''', '''TableTransformerModel''', '''TableTransformerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TableTransformerConfig, TableTransformerOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_table_transformer import ( TABLE_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TableTransformerForObjectDetection, TableTransformerModel, TableTransformerPreTrainedModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
1
'''simple docstring''' import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, BatchEncoding, MBartaaTokenizer, MBartaaTokenizerFast, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from ...test_tokenization_common import TokenizerTesterMixin a : List[str] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mbart.modeling_mbart import shift_tokens_right a : Optional[Any] = 250_004 a : List[Any] = 250_020 @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = MBartaaTokenizer __SCREAMING_SNAKE_CASE = MBartaaTokenizerFast __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = True def A ( self : Dict ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = MBartaaTokenizer(a_ , src_lang="en_XX" , tgt_lang="ro_RO" , keep_accents=a_ ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : Tuple ): """simple docstring""" __snake_case = "<s>" __snake_case = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : int ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "<mask>" ) self.assertEqual(len(a_ ) , 1_054 ) def A ( self : Tuple ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 1_054 ) def A ( self : Any ): """simple docstring""" __snake_case = MBartaaTokenizer(a_ , src_lang="en_XX" , tgt_lang="ro_RO" , keep_accents=a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(a_ ) , [value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] , ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( a_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", "."] , ) __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] ] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) self.assertListEqual( a_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", "."] , ) @slow def A ( self : List[str] ): """simple docstring""" __snake_case = {"input_ids": [[250_004, 11_062, 82_772, 7, 15, 82_772, 538, 51_529, 237, 17_198, 1_290, 206, 9, 215_175, 1_314, 136, 17_198, 1_290, 206, 9, 56_359, 42, 122_009, 9, 16_466, 16, 87_344, 4_537, 9, 4_717, 78_381, 6, 159_958, 7, 15, 24_480, 618, 4, 527, 22_693, 5_428, 4, 2_777, 24_480, 9_874, 4, 43_523, 594, 4, 803, 18_392, 33_189, 18, 4, 43_523, 24_447, 12_399, 100, 24_955, 83_658, 9_626, 144_057, 15, 839, 22_335, 16, 136, 24_955, 83_658, 83_479, 15, 39_102, 724, 16, 678, 645, 2_789, 1_328, 4_589, 42, 122_009, 115_774, 23, 805, 1_328, 46_876, 7, 136, 53_894, 1_940, 42_227, 41_159, 17_721, 823, 425, 4, 27_512, 98_722, 206, 136, 5_531, 4_970, 919, 17_336, 5, 2], [250_004, 20_080, 618, 83, 82_775, 47, 479, 9, 1_517, 73, 53_894, 333, 80_581, 110_117, 18_811, 5_256, 1_295, 51, 152_526, 297, 7_986, 390, 124_416, 538, 35_431, 214, 98, 15_044, 25_737, 136, 7_108, 43_701, 23, 756, 135_355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [250_004, 581, 63_773, 119_455, 6, 147_797, 88_203, 7, 645, 70, 21, 3_285, 10_269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="facebook/mbart-large-50" , revision="d3913889c59cd5c9e456b269c376325eabad57e2" , ) def A ( self : Tuple ): """simple docstring""" if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return __snake_case = (self.rust_tokenizer_class, "hf-internal-testing/tiny-random-mbart50", {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(f'''{tokenizer.__class__.__name__} ({pretrained_name})''' ): __snake_case = self.rust_tokenizer_class.from_pretrained(a_ , **a_ ) __snake_case = self.tokenizer_class.from_pretrained(a_ , **a_ ) __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a_ ) __snake_case = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) __snake_case = tuple(f for f in tokenizer_r_files if "tokenizer.json" not in f ) self.assertSequenceEqual(a_ , a_ ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a_ ) __snake_case = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_ , a_ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=True __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a_ , legacy_format=a_ ) __snake_case = tokenizer_p.save_pretrained(a_ ) # Checks it save with the same files self.assertSequenceEqual(a_ , a_ ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a_ ) __snake_case = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_ , a_ ) ) shutil.rmtree(a_ ) # Save tokenizer rust, legacy_format=False __snake_case = tempfile.mkdtemp() __snake_case = tokenizer_r.save_pretrained(a_ , legacy_format=a_ ) __snake_case = tokenizer_p.save_pretrained(a_ ) # Checks it saved the tokenizer.json file self.assertTrue(any("tokenizer.json" in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way __snake_case = tokenizer_r.from_pretrained(a_ ) __snake_case = tokenizer_p.from_pretrained(a_ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(a_ , a_ ) ) shutil.rmtree(a_ ) @require_torch @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = """facebook/mbart-large-50-one-to-many-mmt""" __SCREAMING_SNAKE_CASE = [ """ UN Chief Says There Is No Military Solution in Syria""", """ Secretary-General Ban Ki-moon says his response to Russia's stepped up military support for Syria is that \"there is no military solution\" to the nearly five-year conflict and more weapons will only worsen the violence and misery for millions of people.""", ] __SCREAMING_SNAKE_CASE = [ """Şeful ONU declară că nu există o soluţie militară în Siria""", """Secretarul General Ban Ki-moon declară că răspunsul său la intensificarea sprijinului militar al Rusiei""" """ pentru Siria este că \"nu există o soluţie militară\" la conflictul de aproape cinci ani şi că noi arme nu vor""" """ face decât să înrăutăţească violenţele şi mizeria pentru milioane de oameni.""", ] __SCREAMING_SNAKE_CASE = [EN_CODE, 8274, 127873, 25916, 7, 8622, 2071, 438, 67485, 53, 187895, 23, 51712, 2] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = MBartaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang="en_XX" , tgt_lang="ro_RO" ) __snake_case = 1 return cls def A ( self : Tuple ): """simple docstring""" self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ar_AR"] , 250_001 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["en_EN"] , 250_004 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["ro_RO"] , 250_020 ) self.assertEqual(self.tokenizer.fairseq_tokens_to_ids["mr_IN"] , 250_038 ) def A ( self : str ): """simple docstring""" __snake_case = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , a_ ) def A ( self : Dict ): """simple docstring""" self.assertIn(a_ , self.tokenizer.all_special_ids ) __snake_case = [RO_CODE, 884, 9_019, 96, 9, 916, 86_792, 36, 18_743, 15_596, 5, 2] __snake_case = self.tokenizer.decode(a_ , skip_special_tokens=a_ ) __snake_case = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=a_ ) self.assertEqual(a_ , a_ ) self.assertNotIn(self.tokenizer.eos_token , a_ ) def A ( self : str ): """simple docstring""" __snake_case = ["this is gunna be a long sentence " * 20] assert isinstance(src_text[0] , a_ ) __snake_case = 10 __snake_case = self.tokenizer(a_ , max_length=a_ , truncation=a_ ).input_ids[0] self.assertEqual(ids[0] , a_ ) self.assertEqual(ids[-1] , 2 ) self.assertEqual(len(a_ ) , a_ ) def A ( self : str ): """simple docstring""" self.assertListEqual(self.tokenizer.convert_tokens_to_ids(["<mask>", "ar_AR"] ) , [250_053, 250_001] ) def A ( self : Dict ): """simple docstring""" __snake_case = tempfile.mkdtemp() __snake_case = self.tokenizer.fairseq_tokens_to_ids self.tokenizer.save_pretrained(a_ ) __snake_case = MBartaaTokenizer.from_pretrained(a_ ) self.assertDictEqual(new_tok.fairseq_tokens_to_ids , a_ ) @require_torch def A ( self : Dict ): """simple docstring""" __snake_case = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=a_ , return_tensors="pt" ) __snake_case = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == RO_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2].tolist() == [2, RO_CODE] @require_torch def A ( self : int ): """simple docstring""" __snake_case = self.tokenizer( self.src_text , text_target=self.tgt_text , padding=a_ , truncation=a_ , max_length=len(self.expected_src_tokens ) , return_tensors="pt" , ) __snake_case = shift_tokens_right(batch["labels"] , self.tokenizer.pad_token_id ) self.assertIsInstance(a_ , a_ ) self.assertEqual((2, 14) , batch.input_ids.shape ) self.assertEqual((2, 14) , batch.attention_mask.shape ) __snake_case = batch.input_ids.tolist()[0] self.assertListEqual(self.expected_src_tokens , a_ ) self.assertEqual(2 , batch.decoder_input_ids[0, 0] ) # decoder_start_token_id # Test that special tokens are reset self.assertEqual(self.tokenizer.prefix_tokens , [EN_CODE] ) self.assertEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) def A ( self : Dict ): """simple docstring""" __snake_case = self.tokenizer(self.src_text , padding=a_ , truncation=a_ , max_length=3 , return_tensors="pt" ) __snake_case = self.tokenizer( text_target=self.tgt_text , padding=a_ , truncation=a_ , max_length=10 , return_tensors="pt" ) __snake_case = targets["input_ids"] __snake_case = shift_tokens_right(a_ , self.tokenizer.pad_token_id ) self.assertEqual(batch.input_ids.shape[1] , 3 ) self.assertEqual(batch.decoder_input_ids.shape[1] , 10 ) @require_torch def A ( self : Any ): """simple docstring""" __snake_case = self.tokenizer._build_translation_inputs( "A test" , return_tensors="pt" , src_lang="en_XX" , tgt_lang="ar_AR" ) self.assertEqual( nested_simplify(a_ ) , { # en_XX, A, test, EOS "input_ids": [[250_004, 62, 3_034, 2]], "attention_mask": [[1, 1, 1, 1]], # ar_AR "forced_bos_token_id": 250_001, } , )
69
'''simple docstring''' import json import os import torch from diffusers import UNetaDModel os.makedirs('''hub/hopper-medium-v2/unet/hor32''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/unet/hor128''', exist_ok=True) os.makedirs('''hub/hopper-medium-v2/value_function''', exist_ok=True) def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: if hor == 1_28: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D") elif hor == 32: __snake_case = ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D") __snake_case = (32, 64, 1_28, 2_56) __snake_case = ("UpResnetBlock1D", "UpResnetBlock1D", "UpResnetBlock1D") __snake_case = torch.load(F'''/Users/bglickenhaus/Documents/diffuser/temporal_unet-hopper-mediumv2-hor{hor}.torch''' ) __snake_case = model.state_dict() __snake_case = { "down_block_types": down_block_types, "block_out_channels": block_out_channels, "up_block_types": up_block_types, "layers_per_block": 1, "use_timestep_embedding": True, "out_block_type": "OutConv1DBlock", "norm_num_groups": 8, "downsample_each_block": False, "in_channels": 14, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "flip_sin_to_cos": False, "freq_shift": 1, "sample_size": 6_55_36, "mid_block_type": "MidResTemporalBlock1D", "act_fn": "mish", } __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(model.state_dict().keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , F'''hub/hopper-medium-v2/unet/hor{hor}/diffusion_pytorch_model.bin''' ) with open(F'''hub/hopper-medium-v2/unet/hor{hor}/config.json''' , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( ) -> List[Any]: __snake_case = { "in_channels": 14, "down_block_types": ("DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D", "DownResnetBlock1D"), "up_block_types": (), "out_block_type": "ValueFunction", "mid_block_type": "ValueFunctionMidBlock1D", "block_out_channels": (32, 64, 1_28, 2_56), "layers_per_block": 1, "downsample_each_block": True, "sample_size": 6_55_36, "out_channels": 14, "extra_in_channels": 0, "time_embedding_type": "positional", "use_timestep_embedding": True, "flip_sin_to_cos": False, "freq_shift": 1, "norm_num_groups": 8, "act_fn": "mish", } __snake_case = torch.load("/Users/bglickenhaus/Documents/diffuser/value_function-hopper-mediumv2-hor32.torch" ) __snake_case = model __snake_case = UNetaDModel(**_UpperCAmelCase ) print(F'''length of state dict: {len(state_dict.keys() )}''' ) print(F'''length of value function dict: {len(hf_value_function.state_dict().keys() )}''' ) __snake_case = dict(zip(state_dict.keys() , hf_value_function.state_dict().keys() ) ) for k, v in mapping.items(): __snake_case = state_dict.pop(_UpperCAmelCase ) hf_value_function.load_state_dict(_UpperCAmelCase ) torch.save(hf_value_function.state_dict() , "hub/hopper-medium-v2/value_function/diffusion_pytorch_model.bin" ) with open("hub/hopper-medium-v2/value_function/config.json" , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase ) if __name__ == "__main__": unet(32) # unet(128) value_function()
69
1
'''simple docstring''' from __future__ import annotations import string from itertools import cycle, product from pathlib import Path a : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) a : list[int] = [ord(letter) for letter in string.ascii_lowercase] a : set[int] = {ord(char) for char in VALID_CHARS} a : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def __UpperCAmelCase ( _UpperCAmelCase : list[int] , _UpperCAmelCase : tuple[int, ...] ) -> str | None: __snake_case = "" __snake_case = 42 __snake_case = 42 __snake_case = 42 for keychar, cipherchar in zip(cycle(_UpperCAmelCase ) , _UpperCAmelCase ): __snake_case = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(_UpperCAmelCase ) return decoded def __UpperCAmelCase ( _UpperCAmelCase : list[int] ) -> list[str]: __snake_case = [] for key in product(_UpperCAmelCase , repeat=3 ): __snake_case = try_key(_UpperCAmelCase , _UpperCAmelCase ) if encoded is not None: possibles.append(_UpperCAmelCase ) return possibles def __UpperCAmelCase ( _UpperCAmelCase : list[str] , _UpperCAmelCase : str ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def __UpperCAmelCase ( _UpperCAmelCase : str = "p059_cipher.txt" ) -> int: __snake_case = 42 __snake_case = 42 __snake_case = 42 __snake_case = 42 __snake_case = Path(_UpperCAmelCase ).parent.joinpath(_UpperCAmelCase ).read_text(encoding="utf-8" ) __snake_case = [int(_UpperCAmelCase ) for number in data.strip().split("," )] __snake_case = filter_valid_chars(_UpperCAmelCase ) for common_word in COMMON_WORDS: __snake_case = filter_common_word(_UpperCAmelCase , _UpperCAmelCase ) if len(_UpperCAmelCase ) == 1: break __snake_case = possibles[0] return sum(ord(_UpperCAmelCase ) for char in decoded_text ) if __name__ == "__main__": print(F'''{solution() = }''')
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1_00_00_00 ) -> int: __snake_case = 1 __snake_case = 1 __snake_case = {1: 1} for inputa in range(2 , _UpperCAmelCase ): __snake_case = 0 __snake_case = inputa while True: if number in counters: counter += counters[number] break if number % 2 == 0: number //= 2 counter += 1 else: __snake_case = (3 * number) + 1 counter += 1 if inputa not in counters: __snake_case = counter if counter > pre_counter: __snake_case = inputa __snake_case = counter return largest_number if __name__ == "__main__": print(solution(int(input().strip())))
69
1
'''simple docstring''' import numpy class SCREAMING_SNAKE_CASE__ : def __init__( self : Any , a_ : numpy.ndarray , a_ : numpy.ndarray ): """simple docstring""" __snake_case = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. __snake_case = numpy.random.rand( self.input_array.shape[1] , 4 ) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. __snake_case = numpy.random.rand( 4 , 3 ) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. __snake_case = numpy.random.rand(3 , 1 ) # Real output values provided. __snake_case = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. __snake_case = numpy.zeros(output_array.shape ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights ) ) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. __snake_case = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. __snake_case = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return self.layer_between_second_hidden_layer_and_output def A ( self : Optional[Any] ): """simple docstring""" __snake_case = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , ) __snake_case = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , ) __snake_case = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output ) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer ) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer ) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def A ( self : Union[str, Any] , a_ : numpy.ndarray , a_ : int , a_ : bool ): """simple docstring""" for iteration in range(1 , iterations + 1 ): __snake_case = self.feedforward() self.back_propagation() if give_loss: __snake_case = numpy.mean(numpy.square(output - self.feedforward() ) ) print(f'''Iteration {iteration} Loss: {loss}''' ) def A ( self : Optional[Any] , a_ : numpy.ndarray ): """simple docstring""" __snake_case = input_arr __snake_case = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights ) ) __snake_case = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , ) ) __snake_case = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , ) ) return int(self.layer_between_second_hidden_layer_and_output > 0.6 ) def __UpperCAmelCase ( _UpperCAmelCase : numpy.ndarray ) -> numpy.ndarray: return 1 / (1 + numpy.exp(-value )) def __UpperCAmelCase ( _UpperCAmelCase : numpy.ndarray ) -> numpy.ndarray: return (value) * (1 - (value)) def __UpperCAmelCase ( ) -> int: __snake_case = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. __snake_case = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. __snake_case = TwoHiddenLayerNeuralNetwork( input_array=_UpperCAmelCase , output_array=_UpperCAmelCase ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=_UpperCAmelCase , iterations=10 , give_loss=_UpperCAmelCase ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
69
'''simple docstring''' from ...processing_utils import ProcessorMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """SpeechT5FeatureExtractor""" __SCREAMING_SNAKE_CASE = """SpeechT5Tokenizer""" def __init__( self : List[Any] , a_ : str , a_ : str ): """simple docstring""" super().__init__(a_ , a_ ) def __call__( self : Dict , *a_ : Tuple , **a_ : List[str] ): """simple docstring""" __snake_case = kwargs.pop("audio" , a_ ) __snake_case = kwargs.pop("text" , a_ ) __snake_case = kwargs.pop("text_target" , a_ ) __snake_case = kwargs.pop("audio_target" , a_ ) __snake_case = kwargs.pop("sampling_rate" , a_ ) if audio is not None and text is not None: raise ValueError( "Cannot process both `audio` and `text` inputs. Did you mean `audio_target` or `text_target`?" ) if audio_target is not None and text_target is not None: raise ValueError( "Cannot process both `audio_target` and `text_target` inputs. Did you mean `audio` or `text`?" ) if audio is None and audio_target is None and text is None and text_target is None: raise ValueError( "You need to specify either an `audio`, `audio_target`, `text`, or `text_target` input to process." ) if audio is not None: __snake_case = self.feature_extractor(a_ , *a_ , sampling_rate=a_ , **a_ ) elif text is not None: __snake_case = self.tokenizer(a_ , **a_ ) else: __snake_case = None if audio_target is not None: __snake_case = self.feature_extractor(audio_target=a_ , *a_ , sampling_rate=a_ , **a_ ) __snake_case = targets["input_values"] elif text_target is not None: __snake_case = self.tokenizer(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : str , **a_ : Dict ): """simple docstring""" __snake_case = kwargs.pop("input_values" , a_ ) __snake_case = kwargs.pop("input_ids" , a_ ) __snake_case = kwargs.pop("labels" , a_ ) if input_values is not None and input_ids is not None: raise ValueError("Cannot process both `input_values` and `input_ids` inputs." ) if input_values is None and input_ids is None and labels is None: raise ValueError( "You need to specify either an `input_values`, `input_ids`, or `labels` input to be padded." ) if input_values is not None: __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) elif input_ids is not None: __snake_case = self.tokenizer.pad(a_ , **a_ ) else: __snake_case = None if labels is not None: if "input_ids" in labels or (isinstance(a_ , a_ ) and "input_ids" in labels[0]): __snake_case = self.tokenizer.pad(a_ , **a_ ) __snake_case = targets["input_ids"] else: __snake_case = self.feature_extractor.feature_size __snake_case = self.feature_extractor.num_mel_bins __snake_case = self.feature_extractor.pad(a_ , *a_ , **a_ ) __snake_case = feature_size_hack __snake_case = targets["input_values"] else: __snake_case = None if inputs is None: return targets if targets is not None: __snake_case = labels __snake_case = targets.get("attention_mask" ) if decoder_attention_mask is not None: __snake_case = decoder_attention_mask return inputs def A ( self : List[str] , *a_ : Any , **a_ : List[str] ): """simple docstring""" return self.tokenizer.batch_decode(*a_ , **a_ ) def A ( self : Optional[int] , *a_ : Union[str, Any] , **a_ : str ): """simple docstring""" return self.tokenizer.decode(*a_ , **a_ )
69
1
'''simple docstring''' from __future__ import annotations def __UpperCAmelCase ( _UpperCAmelCase : list[float] , _UpperCAmelCase : int ) -> List[Any]: print(F'''Vertex\tShortest Distance from vertex {src}''' ) for i, d in enumerate(_UpperCAmelCase ): print(F'''{i}\t\t{d}''' ) def __UpperCAmelCase ( _UpperCAmelCase : list[dict[str, int]] , _UpperCAmelCase : list[float] , _UpperCAmelCase : int ) -> Union[str, Any]: for j in range(_UpperCAmelCase ): __snake_case , __snake_case , __snake_case = (graph[j][k] for k in ["src", "dst", "weight"]) if distance[u] != float("inf" ) and distance[u] + w < distance[v]: return True return False def __UpperCAmelCase ( _UpperCAmelCase : list[dict[str, int]] , _UpperCAmelCase : int , _UpperCAmelCase : int , _UpperCAmelCase : int ) -> list[float]: __snake_case = [float("inf" )] * vertex_count __snake_case = 0.0 for _ in range(vertex_count - 1 ): for j in range(_UpperCAmelCase ): __snake_case , __snake_case , __snake_case = (graph[j][k] for k in ["src", "dst", "weight"]) if distance[u] != float("inf" ) and distance[u] + w < distance[v]: __snake_case = distance[u] + w __snake_case = check_negative_cycle(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) if negative_cycle_exists: raise Exception("Negative cycle found" ) return distance if __name__ == "__main__": import doctest doctest.testmod() a : int = int(input('''Enter number of vertices: ''').strip()) a : List[str] = int(input('''Enter number of edges: ''').strip()) a : list[dict[str, int]] = [{} for _ in range(E)] for i in range(E): print('''Edge ''', i + 1) a , a , a : str = ( int(x) for x in input('''Enter source, destination, weight: ''').strip().split(''' ''') ) a : List[str] = {'''src''': src, '''dst''': dest, '''weight''': weight} a : int = int(input('''\nEnter shortest path source:''').strip()) a : int = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
69
'''simple docstring''' import re from pathlib import Path from unittest import TestCase import pytest @pytest.mark.integration class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def A ( self : Optional[Any] , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"(?!.*\b(?:encoding|rb|w|wb|w+|wb+|ab|ab+)\b)(?<=\s)(open)\((.*)\)" ) __snake_case = input_file.read() __snake_case = regexp.search(a_ ) return match def A ( self : Any , a_ : str ): """simple docstring""" with open(a_ , encoding="utf-8" ) as input_file: __snake_case = re.compile(r"#[^\r\n]*print\(|\"[^\r\n]*print\(|\"\"\".*?print\(.*?\"\"\"|(print\()" , re.DOTALL ) __snake_case = input_file.read() # use `re.finditer` to handle the case where the ignored groups would be matched first by `re.search` __snake_case = regexp.finditer(a_ ) __snake_case = [match for match in matches if match is not None and match.group(1 ) is not None] return matches[0] if matches else None def A ( self : Optional[int] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_encoding_on_file_open(str(a_ ) ): raise AssertionError(f'''open(...) must use utf-8 encoding in {dataset}''' ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Path("./datasets" ) __snake_case = list(dataset_paths.absolute().glob("**/*.py" ) ) for dataset in dataset_files: if self._no_print_statements(str(a_ ) ): raise AssertionError(f'''print statement found in {dataset}. Use datasets.logger/logging instead.''' )
69
1
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging a : Optional[Any] = logging.get_logger(__name__) a : Dict = {'''vocab_file''': '''sentencepiece.model'''} a : Tuple = { '''vocab_file''': { '''google/rembert''': '''https://huggingface.co/google/rembert/resolve/main/sentencepiece.model''', }, } a : str = { '''google/rembert''': 256, } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = VOCAB_FILES_NAMES __SCREAMING_SNAKE_CASE = PRETRAINED_VOCAB_FILES_MAP __SCREAMING_SNAKE_CASE = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES def __init__( self : Optional[Any] , a_ : int , a_ : Any=False , a_ : List[Any]=True , a_ : List[Any]=True , a_ : List[Any]="[CLS]" , a_ : List[Any]="[SEP]" , a_ : List[Any]="[UNK]" , a_ : str="[SEP]" , a_ : List[str]="[PAD]" , a_ : Optional[int]="[CLS]" , a_ : List[str]="[MASK]" , **a_ : str , ): """simple docstring""" super().__init__( do_lower_case=a_ , remove_space=a_ , keep_accents=a_ , bos_token=a_ , eos_token=a_ , unk_token=a_ , sep_token=a_ , pad_token=a_ , cls_token=a_ , mask_token=a_ , **a_ , ) __snake_case = do_lower_case __snake_case = remove_space __snake_case = keep_accents __snake_case = vocab_file __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(a_ ) @property def A ( self : Optional[Any] ): """simple docstring""" return len(self.sp_model ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = {self.convert_ids_to_tokens(a_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Dict ): """simple docstring""" __snake_case = self.__dict__.copy() __snake_case = None return state def __setstate__( self : str , a_ : Optional[int] ): """simple docstring""" __snake_case = d __snake_case = spm.SentencePieceProcessor() self.sp_model.Load(self.vocab_file ) def A ( self : Tuple , a_ : Optional[int] , a_ : int=False ): """simple docstring""" __snake_case = self.sp_model.EncodeAsPieces(a_ ) return pieces def A ( self : Any , a_ : Optional[Any] ): """simple docstring""" return self.sp_model.PieceToId(a_ ) def A ( self : Optional[Any] , a_ : List[str] ): """simple docstring""" return self.sp_model.IdToPiece(a_ ) def A ( self : Optional[Any] , a_ : int ): """simple docstring""" __snake_case = self.sp_model.decode_pieces(a_ ) return out_string def A ( self : Union[str, Any] , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def A ( self : List[str] , a_ : List[int] , a_ : Optional[List[int]] = None , a_ : bool = False ): """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(a_ )) + [1] + ([0] * len(a_ )) + [1] return [1] + ([0] * len(a_ )) + [1] def A ( self : Tuple , a_ : List[int] , a_ : Optional[List[int]] = None ): """simple docstring""" __snake_case = [self.sep_token_id] __snake_case = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def A ( self : List[Any] , a_ : str , a_ : Optional[str] = None ): """simple docstring""" if not os.path.isdir(a_ ): logger.error("Vocabulary path ({}) should be a directory".format(a_ ) ) return __snake_case = os.path.join( a_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(a_ ): copyfile(self.vocab_file , a_ ) return (out_vocab_file,)
69
1
'''simple docstring''' from __future__ import annotations from collections.abc import Callable from typing import Generic, TypeVar a : Union[str, Any] = TypeVar('''T''') a : Union[str, Any] = TypeVar('''U''') class SCREAMING_SNAKE_CASE__ ( Generic[T, U] ): def __init__( self : Optional[Any] , a_ : T | None , a_ : U | None ): """simple docstring""" __snake_case = key __snake_case = val __snake_case = None __snake_case = None def __repr__( self : Dict ): """simple docstring""" return ( f'''Node: key: {self.key}, val: {self.val}, ''' f'''has next: {bool(self.next )}, has prev: {bool(self.prev )}''' ) class SCREAMING_SNAKE_CASE__ ( Generic[T, U] ): def __init__( self : Optional[int] ): """simple docstring""" __snake_case = DoubleLinkedListNode(a_ , a_ ) __snake_case = DoubleLinkedListNode(a_ , a_ ) __snake_case , __snake_case = self.rear, self.head def __repr__( self : Any ): """simple docstring""" __snake_case = ["DoubleLinkedList"] __snake_case = self.head while node.next is not None: rep.append(str(a_ ) ) __snake_case = node.next rep.append(str(self.rear ) ) return ",\n ".join(a_ ) def A ( self : int , a_ : DoubleLinkedListNode[T, U] ): """simple docstring""" __snake_case = self.rear.prev # All nodes other than self.head are guaranteed to have non-None previous assert previous is not None __snake_case = node __snake_case = previous __snake_case = node __snake_case = self.rear def A ( self : Optional[Any] , a_ : DoubleLinkedListNode[T, U] ): """simple docstring""" if node.prev is None or node.next is None: return None __snake_case = node.next __snake_case = node.prev __snake_case = None __snake_case = None return node class SCREAMING_SNAKE_CASE__ ( Generic[T, U] ): __SCREAMING_SNAKE_CASE = {} def __init__( self : Any , a_ : int ): """simple docstring""" __snake_case = DoubleLinkedList() __snake_case = capacity __snake_case = 0 __snake_case = 0 __snake_case = 0 __snake_case = {} def __repr__( self : Union[str, Any] ): """simple docstring""" return ( f'''CacheInfo(hits={self.hits}, misses={self.miss}, ''' f'''capacity={self.capacity}, current size={self.num_keys})''' ) def __contains__( self : Optional[int] , a_ : T ): """simple docstring""" return key in self.cache def A ( self : int , a_ : T ): """simple docstring""" if key in self.cache: self.hits += 1 __snake_case = self.cache[key] __snake_case = self.list.remove(self.cache[key] ) assert node == value_node # node is guaranteed not None because it is in self.cache assert node is not None self.list.add(a_ ) return node.val self.miss += 1 return None def A ( self : Optional[int] , a_ : T , a_ : U ): """simple docstring""" if key not in self.cache: if self.num_keys >= self.capacity: # delete first node (oldest) when over capacity __snake_case = self.list.head.next # guaranteed to have a non-None first node when num_keys > 0 # explain to type checker via assertions assert first_node is not None assert first_node.key is not None assert ( self.list.remove(a_ ) is not None ) # node guaranteed to be in list assert node.key is not None del self.cache[first_node.key] self.num_keys -= 1 __snake_case = DoubleLinkedListNode(a_ , a_ ) self.list.add(self.cache[key] ) self.num_keys += 1 else: # bump node to the end of the list, update value __snake_case = self.list.remove(self.cache[key] ) assert node is not None # node guaranteed to be in list __snake_case = value self.list.add(a_ ) @classmethod def A ( cls : Dict , a_ : int = 128 ): """simple docstring""" def cache_decorator_inner(a_ : Callable[[T], U] ) -> Callable[..., U]: def cache_decorator_wrapper(*a_ : T ) -> U: if func not in cls.decorator_function_to_instance_map: __snake_case = LRUCache(a_ ) __snake_case = cls.decorator_function_to_instance_map[func].get(args[0] ) if result is None: __snake_case = func(*a_ ) cls.decorator_function_to_instance_map[func].put(args[0] , a_ ) return result def cache_info() -> LRUCache[T, U]: return cls.decorator_function_to_instance_map[func] setattr(a_ , "cache_info" , a_ ) # noqa: B010 return cache_decorator_wrapper return cache_decorator_inner if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' import os import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from huggingface_hub.file_download import http_get from requests.exceptions import HTTPError from transformers import ( AlbertTokenizer, AutoTokenizer, BertTokenizer, BertTokenizerFast, GPTaTokenizerFast, is_tokenizers_available, ) from transformers.testing_utils import TOKEN, USER, is_staging_test, require_tokenizers from transformers.tokenization_utils import Trie sys.path.append(str(Path(__file__).parent.parent / '''utils''')) from test_module.custom_tokenization import CustomTokenizer # noqa E402 if is_tokenizers_available(): from test_module.custom_tokenization_fast import CustomTokenizerFast class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = BertTokenizer.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() @require_tokenizers def A ( self : Optional[Any] ): """simple docstring""" __snake_case = mock.Mock() __snake_case = 500 __snake_case = {} __snake_case = HTTPError __snake_case = {} # Download this model to make sure it's in the cache. __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # Under the mock environment we get a 500 error when trying to reach the tokenizer. with mock.patch("requests.Session.request" , return_value=a_ ) as mock_head: __snake_case = GPTaTokenizerFast.from_pretrained("gpt2" ) # This check we did call the fake head request mock_head.assert_called() def A ( self : Optional[Any] ): """simple docstring""" try: __snake_case = tempfile.mktemp() with open(a_ , "wb" ) as f: http_get("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" , a_ ) __snake_case = AlbertTokenizer.from_pretrained(a_ ) finally: os.remove(a_ ) # Supporting this legacy load introduced a weird bug where the tokenizer would load local files if they are in # the current folder and have the right name. if os.path.isfile("tokenizer.json" ): # We skip the test if the user has a `tokenizer.json` in this folder to avoid deleting it. return try: with open("tokenizer.json" , "wb" ) as f: http_get("https://huggingface.co/hf-internal-testing/tiny-random-bert/blob/main/tokenizer.json" , a_ ) __snake_case = AutoTokenizer.from_pretrained("hf-internal-testing/tiny-random-gpt2" ) # The tiny random BERT has a vocab size of 1024, tiny gpt2 as a vocab size of 1000 self.assertEqual(tokenizer.vocab_size , 1_000 ) # Tokenizer should depend on the remote checkpoint, not the local tokenizer.json file. finally: os.remove("tokenizer.json" ) def A ( self : str ): """simple docstring""" __snake_case = AlbertTokenizer.from_pretrained("https://huggingface.co/albert-base-v1/resolve/main/spiece.model" ) @is_staging_test class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): __SCREAMING_SNAKE_CASE = ["""[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """bla""", """blou"""] @classmethod def A ( cls : List[Any] ): """simple docstring""" __snake_case = TOKEN HfFolder.save_token(a_ ) @classmethod def A ( cls : List[Any] ): """simple docstring""" try: delete_repo(token=cls._token , repo_id="test-tokenizer" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-tokenizer-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-tokenizer" ) except HTTPError: pass def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("test-tokenizer" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="test-tokenizer" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained(a_ , repo_id="test-tokenizer" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained(f'''{USER}/test-tokenizer''' ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) def A ( self : int ): """simple docstring""" with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizer(a_ ) tokenizer.push_to_hub("valid_org/test-tokenizer-org" , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-tokenizer-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: tokenizer.save_pretrained( a_ , repo_id="valid_org/test-tokenizer-org" , push_to_hub=a_ , use_auth_token=self._token ) __snake_case = BertTokenizer.from_pretrained("valid_org/test-tokenizer-org" ) self.assertDictEqual(new_tokenizer.vocab , tokenizer.vocab ) @require_tokenizers def A ( self : List[str] ): """simple docstring""" CustomTokenizer.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = CustomTokenizer(a_ ) # No fast custom tokenizer tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the CustomTokenizer class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) # Fast and slow custom tokenizer CustomTokenizerFast.register_for_auto_class() with tempfile.TemporaryDirectory() as tmp_dir: __snake_case = os.path.join(a_ , "vocab.txt" ) with open(a_ , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in self.vocab_tokens] ) ) __snake_case = BertTokenizerFast.from_pretrained(a_ ) bert_tokenizer.save_pretrained(a_ ) __snake_case = CustomTokenizerFast.from_pretrained(a_ ) tokenizer.push_to_hub("test-dynamic-tokenizer" , use_auth_token=self._token ) __snake_case = AutoTokenizer.from_pretrained(f'''{USER}/test-dynamic-tokenizer''' , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizerFast" ) __snake_case = AutoTokenizer.from_pretrained( f'''{USER}/test-dynamic-tokenizer''' , use_fast=a_ , trust_remote_code=a_ ) # Can't make an isinstance check because the new_model.config is from the FakeConfig class of a dynamic module self.assertEqual(tokenizer.__class__.__name__ , "CustomTokenizer" ) class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("Hello 友達" ) self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {" ": {"友": {"達": {"": 1}}}}}}}}} ) trie.add("Hello" ) trie.data self.assertEqual(trie.data , {"H": {"e": {"l": {"l": {"o": {"": 1, " ": {"友": {"達": {"": 1}}}}}}}}} ) def A ( self : str ): """simple docstring""" __snake_case = Trie() self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS] This is a extra_id_100"] ) trie.add("[CLS]" ) trie.add("extra_id_1" ) trie.add("extra_id_100" ) self.assertEqual(trie.split("[CLS] This is a extra_id_100" ) , ["[CLS]", " This is a ", "extra_id_100"] ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = Trie() trie.add("A" ) self.assertEqual(trie.split("ABC" ) , ["A", "BC"] ) self.assertEqual(trie.split("BCA" ) , ["BC", "A"] ) def A ( self : List[Any] ): """simple docstring""" __snake_case = Trie() trie.add("TOKEN]" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : str ): """simple docstring""" __snake_case = Trie() trie.add("A" ) trie.add("P" ) trie.add("[SPECIAL_TOKEN]" ) self.assertEqual(trie.split("This is something [SPECIAL_TOKEN]" ) , ["This is something ", "[SPECIAL_TOKEN]"] ) def A ( self : Optional[int] ): """simple docstring""" __snake_case = Trie() trie.add("AB" ) trie.add("B" ) trie.add("C" ) self.assertEqual(trie.split("ABC" ) , ["AB", "C"] ) def A ( self : Tuple ): """simple docstring""" __snake_case = Trie() trie.add("ABC" ) trie.add("B" ) trie.add("CD" ) self.assertEqual(trie.split("ABCD" ) , ["ABC", "D"] ) def A ( self : Any ): """simple docstring""" __snake_case = Trie() __snake_case = trie.cut_text("ABC" , [0, 0, 2, 1, 2, 3] ) self.assertEqual(a_ , ["AB", "C"] )
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a : Any = { '''configuration_git''': ['''GIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''GitConfig''', '''GitVisionConfig'''], '''processing_git''': ['''GitProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a : Any = [ '''GIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''GitForCausalLM''', '''GitModel''', '''GitPreTrainedModel''', '''GitVisionModel''', ] if TYPE_CHECKING: from .configuration_git import GIT_PRETRAINED_CONFIG_ARCHIVE_MAP, GitConfig, GitVisionConfig from .processing_git import GitProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_git import ( GIT_PRETRAINED_MODEL_ARCHIVE_LIST, GitForCausalLM, GitModel, GitPreTrainedModel, GitVisionModel, ) else: import sys a : List[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> int: assert ( isinstance(_UpperCAmelCase , _UpperCAmelCase ) and number_of_steps > 0 ), F'''number_of_steps needs to be positive integer, your input {number_of_steps}''' if number_of_steps == 1: return 1 __snake_case , __snake_case = 1, 1 for _ in range(number_of_steps - 1 ): __snake_case , __snake_case = current + previous, current return current if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import json import os import unittest from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES, XLMTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = XLMTokenizer __SCREAMING_SNAKE_CASE = False def A ( self : str ): """simple docstring""" super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt __snake_case = [ "l", "o", "w", "e", "r", "s", "t", "i", "d", "n", "w</w>", "r</w>", "t</w>", "lo", "low", "er</w>", "low</w>", "lowest</w>", "newer</w>", "wider</w>", "<unk>", ] __snake_case = dict(zip(a_ , range(len(a_ ) ) ) ) __snake_case = ["l o 123", "lo w 1456", "e r</w> 1789", ""] __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) __snake_case = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["merges_file"] ) with open(self.vocab_file , "w" ) as fp: fp.write(json.dumps(a_ ) ) with open(self.merges_file , "w" ) as fp: fp.write("\n".join(a_ ) ) def A ( self : Union[str, Any] , a_ : Optional[int] ): """simple docstring""" __snake_case = "lower newer" __snake_case = "lower newer" return input_text, output_text def A ( self : Dict ): """simple docstring""" __snake_case = XLMTokenizer(self.vocab_file , self.merges_file ) __snake_case = "lower" __snake_case = ["low", "er</w>"] __snake_case = tokenizer.tokenize(a_ ) self.assertListEqual(a_ , a_ ) __snake_case = tokens + ["<unk>"] __snake_case = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , a_ ) @slow def A ( self : List[str] ): """simple docstring""" __snake_case = XLMTokenizer.from_pretrained("xlm-mlm-en-2048" ) __snake_case = tokenizer.encode("sequence builders" , add_special_tokens=a_ ) __snake_case = tokenizer.encode("multi-sequence build" , add_special_tokens=a_ ) __snake_case = tokenizer.build_inputs_with_special_tokens(a_ ) __snake_case = tokenizer.build_inputs_with_special_tokens(a_ , a_ ) assert encoded_sentence == [0] + text + [1] assert encoded_pair == [0] + text + [1] + text_a + [1]
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str ) -> str: return " ".join( "".join(word[::-1] ) if len(_UpperCAmelCase ) > 4 else word for word in sentence.split() ) if __name__ == "__main__": import doctest doctest.testmod() print(reverse_long_words('''Hey wollef sroirraw'''))
69
1
'''simple docstring''' import torch from diffusers import EulerDiscreteScheduler from diffusers.utils import torch_device from .test_schedulers import SchedulerCommonTest class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = (EulerDiscreteScheduler,) __SCREAMING_SNAKE_CASE = 10 def A ( self : int , **a_ : int ): """simple docstring""" __snake_case = { "num_train_timesteps": 1_100, "beta_start": 0.0001, "beta_end": 0.02, "beta_schedule": "linear", } config.update(**a_ ) return config def A ( self : Tuple ): """simple docstring""" for timesteps in [10, 50, 100, 1_000]: self.check_over_configs(num_train_timesteps=a_ ) def A ( self : Optional[Any] ): """simple docstring""" for beta_start, beta_end in zip([0.00001, 0.0001, 0.001] , [0.0002, 0.002, 0.02] ): self.check_over_configs(beta_start=a_ , beta_end=a_ ) def A ( self : Optional[Any] ): """simple docstring""" for schedule in ["linear", "scaled_linear"]: self.check_over_configs(beta_schedule=a_ ) def A ( self : List[Any] ): """simple docstring""" for prediction_type in ["epsilon", "v_prediction"]: self.check_over_configs(prediction_type=a_ ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) __snake_case = torch.manual_seed(0 ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma __snake_case = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ , generator=a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def A ( self : Tuple ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config(prediction_type="v_prediction" ) __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps ) __snake_case = torch.manual_seed(0 ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma __snake_case = sample.to(a_ ) for i, t in enumerate(scheduler.timesteps ): __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ , generator=a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 0.0002 ) < 1e-2 assert abs(result_mean.item() - 2.2_676e-06 ) < 1e-3 def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a_ ) scheduler.set_timesteps(self.num_inference_steps , device=a_ ) __snake_case = torch.manual_seed(0 ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() __snake_case = sample.to(a_ ) for t in scheduler.timesteps: __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ , generator=a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 10.0807 ) < 1e-2 assert abs(result_mean.item() - 0.0131 ) < 1e-3 def A ( self : List[Any] ): """simple docstring""" __snake_case = self.scheduler_classes[0] __snake_case = self.get_scheduler_config() __snake_case = scheduler_class(**a_ , use_karras_sigmas=a_ ) scheduler.set_timesteps(self.num_inference_steps , device=a_ ) __snake_case = torch.manual_seed(0 ) __snake_case = self.dummy_model() __snake_case = self.dummy_sample_deter * scheduler.init_noise_sigma.cpu() __snake_case = sample.to(a_ ) for t in scheduler.timesteps: __snake_case = scheduler.scale_model_input(a_ , a_ ) __snake_case = model(a_ , a_ ) __snake_case = scheduler.step(a_ , a_ , a_ , generator=a_ ) __snake_case = output.prev_sample __snake_case = torch.sum(torch.abs(a_ ) ) __snake_case = torch.mean(torch.abs(a_ ) ) assert abs(result_sum.item() - 124.52299499511719 ) < 1e-2 assert abs(result_mean.item() - 0.16213932633399963 ) < 1e-3
69
'''simple docstring''' import unittest from transformers import MPNetConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : Any , a_ : Union[str, Any]=13 , a_ : Any=7 , a_ : Any=True , a_ : Dict=True , a_ : Union[str, Any]=False , a_ : Tuple=True , a_ : str=99 , a_ : Tuple=64 , a_ : Tuple=5 , a_ : Union[str, Any]=4 , a_ : Dict=64 , a_ : Union[str, Any]="gelu" , a_ : Dict=0.1 , a_ : List[str]=0.1 , a_ : Dict=512 , a_ : Tuple=16 , a_ : str=2 , a_ : Any=0.02 , a_ : List[Any]=3 , a_ : Tuple=4 , a_ : Optional[int]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : int ): """simple docstring""" return MPNetConfig.from_pretrained("microsoft/mpnet-base" ) def A ( self : str ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : List[str] ): """simple docstring""" return MPNetConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def A ( self : Tuple , a_ : int , a_ : str , a_ : Optional[int] , a_ : List[Any] , a_ : str , a_ : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def A ( self : Any , a_ : int , a_ : Tuple , a_ : str , a_ : int , a_ : str , a_ : List[Any] ): """simple docstring""" __snake_case = MPNetForQuestionAnswering(config=a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , start_positions=a_ , end_positions=a_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A ( self : Any , a_ : Any , a_ : int , a_ : Union[str, Any] , a_ : Dict , a_ : Optional[Any] , a_ : Any ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A ( self : Optional[Any] , a_ : Any , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : Union[str, Any] , a_ : List[Any] , a_ : List[Any] ): """simple docstring""" __snake_case = self.num_choices __snake_case = MPNetForMultipleChoice(config=a_ ) model.to(a_ ) model.eval() __snake_case = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __snake_case = model( a_ , attention_mask=a_ , labels=a_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A ( self : Dict , a_ : List[str] , a_ : str , a_ : Union[str, Any] , a_ : str , a_ : Optional[int] , a_ : Optional[Any] ): """simple docstring""" __snake_case = self.num_labels __snake_case = MPNetForTokenClassification(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ((__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case) , (__snake_case)) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( MPNetForMaskedLM, MPNetForMultipleChoice, MPNetForQuestionAnswering, MPNetForSequenceClassification, MPNetForTokenClassification, MPNetModel, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": MPNetModel, """fill-mask""": MPNetForMaskedLM, """question-answering""": MPNetForQuestionAnswering, """text-classification""": MPNetForSequenceClassification, """token-classification""": MPNetForTokenClassification, """zero-shot""": MPNetForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True def A ( self : List[Any] ): """simple docstring""" __snake_case = MPNetModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : List[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_model(*a_ ) def A ( self : Dict ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_sequence_classification(*a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_multiple_choice(*a_ ) def A ( self : int ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_token_classification(*a_ ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mpnet_for_question_answering(*a_ ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Optional[Any] ): """simple docstring""" __snake_case = MPNetModel.from_pretrained("microsoft/mpnet-base" ) __snake_case = torch.tensor([[0, 345, 232, 328, 740, 140, 1_695, 69, 6_078, 1_588, 2]] ) __snake_case = model(a_ )[0] __snake_case = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , a_ ) __snake_case = torch.tensor( [[[-0.0550, 0.1943, -0.0740], [-0.0562, 0.2211, -0.0579], [-0.0437, 0.3337, -0.0641]]] ) # compare the actual values for a slice. self.assertTrue(torch.allclose(output[:, :3, :3] , a_ , atol=1e-4 ) )
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : int ) -> int: return abs(_UpperCAmelCase ) if a == 0 else greatest_common_divisor(b % a , _UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int , _UpperCAmelCase : int ) -> int: while y: # --> when y=0 then loop will terminate and return x as final GCD. __snake_case , __snake_case = y, x % y return abs(_UpperCAmelCase ) def __UpperCAmelCase ( ) -> Union[str, Any]: try: __snake_case = input("Enter two integers separated by comma (,): " ).split("," ) __snake_case = int(nums[0] ) __snake_case = int(nums[1] ) print( F'''greatest_common_divisor({num_a}, {num_a}) = ''' F'''{greatest_common_divisor(_UpperCAmelCase , _UpperCAmelCase )}''' ) print(F'''By iterative gcd({num_a}, {num_a}) = {gcd_by_iterative(_UpperCAmelCase , _UpperCAmelCase )}''' ) except (IndexError, UnboundLocalError, ValueError): print("Wrong input" ) if __name__ == "__main__": main()
69
'''simple docstring''' # Logistic Regression from scratch # In[62]: # In[63]: # importing all the required libraries import numpy as np from matplotlib import pyplot as plt from sklearn import datasets def __UpperCAmelCase ( _UpperCAmelCase : str ) -> Optional[int]: return 1 / (1 + np.exp(-z )) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Dict ) -> List[str]: return (-y * np.log(_UpperCAmelCase ) - (1 - y) * np.log(1 - h )).mean() def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any] ) -> Optional[Any]: __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) return np.sum(y * scores - np.log(1 + np.exp(_UpperCAmelCase ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict , _UpperCAmelCase : List[str]=7_00_00 ) -> Union[str, Any]: __snake_case = np.zeros(x.shape[1] ) for iterations in range(_UpperCAmelCase ): __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = np.dot(x.T , h - y ) / y.size __snake_case = theta - alpha * gradient # updating the weights __snake_case = np.dot(_UpperCAmelCase , _UpperCAmelCase ) __snake_case = sigmoid_function(_UpperCAmelCase ) __snake_case = cost_function(_UpperCAmelCase , _UpperCAmelCase ) if iterations % 1_00 == 0: print(F'''loss: {j} \t''' ) # printing the loss after every 100 iterations return theta # In[68]: if __name__ == "__main__": a : int = datasets.load_iris() a : int = iris.data[:, :2] a : Optional[Any] = (iris.target != 0) * 1 a : Tuple = 0.1 a : List[str] = logistic_reg(alpha, x, y, max_iterations=70_000) print('''theta: ''', theta) # printing the theta i.e our weights vector def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> Union[str, Any]: return sigmoid_function( np.dot(_UpperCAmelCase , _UpperCAmelCase ) ) # predicting the value of probability from the logistic regression algorithm plt.figure(figsize=(10, 6)) plt.scatter(x[y == 0][:, 0], x[y == 0][:, 1], color='''b''', label='''0''') plt.scatter(x[y == 1][:, 0], x[y == 1][:, 1], color='''r''', label='''1''') ((a) , (a)) : Any = (x[:, 0].min(), x[:, 0].max()) ((a) , (a)) : Any = (x[:, 1].min(), x[:, 1].max()) ((a) , (a)) : Any = np.meshgrid(np.linspace(xa_min, xa_max), np.linspace(xa_min, xa_max)) a : Optional[Any] = np.c_[xxa.ravel(), xxa.ravel()] a : List[Any] = predict_prob(grid).reshape(xxa.shape) plt.contour(xxa, xxa, probs, [0.5], linewidths=1, colors='''black''') plt.legend() plt.show()
69
1
'''simple docstring''' from typing import TYPE_CHECKING from ....utils import _LazyModule a : Any = {'''tokenization_tapex''': ['''TapexTokenizer''']} if TYPE_CHECKING: from .tokenization_tapex import TapexTokenizer else: import sys a : List[str] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> bool: return number & 1 == 0 if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import math import torch from torch import nn from ..configuration_utils import ConfigMixin, register_to_config from .attention_processor import Attention from .embeddings import get_timestep_embedding from .modeling_utils import ModelMixin class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase ): @register_to_config def __init__( self : Dict , a_ : int = 128 , a_ : int = 256 , a_ : float = 2000.0 , a_ : int = 768 , a_ : int = 12 , a_ : int = 12 , a_ : int = 64 , a_ : int = 2_048 , a_ : float = 0.1 , ): """simple docstring""" super().__init__() __snake_case = nn.Sequential( nn.Linear(a_ , d_model * 4 , bias=a_ ) , nn.SiLU() , nn.Linear(d_model * 4 , d_model * 4 , bias=a_ ) , nn.SiLU() , ) __snake_case = nn.Embedding(a_ , a_ ) __snake_case = False __snake_case = nn.Linear(a_ , a_ , bias=a_ ) __snake_case = nn.Dropout(p=a_ ) __snake_case = nn.ModuleList() for lyr_num in range(a_ ): # FiLM conditional T5 decoder __snake_case = DecoderLayer(d_model=a_ , d_kv=a_ , num_heads=a_ , d_ff=a_ , dropout_rate=a_ ) self.decoders.append(a_ ) __snake_case = TaLayerNorm(a_ ) __snake_case = nn.Dropout(p=a_ ) __snake_case = nn.Linear(a_ , a_ , bias=a_ ) def A ( self : Any , a_ : Tuple , a_ : Optional[Any] ): """simple docstring""" __snake_case = torch.mul(query_input.unsqueeze(-1 ) , key_input.unsqueeze(-2 ) ) return mask.unsqueeze(-3 ) def A ( self : int , a_ : Any , a_ : Optional[int] , a_ : Any ): """simple docstring""" __snake_case , __snake_case , __snake_case = decoder_input_tokens.shape assert decoder_noise_time.shape == (batch,) # decoder_noise_time is in [0, 1), so rescale to expected timing range. __snake_case = get_timestep_embedding( decoder_noise_time * self.config.max_decoder_noise_time , embedding_dim=self.config.d_model , max_period=self.config.max_decoder_noise_time , ).to(dtype=self.dtype ) __snake_case = self.conditioning_emb(a_ ).unsqueeze(1 ) assert conditioning_emb.shape == (batch, 1, self.config.d_model * 4) __snake_case = decoder_input_tokens.shape[1] # If we want to use relative positions for audio context, we can just offset # this sequence by the length of encodings_and_masks. __snake_case = torch.broadcast_to( torch.arange(a_ , device=decoder_input_tokens.device ) , (batch, seq_length) , ) __snake_case = self.position_encoding(a_ ) __snake_case = self.continuous_inputs_projection(a_ ) inputs += position_encodings __snake_case = self.dropout(a_ ) # decoder: No padding present. __snake_case = torch.ones( decoder_input_tokens.shape[:2] , device=decoder_input_tokens.device , dtype=inputs.dtype ) # Translate encoding masks to encoder-decoder masks. __snake_case = [(x, self.encoder_decoder_mask(a_ , a_ )) for x, y in encodings_and_masks] # cross attend style: concat encodings __snake_case = torch.cat([x[0] for x in encodings_and_encdec_masks] , dim=1 ) __snake_case = torch.cat([x[1] for x in encodings_and_encdec_masks] , dim=-1 ) for lyr in self.decoders: __snake_case = lyr( a_ , conditioning_emb=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , )[0] __snake_case = self.decoder_norm(a_ ) __snake_case = self.post_dropout(a_ ) __snake_case = self.spec_out(a_ ) return spec_out class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Dict , a_ : Optional[int] , a_ : Dict , a_ : List[str] , a_ : Any , a_ : Tuple , a_ : Optional[int]=1e-6 ): """simple docstring""" super().__init__() __snake_case = nn.ModuleList() # cond self attention: layer 0 self.layer.append( TaLayerSelfAttentionCond(d_model=a_ , d_kv=a_ , num_heads=a_ , dropout_rate=a_ ) ) # cross attention: layer 1 self.layer.append( TaLayerCrossAttention( d_model=a_ , d_kv=a_ , num_heads=a_ , dropout_rate=a_ , layer_norm_epsilon=a_ , ) ) # Film Cond MLP + dropout: last layer self.layer.append( TaLayerFFCond(d_model=a_ , d_ff=a_ , dropout_rate=a_ , layer_norm_epsilon=a_ ) ) def A ( self : Any , a_ : List[str] , a_ : Tuple=None , a_ : Any=None , a_ : Union[str, Any]=None , a_ : List[str]=None , a_ : Dict=None , ): """simple docstring""" __snake_case = self.layer[0]( a_ , conditioning_emb=a_ , attention_mask=a_ , ) if encoder_hidden_states is not None: __snake_case = torch.where(encoder_attention_mask > 0 , 0 , -1e10 ).to( encoder_hidden_states.dtype ) __snake_case = self.layer[1]( a_ , key_value_states=a_ , attention_mask=a_ , ) # Apply Film Conditional Feed Forward layer __snake_case = self.layer[-1](a_ , a_ ) return (hidden_states,) class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Union[str, Any] , a_ : Optional[Any] , a_ : Optional[Any] , a_ : List[Any] , a_ : Tuple ): """simple docstring""" super().__init__() __snake_case = TaLayerNorm(a_ ) __snake_case = TaFiLMLayer(in_features=d_model * 4 , out_features=a_ ) __snake_case = Attention(query_dim=a_ , heads=a_ , dim_head=a_ , out_bias=a_ , scale_qk=a_ ) __snake_case = nn.Dropout(a_ ) def A ( self : Optional[Any] , a_ : str , a_ : Union[str, Any]=None , a_ : Tuple=None , ): """simple docstring""" __snake_case = self.layer_norm(a_ ) if conditioning_emb is not None: __snake_case = self.FiLMLayer(a_ , a_ ) # Self-attention block __snake_case = self.attention(a_ ) __snake_case = hidden_states + self.dropout(a_ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Any , a_ : List[str] , a_ : List[str] , a_ : List[str] , a_ : Optional[Any] , a_ : int ): """simple docstring""" super().__init__() __snake_case = Attention(query_dim=a_ , heads=a_ , dim_head=a_ , out_bias=a_ , scale_qk=a_ ) __snake_case = TaLayerNorm(a_ , eps=a_ ) __snake_case = nn.Dropout(a_ ) def A ( self : int , a_ : Any , a_ : Dict=None , a_ : Optional[Any]=None , ): """simple docstring""" __snake_case = self.layer_norm(a_ ) __snake_case = self.attention( a_ , encoder_hidden_states=a_ , attention_mask=attention_mask.squeeze(1 ) , ) __snake_case = hidden_states + self.dropout(a_ ) return layer_output class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Union[str, Any] , a_ : Optional[int] , a_ : Tuple , a_ : Optional[int] , a_ : List[str] ): """simple docstring""" super().__init__() __snake_case = TaDenseGatedActDense(d_model=a_ , d_ff=a_ , dropout_rate=a_ ) __snake_case = TaFiLMLayer(in_features=d_model * 4 , out_features=a_ ) __snake_case = TaLayerNorm(a_ , eps=a_ ) __snake_case = nn.Dropout(a_ ) def A ( self : Optional[int] , a_ : Union[str, Any] , a_ : Union[str, Any]=None ): """simple docstring""" __snake_case = self.layer_norm(a_ ) if conditioning_emb is not None: __snake_case = self.film(a_ , a_ ) __snake_case = self.DenseReluDense(a_ ) __snake_case = hidden_states + self.dropout(a_ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : Optional[int] , a_ : str , a_ : Any , a_ : Dict ): """simple docstring""" super().__init__() __snake_case = nn.Linear(a_ , a_ , bias=a_ ) __snake_case = nn.Linear(a_ , a_ , bias=a_ ) __snake_case = nn.Linear(a_ , a_ , bias=a_ ) __snake_case = nn.Dropout(a_ ) __snake_case = NewGELUActivation() def A ( self : Any , a_ : str ): """simple docstring""" __snake_case = self.act(self.wi_a(a_ ) ) __snake_case = self.wi_a(a_ ) __snake_case = hidden_gelu * hidden_linear __snake_case = self.dropout(a_ ) __snake_case = self.wo(a_ ) return hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : List[str] , a_ : Optional[Any] , a_ : Optional[Any]=1e-6 ): """simple docstring""" super().__init__() __snake_case = nn.Parameter(torch.ones(a_ ) ) __snake_case = eps def A ( self : List[str] , a_ : Any ): """simple docstring""" __snake_case = hidden_states.to(torch.floataa ).pow(2 ).mean(-1 , keepdim=a_ ) __snake_case = hidden_states * torch.rsqrt(variance + self.variance_epsilon ) # convert into half-precision if necessary if self.weight.dtype in [torch.floataa, torch.bfloataa]: __snake_case = hidden_states.to(self.weight.dtype ) return self.weight * hidden_states class SCREAMING_SNAKE_CASE__ ( nn.Module ): def A ( self : List[str] , a_ : torch.Tensor ): """simple docstring""" return 0.5 * input * (1.0 + torch.tanh(math.sqrt(2.0 / math.pi ) * (input + 0.044715 * torch.pow(a_ , 3.0 )) )) class SCREAMING_SNAKE_CASE__ ( nn.Module ): def __init__( self : int , a_ : Union[str, Any] , a_ : List[str] ): """simple docstring""" super().__init__() __snake_case = nn.Linear(a_ , out_features * 2 , bias=a_ ) def A ( self : Optional[int] , a_ : Union[str, Any] , a_ : Optional[int] ): """simple docstring""" __snake_case = self.scale_bias(a_ ) __snake_case , __snake_case = torch.chunk(a_ , 2 , -1 ) __snake_case = x * (1 + scale) + shift return x
69
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int = 1 , _UpperCAmelCase : int = 10_00 ) -> int: __snake_case = 1 __snake_case = 0 for divide_by_number in range(_UpperCAmelCase , digit + 1 ): __snake_case = [] __snake_case = numerator for _ in range(1 , digit + 1 ): if now_divide in has_been_divided: if longest_list_length < len(_UpperCAmelCase ): __snake_case = len(_UpperCAmelCase ) __snake_case = divide_by_number else: has_been_divided.append(_UpperCAmelCase ) __snake_case = now_divide * 10 % divide_by_number return the_digit # Tests if __name__ == "__main__": import doctest doctest.testmod()
69
'''simple docstring''' from typing import List, Optional from ...configuration_utils import PretrainedConfig from ...utils import logging a : List[str] = logging.get_logger(__name__) a : Tuple = { '''huggingface/autoformer-tourism-monthly''': '''https://huggingface.co/huggingface/autoformer-tourism-monthly/resolve/main/config.json''', } class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = """autoformer""" __SCREAMING_SNAKE_CASE = { """hidden_size""": """d_model""", """num_attention_heads""": """encoder_attention_heads""", """num_hidden_layers""": """encoder_layers""", } def __init__( self : List[Any] , a_ : Optional[int] = None , a_ : Optional[int] = None , a_ : str = "student_t" , a_ : str = "nll" , a_ : int = 1 , a_ : List[int] = [1, 2, 3, 4, 5, 6, 7] , a_ : bool = True , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : int = 0 , a_ : Optional[List[int]] = None , a_ : Optional[List[int]] = None , a_ : int = 64 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 2 , a_ : int = 32 , a_ : int = 32 , a_ : str = "gelu" , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : float = 0.1 , a_ : int = 100 , a_ : float = 0.02 , a_ : bool = True , a_ : Union[str, Any]=True , a_ : int = 10 , a_ : int = 25 , a_ : int = 3 , **a_ : Tuple , ): """simple docstring""" __snake_case = prediction_length __snake_case = context_length if context_length is not None else prediction_length __snake_case = distribution_output __snake_case = loss __snake_case = input_size __snake_case = num_time_features __snake_case = lags_sequence __snake_case = scaling __snake_case = num_dynamic_real_features __snake_case = num_static_real_features __snake_case = num_static_categorical_features if cardinality is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The cardinality should be a list of the same length as `num_static_categorical_features`" ) __snake_case = cardinality else: __snake_case = [0] if embedding_dimension is not None and num_static_categorical_features > 0: if len(a_ ) != num_static_categorical_features: raise ValueError( "The embedding dimension should be a list of the same length as `num_static_categorical_features`" ) __snake_case = embedding_dimension else: __snake_case = [min(50 , (cat + 1) // 2 ) for cat in self.cardinality] __snake_case = num_parallel_samples # Transformer architecture configuration __snake_case = input_size * len(self.lags_sequence ) + self._number_of_features __snake_case = d_model __snake_case = encoder_attention_heads __snake_case = decoder_attention_heads __snake_case = encoder_ffn_dim __snake_case = decoder_ffn_dim __snake_case = encoder_layers __snake_case = decoder_layers __snake_case = dropout __snake_case = attention_dropout __snake_case = activation_dropout __snake_case = encoder_layerdrop __snake_case = decoder_layerdrop __snake_case = activation_function __snake_case = init_std __snake_case = use_cache # Autoformer __snake_case = label_length __snake_case = moving_average __snake_case = autocorrelation_factor super().__init__(is_encoder_decoder=a_ , **a_ ) @property def A ( self : Optional[int] ): """simple docstring""" return ( sum(self.embedding_dimension ) + self.num_dynamic_real_features + self.num_time_features + self.num_static_real_features + self.input_size * 2 # the log1p(abs(loc)) and log(scale) features )
69
1
'''simple docstring''' from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL a : Tuple = logging.get_logger(__name__) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = ["""pixel_values"""] def __init__( self : Tuple , a_ : bool = True , a_ : Dict[str, int] = None , a_ : int = 0.9 , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : bool = True , a_ : Dict[str, int] = None , a_ : Union[int, float] = 1 / 255 , a_ : bool = True , a_ : bool = True , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , **a_ : Optional[int] , ): """simple docstring""" super().__init__(**a_ ) __snake_case = size if size is not None else {"shortest_edge": 224} __snake_case = get_size_dict(a_ , default_to_square=a_ ) __snake_case = crop_size if crop_size is not None else {"height": 224, "width": 224} __snake_case = get_size_dict(a_ , param_name="crop_size" ) __snake_case = do_resize __snake_case = size __snake_case = crop_pct __snake_case = resample __snake_case = do_center_crop __snake_case = crop_size __snake_case = do_rescale __snake_case = rescale_factor __snake_case = do_normalize __snake_case = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN __snake_case = image_std if image_std is not None else IMAGENET_DEFAULT_STD def A ( self : Tuple , a_ : np.ndarray , a_ : Dict[str, int] , a_ : Optional[float] = None , a_ : PILImageResampling = PILImageResampling.BICUBIC , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Any , ): """simple docstring""" __snake_case = get_size_dict(a_ , default_to_square=a_ ) if "shortest_edge" not in size and ("height" not in size or "width" not in size): raise ValueError(f'''size must contain \'height\' and \'width\' or \'shortest_edge\' as keys. Got {size.keys()}''' ) if crop_pct is not None: if "shortest_edge" in size: __snake_case = int(size["shortest_edge"] / crop_pct ) elif "height" in size and "width" in size: if size["height"] == size["width"]: __snake_case = int(size["height"] / crop_pct ) else: __snake_case = (int(size["height"] / crop_pct ), int(size["width"] / crop_pct )) else: raise ValueError("Invalid size for resize: {}".format(a_ ) ) __snake_case = get_resize_output_image_size(a_ , size=a_ , default_to_square=a_ ) else: if "shortest_edge" in size: __snake_case = get_resize_output_image_size(a_ , size=size["shortest_edge"] , default_to_square=a_ ) elif "height" in size and "width" in size: __snake_case = (size["height"], size["width"]) else: raise ValueError("Invalid size for resize: {}".format(a_ ) ) return resize(a_ , size=a_ , resample=a_ , data_format=a_ , **a_ ) def A ( self : int , a_ : np.ndarray , a_ : Dict[str, int] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : str , ): """simple docstring""" __snake_case = get_size_dict(a_ ) if "height" not in size or "width" not in size: raise ValueError(f'''size must contain \'height\' and \'width\' as keys. Got {size.keys()}''' ) return center_crop(a_ , size=(size["height"], size["width"]) , data_format=a_ , **a_ ) def A ( self : Dict , a_ : np.ndarray , a_ : Union[int, float] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Any , ): """simple docstring""" return rescale(a_ , scale=a_ , data_format=a_ , **a_ ) def A ( self : str , a_ : np.ndarray , a_ : Union[float, List[float]] , a_ : Union[float, List[float]] , a_ : Optional[Union[str, ChannelDimension]] = None , **a_ : Dict , ): """simple docstring""" return normalize(a_ , mean=a_ , std=a_ , data_format=a_ , **a_ ) def A ( self : Dict , a_ : ImageInput , a_ : bool = None , a_ : Dict[str, int] = None , a_ : int = None , a_ : PILImageResampling = None , a_ : bool = None , a_ : Dict[str, int] = None , a_ : bool = None , a_ : float = None , a_ : bool = None , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[float, List[float]]] = None , a_ : Optional[Union[str, TensorType]] = None , a_ : ChannelDimension = ChannelDimension.FIRST , **a_ : int , ): """simple docstring""" __snake_case = do_resize if do_resize is not None else self.do_resize __snake_case = crop_pct if crop_pct is not None else self.crop_pct __snake_case = resample if resample is not None else self.resample __snake_case = do_center_crop if do_center_crop is not None else self.do_center_crop __snake_case = do_rescale if do_rescale is not None else self.do_rescale __snake_case = rescale_factor if rescale_factor is not None else self.rescale_factor __snake_case = do_normalize if do_normalize is not None else self.do_normalize __snake_case = image_mean if image_mean is not None else self.image_mean __snake_case = image_std if image_std is not None else self.image_std __snake_case = size if size is not None else self.size __snake_case = get_size_dict(a_ , default_to_square=a_ ) __snake_case = crop_size if crop_size is not None else self.crop_size __snake_case = get_size_dict(a_ , param_name="crop_size" ) __snake_case = make_list_of_images(a_ ) if not valid_images(a_ ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_pct is None: raise ValueError("Crop_pct must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. __snake_case = [to_numpy_array(a_ ) for image in images] if do_resize: __snake_case = [self.resize(image=a_ , size=a_ , crop_pct=a_ , resample=a_ ) for image in images] if do_center_crop: __snake_case = [self.center_crop(image=a_ , size=a_ ) for image in images] if do_rescale: __snake_case = [self.rescale(image=a_ , scale=a_ ) for image in images] if do_normalize: __snake_case = [self.normalize(image=a_ , mean=a_ , std=a_ ) for image in images] __snake_case = [to_channel_dimension_format(a_ , a_ ) for image in images] __snake_case = {"pixel_values": images} return BatchFeature(data=a_ , tensor_type=a_ )
69
'''simple docstring''' import unittest from transformers import GPTSwaTokenizer from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin a : List[Any] = get_tests_dir('''fixtures/test_sentencepiece_with_bytefallback.model''') @require_sentencepiece @require_tokenizers class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = GPTSwaTokenizer __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = True __SCREAMING_SNAKE_CASE = False def A ( self : int ): """simple docstring""" super().setUp() # We have a SentencePiece fixture for testing __snake_case = GPTSwaTokenizer(a_ , eos_token="<unk>" , bos_token="<unk>" , pad_token="<unk>" ) tokenizer.save_pretrained(self.tmpdirname ) def A ( self : str , a_ : List[Any] ): """simple docstring""" __snake_case = "This is a test" __snake_case = "This is a test" return input_text, output_text def A ( self : Union[str, Any] ): """simple docstring""" __snake_case = "<s>" __snake_case = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(a_ ) , a_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(a_ ) , a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<unk>" ) self.assertEqual(vocab_keys[1] , "<s>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(a_ ) , 2_000 ) def A ( self : Optional[int] ): """simple docstring""" self.assertEqual(self.get_tokenizer().vocab_size , 2_000 ) def A ( self : Dict ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = tokenizer.tokenize("This is a test" ) self.assertListEqual(a_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(a_ ) , [465, 287, 265, 631, 842] ) __snake_case = tokenizer.tokenize("I was born in 92000, and this is falsé." ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] , ) # fmt: on __snake_case = tokenizer.convert_tokens_to_ids(a_ ) self.assertListEqual( a_ , [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260] , ) __snake_case = tokenizer.convert_ids_to_tokens(a_ ) # fmt: off self.assertListEqual( a_ , ["▁I", "▁was", "▁bor", "n", "▁in", "▁", "<0x39>", "2", "0", "0", "0", ",", "▁and", "▁this", "▁is", "▁f", "al", "s", "<0xC3>", "<0xA9>", "."] ) # fmt: on def A ( self : List[str] ): """simple docstring""" __snake_case = GPTSwaTokenizer(a_ ) __snake_case = ["This is a test", "I was born in 92000, and this is falsé."] __snake_case = [ [465, 287, 265, 631, 842], [262, 272, 1_525, 286, 271, 268, 60, 916, 633, 633, 633, 259, 266, 301, 287, 384, 367, 263, 198, 172, 260], ] # Test that encode_fast returns the same as tokenize + convert_tokens_to_ids for text, expected_ids in zip(a_ , a_ ): self.assertListEqual(tokenizer.encode_fast(a_ ) , a_ ) # Test that decode_fast returns the input text for text, token_ids in zip(a_ , a_ ): self.assertEqual(tokenizer.decode_fast(a_ ) , a_ ) @slow def A ( self : Any ): """simple docstring""" __snake_case = [ "<|python|>def fibonacci(n)\n if n < 0:\n print('Incorrect input')", "Hey there, how are you doing this fine day?", "This is a text with a trailing spaces followed by a dot .", "Häj sväjs lillebrör! =)", "Det är inget fel på Mr. Cool", ] # fmt: off __snake_case = {"input_ids": [[63_423, 5, 6_811, 14_954, 282, 816, 3_821, 63_466, 63_425, 63_462, 18, 63_978, 678, 301, 1_320, 63_423, 63_455, 63_458, 18, 63_982, 4_246, 3_940, 1_901, 47_789, 5_547, 18_994], [19_630, 1_100, 63_446, 1_342, 633, 544, 4_488, 593, 5_102, 2_416, 63_495, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1_652, 428, 268, 1_936, 515, 268, 58_593, 22_413, 9_106, 546, 268, 33_213, 63_979, 698, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [55_130, 63_450, 924, 63_449, 2_249, 4_062, 1_558, 318, 63_504, 21_498, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [509, 377, 2_827, 2_559, 332, 6_575, 63_443, 26_801, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "token_type_ids": [[0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # fmt: on self.tokenizer_integration_test_util( expected_encoding=a_ , model_name="AI-Sweden/gpt-sw3-126m" , sequences=a_ , )
69
1
'''simple docstring''' import argparse import logging from collections import namedtuple import torch from model_bertabs import BertAbsSummarizer from models.model_builder import AbsSummarizer # The authors' implementation from transformers import BertTokenizer logging.basicConfig(level=logging.INFO) a : Tuple = logging.getLogger(__name__) a : Union[str, Any] = '''Hello world! cécé herlolip''' a : Optional[Any] = namedtuple( '''BertAbsConfig''', [ '''temp_dir''', '''large''', '''use_bert_emb''', '''finetune_bert''', '''encoder''', '''share_emb''', '''max_pos''', '''enc_layers''', '''enc_hidden_size''', '''enc_heads''', '''enc_ff_size''', '''enc_dropout''', '''dec_layers''', '''dec_hidden_size''', '''dec_heads''', '''dec_ff_size''', '''dec_dropout''', ], ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[Any] ) -> Union[str, Any]: __snake_case = BertAbsConfig( temp_dir="." , finetune_bert=_UpperCAmelCase , large=_UpperCAmelCase , share_emb=_UpperCAmelCase , use_bert_emb=_UpperCAmelCase , encoder="bert" , max_pos=5_12 , enc_layers=6 , enc_hidden_size=5_12 , enc_heads=8 , enc_ff_size=5_12 , enc_dropout=0.2 , dec_layers=6 , dec_hidden_size=7_68 , dec_heads=8 , dec_ff_size=20_48 , dec_dropout=0.2 , ) __snake_case = torch.load(_UpperCAmelCase , lambda _UpperCAmelCase , _UpperCAmelCase : storage ) __snake_case = AbsSummarizer(_UpperCAmelCase , torch.device("cpu" ) , _UpperCAmelCase ) original.eval() __snake_case = BertAbsSummarizer(_UpperCAmelCase , torch.device("cpu" ) ) new_model.eval() # ------------------- # Convert the weights # ------------------- logging.info("convert the model" ) new_model.bert.load_state_dict(original.bert.state_dict() ) new_model.decoder.load_state_dict(original.decoder.state_dict() ) new_model.generator.load_state_dict(original.generator.state_dict() ) # ---------------------------------- # Make sure the outpus are identical # ---------------------------------- logging.info("Make sure that the models' outputs are identical" ) __snake_case = BertTokenizer.from_pretrained("bert-base-uncased" ) # prepare the model inputs __snake_case = tokenizer.encode("This is sample éàalj'-." ) encoder_input_ids.extend([tokenizer.pad_token_id] * (5_12 - len(_UpperCAmelCase )) ) __snake_case = torch.tensor(_UpperCAmelCase ).unsqueeze(0 ) __snake_case = tokenizer.encode("This is sample 3 éàalj'-." ) decoder_input_ids.extend([tokenizer.pad_token_id] * (5_12 - len(_UpperCAmelCase )) ) __snake_case = torch.tensor(_UpperCAmelCase ).unsqueeze(0 ) # failsafe to make sure the weights reset does not affect the # loaded weights. assert torch.max(torch.abs(original.generator[0].weight - new_model.generator[0].weight ) ) == 0 # forward pass __snake_case = encoder_input_ids __snake_case = decoder_input_ids __snake_case = __snake_case = None __snake_case = None __snake_case = __snake_case = None __snake_case = __snake_case = None __snake_case = None # The original model does not apply the geneator layer immediatly but rather in # the beam search (where it combines softmax + linear layer). Since we already # apply the softmax in our generation process we only apply the linear layer here. # We make sure that the outputs of the full stack are identical __snake_case = original(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase )[0] __snake_case = original.generator(_UpperCAmelCase ) __snake_case = new_model( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase )[0] __snake_case = new_model.generator(_UpperCAmelCase ) __snake_case = torch.max(torch.abs(output_converted_model - output_original_model ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(_UpperCAmelCase ) ) __snake_case = torch.max(torch.abs(output_converted_generator - output_original_generator ) ).item() print("Maximum absolute difference beween weights: {:.2f}".format(_UpperCAmelCase ) ) __snake_case = torch.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ) if are_identical: logging.info("all weights are equal up to 1e-3" ) else: raise ValueError("the weights are different. The new model is likely different from the original one." ) # The model has been saved with torch.save(model) and this is bound to the exact # directory structure. We save the state_dict instead. logging.info("saving the model's state dictionary" ) torch.save( new_model.state_dict() , "./bertabs-finetuned-cnndm-extractive-abstractive-summarization/pytorch_model.bin" ) if __name__ == "__main__": a : Dict = argparse.ArgumentParser() parser.add_argument( '''--bertabs_checkpoint_path''', default=None, type=str, required=True, help='''Path the official PyTorch dump.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''', ) a : Dict = parser.parse_args() convert_bertabs_checkpoints( args.bertabs_checkpoint_path, args.pytorch_dump_folder_path, )
69
'''simple docstring''' import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, FEATURE_EXTRACTOR_MAPPING, AutoConfig, AutoFeatureExtractor, WavaVecaConfig, WavaVecaFeatureExtractor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER, get_tests_dir sys.path.append(str(Path(__file__).parent.parent.parent.parent / '''utils''')) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_feature_extraction import CustomFeatureExtractor # noqa E402 a : Tuple = get_tests_dir('''fixtures''') a : Dict = get_tests_dir('''fixtures/dummy_feature_extractor_config.json''') a : int = get_tests_dir('''fixtures/dummy-config.json''') class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): def A ( self : Tuple ): """simple docstring""" __snake_case = 0 def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained("facebook/wav2vec2-base-960h" ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : str ): """simple docstring""" with tempfile.TemporaryDirectory() as tmpdirname: __snake_case = WavaVecaConfig() # remove feature_extractor_type to make sure config.json alone is enough to load feature processor locally __snake_case = AutoFeatureExtractor.from_pretrained(a_ ).to_dict() config_dict.pop("feature_extractor_type" ) __snake_case = WavaVecaFeatureExtractor(**a_ ) # save in new folder model_config.save_pretrained(a_ ) config.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) # make sure private variable is not incorrectly saved __snake_case = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(a_ , a_ ) def A ( self : List[Any] ): """simple docstring""" __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) def A ( self : Optional[Any] ): """simple docstring""" with self.assertRaisesRegex( a_ , "bert-base is not a local folder and is not a valid model identifier" ): __snake_case = AutoFeatureExtractor.from_pretrained("bert-base" ) def A ( self : Dict ): """simple docstring""" with self.assertRaisesRegex( a_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): __snake_case = AutoFeatureExtractor.from_pretrained(a_ , revision="aaaaaa" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaisesRegex( a_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): __snake_case = AutoFeatureExtractor.from_pretrained("hf-internal-testing/config-no-model" ) def A ( self : Tuple ): """simple docstring""" with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(a_ ): __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) # Test feature extractor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ , trust_remote_code=a_ ) self.assertEqual(reloaded_feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) def A ( self : int ): """simple docstring""" try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a_ ): AutoFeatureExtractor.register(a_ , a_ ) # Now that the config is registered, it can be used as any other config with the auto-API __snake_case = CustomFeatureExtractor.from_pretrained(a_ ) with tempfile.TemporaryDirectory() as tmp_dir: feature_extractor.save_pretrained(a_ ) __snake_case = AutoFeatureExtractor.from_pretrained(a_ ) self.assertIsInstance(a_ , a_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig] def A ( self : Dict ): """simple docstring""" class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): __SCREAMING_SNAKE_CASE = True try: AutoConfig.register("custom" , a_ ) AutoFeatureExtractor.register(a_ , a_ ) # If remote code is not set, the default is to use local __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote code is disabled, we load the local one. __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(feature_extractor.is_local ) # If remote is enabled, we load from the Hub __snake_case = AutoFeatureExtractor.from_pretrained( "hf-internal-testing/test_dynamic_feature_extractor" , trust_remote_code=a_ ) self.assertEqual(feature_extractor.__class__.__name__ , "NewFeatureExtractor" ) self.assertTrue(not hasattr(a_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in FEATURE_EXTRACTOR_MAPPING._extra_content: del FEATURE_EXTRACTOR_MAPPING._extra_content[CustomConfig]
69
1
'''simple docstring''' import argparse import json import os import numpy as np import PIL import requests import tensorflow.keras.applications.efficientnet as efficientnet import torch from huggingface_hub import hf_hub_download from PIL import Image from tensorflow.keras.preprocessing import image from transformers import ( EfficientNetConfig, EfficientNetForImageClassification, EfficientNetImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() a : Optional[Any] = logging.get_logger(__name__) a : Union[str, Any] = { '''b0''': efficientnet.EfficientNetBa, '''b1''': efficientnet.EfficientNetBa, '''b2''': efficientnet.EfficientNetBa, '''b3''': efficientnet.EfficientNetBa, '''b4''': efficientnet.EfficientNetBa, '''b5''': efficientnet.EfficientNetBa, '''b6''': efficientnet.EfficientNetBa, '''b7''': efficientnet.EfficientNetBa, } a : int = { '''b0''': { '''hidden_dim''': 1_280, '''width_coef''': 1.0, '''depth_coef''': 1.0, '''image_size''': 224, '''dropout_rate''': 0.2, '''dw_padding''': [], }, '''b1''': { '''hidden_dim''': 1_280, '''width_coef''': 1.0, '''depth_coef''': 1.1, '''image_size''': 240, '''dropout_rate''': 0.2, '''dw_padding''': [16], }, '''b2''': { '''hidden_dim''': 1_408, '''width_coef''': 1.1, '''depth_coef''': 1.2, '''image_size''': 260, '''dropout_rate''': 0.3, '''dw_padding''': [5, 8, 16], }, '''b3''': { '''hidden_dim''': 1_536, '''width_coef''': 1.2, '''depth_coef''': 1.4, '''image_size''': 300, '''dropout_rate''': 0.3, '''dw_padding''': [5, 18], }, '''b4''': { '''hidden_dim''': 1_792, '''width_coef''': 1.4, '''depth_coef''': 1.8, '''image_size''': 380, '''dropout_rate''': 0.4, '''dw_padding''': [6], }, '''b5''': { '''hidden_dim''': 2_048, '''width_coef''': 1.6, '''depth_coef''': 2.2, '''image_size''': 456, '''dropout_rate''': 0.4, '''dw_padding''': [13, 27], }, '''b6''': { '''hidden_dim''': 2_304, '''width_coef''': 1.8, '''depth_coef''': 2.6, '''image_size''': 528, '''dropout_rate''': 0.5, '''dw_padding''': [31], }, '''b7''': { '''hidden_dim''': 2_560, '''width_coef''': 2.0, '''depth_coef''': 3.1, '''image_size''': 600, '''dropout_rate''': 0.5, '''dw_padding''': [18], }, } def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> List[Any]: __snake_case = EfficientNetConfig() __snake_case = CONFIG_MAP[model_name]["hidden_dim"] __snake_case = CONFIG_MAP[model_name]["width_coef"] __snake_case = CONFIG_MAP[model_name]["depth_coef"] __snake_case = CONFIG_MAP[model_name]["image_size"] __snake_case = CONFIG_MAP[model_name]["dropout_rate"] __snake_case = CONFIG_MAP[model_name]["dw_padding"] __snake_case = "huggingface/label-files" __snake_case = "imagenet-1k-id2label.json" __snake_case = 10_00 __snake_case = json.load(open(hf_hub_download(_UpperCAmelCase , _UpperCAmelCase , repo_type="dataset" ) , "r" ) ) __snake_case = {int(_UpperCAmelCase ): v for k, v in idalabel.items()} __snake_case = idalabel __snake_case = {v: k for k, v in idalabel.items()} return config def __UpperCAmelCase ( ) -> int: __snake_case = "http://images.cocodataset.org/val2017/000000039769.jpg" __snake_case = Image.open(requests.get(_UpperCAmelCase , stream=_UpperCAmelCase ).raw ) return im def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] ) -> Tuple: __snake_case = CONFIG_MAP[model_name]["image_size"] __snake_case = EfficientNetImageProcessor( size={"height": size, "width": size} , image_mean=[0.485, 0.456, 0.406] , image_std=[0.4785_3944, 0.473_2864, 0.4743_4163] , do_center_crop=_UpperCAmelCase , ) return preprocessor def __UpperCAmelCase ( _UpperCAmelCase : List[str] ) -> str: __snake_case = [v.split("_" )[0].split("block" )[1] for v in original_param_names if v.startswith("block" )] __snake_case = sorted(set(_UpperCAmelCase ) ) __snake_case = len(_UpperCAmelCase ) __snake_case = {b: str(_UpperCAmelCase ) for b, i in zip(_UpperCAmelCase , range(_UpperCAmelCase ) )} __snake_case = [] rename_keys.append(("stem_conv/kernel:0", "embeddings.convolution.weight") ) rename_keys.append(("stem_bn/gamma:0", "embeddings.batchnorm.weight") ) rename_keys.append(("stem_bn/beta:0", "embeddings.batchnorm.bias") ) rename_keys.append(("stem_bn/moving_mean:0", "embeddings.batchnorm.running_mean") ) rename_keys.append(("stem_bn/moving_variance:0", "embeddings.batchnorm.running_var") ) for b in block_names: __snake_case = block_name_mapping[b] rename_keys.append((F'''block{b}_expand_conv/kernel:0''', F'''encoder.blocks.{hf_b}.expansion.expand_conv.weight''') ) rename_keys.append((F'''block{b}_expand_bn/gamma:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.weight''') ) rename_keys.append((F'''block{b}_expand_bn/beta:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.bias''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_mean''') ) rename_keys.append( (F'''block{b}_expand_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.expansion.expand_bn.running_var''') ) rename_keys.append( (F'''block{b}_dwconv/depthwise_kernel:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_conv.weight''') ) rename_keys.append((F'''block{b}_bn/gamma:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.weight''') ) rename_keys.append((F'''block{b}_bn/beta:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.bias''') ) rename_keys.append( (F'''block{b}_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_mean''') ) rename_keys.append( (F'''block{b}_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.depthwise_conv.depthwise_norm.running_var''') ) rename_keys.append((F'''block{b}_se_reduce/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.weight''') ) rename_keys.append((F'''block{b}_se_reduce/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.reduce.bias''') ) rename_keys.append((F'''block{b}_se_expand/kernel:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.weight''') ) rename_keys.append((F'''block{b}_se_expand/bias:0''', F'''encoder.blocks.{hf_b}.squeeze_excite.expand.bias''') ) rename_keys.append( (F'''block{b}_project_conv/kernel:0''', F'''encoder.blocks.{hf_b}.projection.project_conv.weight''') ) rename_keys.append((F'''block{b}_project_bn/gamma:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.weight''') ) rename_keys.append((F'''block{b}_project_bn/beta:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.bias''') ) rename_keys.append( (F'''block{b}_project_bn/moving_mean:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_mean''') ) rename_keys.append( (F'''block{b}_project_bn/moving_variance:0''', F'''encoder.blocks.{hf_b}.projection.project_bn.running_var''') ) rename_keys.append(("top_conv/kernel:0", "encoder.top_conv.weight") ) rename_keys.append(("top_bn/gamma:0", "encoder.top_bn.weight") ) rename_keys.append(("top_bn/beta:0", "encoder.top_bn.bias") ) rename_keys.append(("top_bn/moving_mean:0", "encoder.top_bn.running_mean") ) rename_keys.append(("top_bn/moving_variance:0", "encoder.top_bn.running_var") ) __snake_case = {} for item in rename_keys: if item[0] in original_param_names: __snake_case = "efficientnet." + item[1] __snake_case = "classifier.weight" __snake_case = "classifier.bias" return key_mapping def __UpperCAmelCase ( _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Dict ) -> Dict: for key, value in tf_params.items(): if "normalization" in key: continue __snake_case = key_mapping[key] if "_conv" in key and "kernel" in key: __snake_case = torch.from_numpy(_UpperCAmelCase ).permute(3 , 2 , 0 , 1 ) elif "depthwise_kernel" in key: __snake_case = torch.from_numpy(_UpperCAmelCase ).permute(2 , 3 , 0 , 1 ) elif "kernel" in key: __snake_case = torch.from_numpy(np.transpose(_UpperCAmelCase ) ) else: __snake_case = torch.from_numpy(_UpperCAmelCase ) # Replace HF parameters with original TF model parameters assert hf_params[hf_key].shape == new_hf_value.shape hf_params[hf_key].copy_(_UpperCAmelCase ) @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Any , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] ) -> List[str]: __snake_case = model_classes[model_name]( include_top=_UpperCAmelCase , weights="imagenet" , input_tensor=_UpperCAmelCase , input_shape=_UpperCAmelCase , pooling=_UpperCAmelCase , classes=10_00 , classifier_activation="softmax" , ) __snake_case = original_model.trainable_variables __snake_case = original_model.non_trainable_variables __snake_case = {param.name: param.numpy() for param in tf_params} for param in tf_non_train_params: __snake_case = param.numpy() __snake_case = list(tf_params.keys() ) # Load HuggingFace model __snake_case = get_efficientnet_config(_UpperCAmelCase ) __snake_case = EfficientNetForImageClassification(_UpperCAmelCase ).eval() __snake_case = hf_model.state_dict() # Create src-to-dst parameter name mapping dictionary print("Converting parameters..." ) __snake_case = rename_keys(_UpperCAmelCase ) replace_params(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # Initialize preprocessor and preprocess input image __snake_case = convert_image_processor(_UpperCAmelCase ) __snake_case = preprocessor(images=prepare_img() , return_tensors="pt" ) # HF model inference hf_model.eval() with torch.no_grad(): __snake_case = hf_model(**_UpperCAmelCase ) __snake_case = outputs.logits.detach().numpy() # Original model inference __snake_case = False __snake_case = CONFIG_MAP[model_name]["image_size"] __snake_case = prepare_img().resize((image_size, image_size) , resample=PIL.Image.NEAREST ) __snake_case = image.img_to_array(_UpperCAmelCase ) __snake_case = np.expand_dims(_UpperCAmelCase , axis=0 ) __snake_case = original_model.predict(_UpperCAmelCase ) # Check whether original and HF model outputs match -> np.allclose assert np.allclose(_UpperCAmelCase , _UpperCAmelCase , atol=1E-3 ), "The predicted logits are not the same." print("Model outputs match!" ) if save_model: # Create folder to save model if not os.path.isdir(_UpperCAmelCase ): os.mkdir(_UpperCAmelCase ) # Save converted model and image processor hf_model.save_pretrained(_UpperCAmelCase ) preprocessor.save_pretrained(_UpperCAmelCase ) if push_to_hub: # Push model and image processor to hub print(F'''Pushing converted {model_name} to the hub...''' ) __snake_case = F'''efficientnet-{model_name}''' preprocessor.push_to_hub(_UpperCAmelCase ) hf_model.push_to_hub(_UpperCAmelCase ) if __name__ == "__main__": a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--model_name''', default='''b0''', type=str, help='''Version name of the EfficientNet model you want to convert, select from [b0, b1, b2, b3, b4, b5, b6, b7].''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default='''hf_model''', type=str, help='''Path to the output PyTorch model directory.''', ) parser.add_argument('''--save_model''', action='''store_true''', help='''Save model to local''') parser.add_argument('''--push_to_hub''', action='''store_true''', help='''Push model and image processor to the hub''') a : Optional[int] = parser.parse_args() convert_efficientnet_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.save_model, args.push_to_hub)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # bit count represents no. of bits in the gray code if bit_count < 0: raise ValueError("The given input must be positive" ) # get the generated string sequence __snake_case = gray_code_sequence_string(_UpperCAmelCase ) # # convert them to integers for i in range(len(_UpperCAmelCase ) ): __snake_case = int(sequence[i] , 2 ) return sequence def __UpperCAmelCase ( _UpperCAmelCase : int ) -> list: # The approach is a recursive one # Base case achieved when either n = 0 or n=1 if bit_count == 0: return ["0"] if bit_count == 1: return ["0", "1"] __snake_case = 1 << bit_count # defines the length of the sequence # 1<< n is equivalent to 2^n # recursive answer will generate answer for n-1 bits __snake_case = gray_code_sequence_string(bit_count - 1 ) __snake_case = [] # append 0 to first half of the smaller sequence generated for i in range(seq_len // 2 ): __snake_case = "0" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) # append 1 to second half ... start from the end of the list for i in reversed(range(seq_len // 2 ) ): __snake_case = "1" + smaller_sequence[i] sequence.append(_UpperCAmelCase ) return sequence if __name__ == "__main__": import doctest doctest.testmod()
69
1
'''simple docstring''' import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : str , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : List[Any]=True , _UpperCAmelCase : List[str]="pt" ) -> List[Any]: __snake_case = {"add_prefix_space": True} if isinstance(_UpperCAmelCase , _UpperCAmelCase ) and not line.startswith(" " ) else {} __snake_case = padding_side return tokenizer( [line] , max_length=_UpperCAmelCase , padding="max_length" if pad_to_max_length else None , truncation=_UpperCAmelCase , return_tensors=_UpperCAmelCase , add_special_tokens=_UpperCAmelCase , **_UpperCAmelCase , ) def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : int=None , ) -> List[Any]: __snake_case = input_ids.ne(_UpperCAmelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase ): def __init__( self : Dict , a_ : List[str] , a_ : List[Any] , a_ : int , a_ : List[Any] , a_ : int="train" , a_ : Optional[int]=None , a_ : List[str]=None , a_ : Dict=None , a_ : str="" , ): """simple docstring""" super().__init__() __snake_case = Path(a_ ).joinpath(type_path + ".source" ) __snake_case = Path(a_ ).joinpath(type_path + ".target" ) __snake_case = self.get_char_lens(self.src_file ) __snake_case = max_source_length __snake_case = max_target_length assert min(self.src_lens ) > 0, f'''found empty line in {self.src_file}''' __snake_case = tokenizer __snake_case = prefix if n_obs is not None: __snake_case = self.src_lens[:n_obs] __snake_case = src_lang __snake_case = tgt_lang def __len__( self : Any ): """simple docstring""" return len(self.src_lens ) def __getitem__( self : Tuple , a_ : Union[str, Any] ): """simple docstring""" __snake_case = index + 1 # linecache starts at 1 __snake_case = self.prefix + linecache.getline(str(self.src_file ) , a_ ).rstrip("\n" ) __snake_case = linecache.getline(str(self.tgt_file ) , a_ ).rstrip("\n" ) assert source_line, f'''empty source line for index {index}''' assert tgt_line, f'''empty tgt line for index {index}''' # Need to add eos token manually for T5 if isinstance(self.tokenizer , a_ ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right __snake_case = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , a_ ) else self.tokenizer ) __snake_case = self.tokenizer.generator if isinstance(self.tokenizer , a_ ) else self.tokenizer __snake_case = encode_line(a_ , a_ , self.max_source_length , "right" ) __snake_case = encode_line(a_ , a_ , self.max_target_length , "right" ) __snake_case = source_inputs["input_ids"].squeeze() __snake_case = target_inputs["input_ids"].squeeze() __snake_case = source_inputs["attention_mask"].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def A ( a_ : List[str] ): """simple docstring""" return [len(a_ ) for x in Path(a_ ).open().readlines()] def A ( self : List[Any] , a_ : str ): """simple docstring""" __snake_case = torch.stack([x["input_ids"] for x in batch] ) __snake_case = torch.stack([x["attention_mask"] for x in batch] ) __snake_case = torch.stack([x["decoder_input_ids"] for x in batch] ) __snake_case = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , a_ ) else self.tokenizer.pad_token_id ) __snake_case = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , a_ ) else self.tokenizer.pad_token_id ) __snake_case = trim_batch(a_ , a_ ) __snake_case , __snake_case = trim_batch(a_ , a_ , attention_mask=a_ ) __snake_case = { "input_ids": source_ids, "attention_mask": source_mask, "decoder_input_ids": y, } return batch a : int = getLogger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : List[List] ) -> Tuple: return list(itertools.chain.from_iterable(_UpperCAmelCase ) ) def __UpperCAmelCase ( _UpperCAmelCase : str ) -> None: __snake_case = get_git_info() save_json(_UpperCAmelCase , os.path.join(_UpperCAmelCase , "git_log.json" ) ) def __UpperCAmelCase ( _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : List[str]=4 , **_UpperCAmelCase : str ) -> Optional[Any]: with open(_UpperCAmelCase , "w" ) as f: json.dump(_UpperCAmelCase , _UpperCAmelCase , indent=_UpperCAmelCase , **_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> List[str]: with open(_UpperCAmelCase ) as f: return json.load(_UpperCAmelCase ) def __UpperCAmelCase ( ) -> int: __snake_case = git.Repo(search_parent_directories=_UpperCAmelCase ) __snake_case = { "repo_id": str(_UpperCAmelCase ), "repo_sha": str(repo.head.object.hexsha ), "repo_branch": str(repo.active_branch ), "hostname": str(socket.gethostname() ), } return repo_infos def __UpperCAmelCase ( _UpperCAmelCase : Callable , _UpperCAmelCase : Iterable ) -> List: return list(map(_UpperCAmelCase , _UpperCAmelCase ) ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple ) -> Dict: with open(_UpperCAmelCase , "wb" ) as f: return pickle.dump(_UpperCAmelCase , _UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[int] ) -> List[str]: def remove_articles(_UpperCAmelCase : int ): return re.sub(R"\b(a|an|the)\b" , " " , _UpperCAmelCase ) def white_space_fix(_UpperCAmelCase : int ): return " ".join(text.split() ) def remove_punc(_UpperCAmelCase : Optional[int] ): __snake_case = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_UpperCAmelCase : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_UpperCAmelCase ) ) ) ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Tuple ) -> int: __snake_case = normalize_answer(_UpperCAmelCase ).split() __snake_case = normalize_answer(_UpperCAmelCase ).split() __snake_case = Counter(_UpperCAmelCase ) & Counter(_UpperCAmelCase ) __snake_case = sum(common.values() ) if num_same == 0: return 0 __snake_case = 1.0 * num_same / len(_UpperCAmelCase ) __snake_case = 1.0 * num_same / len(_UpperCAmelCase ) __snake_case = (2 * precision * recall) / (precision + recall) return fa def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Tuple ) -> List[str]: return normalize_answer(_UpperCAmelCase ) == normalize_answer(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : List[str] ) -> Dict: assert len(_UpperCAmelCase ) == len(_UpperCAmelCase ) __snake_case = 0 for hypo, pred in zip(_UpperCAmelCase , _UpperCAmelCase ): em += exact_match_score(_UpperCAmelCase , _UpperCAmelCase ) if len(_UpperCAmelCase ) > 0: em /= len(_UpperCAmelCase ) return {"em": em} def __UpperCAmelCase ( _UpperCAmelCase : Tuple ) -> Union[str, Any]: return model_prefix.startswith("rag" ) def __UpperCAmelCase ( _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Any , _UpperCAmelCase : int ) -> int: __snake_case = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead __snake_case = "dropout_rate" for p in extra_params: if getattr(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): if not hasattr(_UpperCAmelCase , _UpperCAmelCase ) and not hasattr(_UpperCAmelCase , equivalent_param[p] ): logger.info("config doesn't have a `{}` attribute".format(_UpperCAmelCase ) ) delattr(_UpperCAmelCase , _UpperCAmelCase ) continue __snake_case = p if hasattr(_UpperCAmelCase , _UpperCAmelCase ) else equivalent_param[p] setattr(_UpperCAmelCase , _UpperCAmelCase , getattr(_UpperCAmelCase , _UpperCAmelCase ) ) delattr(_UpperCAmelCase , _UpperCAmelCase ) return hparams, config
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
1
'''simple docstring''' from __future__ import annotations def __UpperCAmelCase ( _UpperCAmelCase : list ) -> list: if len(_UpperCAmelCase ) == 0: return [] __snake_case , __snake_case = min(_UpperCAmelCase ), max(_UpperCAmelCase ) __snake_case = int(max_value - min_value ) + 1 __snake_case = [[] for _ in range(_UpperCAmelCase )] for i in my_list: buckets[int(i - min_value )].append(_UpperCAmelCase ) return [v for bucket in buckets for v in sorted(_UpperCAmelCase )] if __name__ == "__main__": from doctest import testmod testmod() assert bucket_sort([4, 5, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bucket_sort([0, 1, -10, 15, 2, -2]) == [-10, -2, 0, 1, 2, 15]
69
'''simple docstring''' a : Dict = range(2, 20 + 1) a : Optional[int] = [10**k for k in range(ks[-1] + 1)] a : dict[int, dict[int, list[list[int]]]] = {} def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Dict , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[Any] ) -> int: __snake_case = sum(a_i[j] for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ) ) __snake_case = sum(a_i[j] * base[j] for j in range(min(len(_UpperCAmelCase ) , _UpperCAmelCase ) ) ) __snake_case , __snake_case = 0, 0 __snake_case = n - i __snake_case = memo.get(_UpperCAmelCase ) if sub_memo is not None: __snake_case = sub_memo.get(_UpperCAmelCase ) if jumps is not None and len(_UpperCAmelCase ) > 0: # find and make the largest jump without going over __snake_case = -1 for _k in range(len(_UpperCAmelCase ) - 1 , -1 , -1 ): if jumps[_k][2] <= k and jumps[_k][1] <= max_dn: __snake_case = _k break if max_jump >= 0: __snake_case , __snake_case , __snake_case = jumps[max_jump] # since the difference between jumps is cached, add c __snake_case = diff + c for j in range(min(_UpperCAmelCase , len(_UpperCAmelCase ) ) ): __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) if new_c > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) else: __snake_case = [] else: __snake_case = {c: []} __snake_case = sub_memo if dn >= max_dn or c + diff >= base[k]: return diff, dn if k > ks[0]: while True: # keep doing smaller jumps __snake_case , __snake_case = next_term(_UpperCAmelCase , k - 1 , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped if dn >= max_dn or c + diff >= base[k]: break else: # would be too small a jump, just compute sequential terms instead __snake_case , __snake_case = compute(_UpperCAmelCase , _UpperCAmelCase , i + dn , _UpperCAmelCase ) diff += _diff dn += terms_jumped __snake_case = sub_memo[c] # keep jumps sorted by # of terms skipped __snake_case = 0 while j < len(_UpperCAmelCase ): if jumps[j][1] > dn: break j += 1 # cache the jump for this value digitsum(b) and c sub_memo[c].insert(_UpperCAmelCase , (diff, dn, k) ) return (diff, dn) def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Optional[Any] , _UpperCAmelCase : Optional[int] ) -> Optional[int]: if i >= n: return 0, i if k > len(_UpperCAmelCase ): a_i.extend([0 for _ in range(k - len(_UpperCAmelCase ) )] ) # note: a_i -> b * 10^k + c # ds_b -> digitsum(b) # ds_c -> digitsum(c) __snake_case = i __snake_case , __snake_case , __snake_case = 0, 0, 0 for j in range(len(_UpperCAmelCase ) ): if j >= k: ds_b += a_i[j] else: ds_c += a_i[j] while i < n: i += 1 __snake_case = ds_c + ds_b diff += addend __snake_case = 0 for j in range(_UpperCAmelCase ): __snake_case = a_i[j] + addend __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) ds_c += a_i[j] if addend > 0: break if addend > 0: add(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) return diff, i - start_i def __UpperCAmelCase ( _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : Tuple , _UpperCAmelCase : str ) -> Tuple: for j in range(_UpperCAmelCase , len(_UpperCAmelCase ) ): __snake_case = digits[j] + addend if s >= 10: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) __snake_case = addend // 10 + quotient else: __snake_case = s __snake_case = addend // 10 if addend == 0: break while addend > 0: __snake_case , __snake_case = divmod(_UpperCAmelCase , 10 ) digits.append(_UpperCAmelCase ) def __UpperCAmelCase ( _UpperCAmelCase : int = 10**15 ) -> int: __snake_case = [1] __snake_case = 1 __snake_case = 0 while True: __snake_case , __snake_case = next_term(_UpperCAmelCase , 20 , i + dn , _UpperCAmelCase ) dn += terms_jumped if dn == n - i: break __snake_case = 0 for j in range(len(_UpperCAmelCase ) ): a_n += digits[j] * 10**j return a_n if __name__ == "__main__": print(F'''{solution() = }''')
69
1
'''simple docstring''' import argparse from pathlib import Path import torch from transformers import OPTConfig, OPTModel from transformers.utils import logging logging.set_verbosity_info() a : List[str] = logging.get_logger(__name__) def __UpperCAmelCase ( _UpperCAmelCase : Dict ) -> Union[str, Any]: __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" ) if "model" in sd.keys(): __snake_case = torch.load(_UpperCAmelCase , map_location="cpu" )["model"] # pop unnecessary weights __snake_case = [ "decoder.version", "decoder.output_projection.weight", ] for key in keys_to_delete: if key in sd: sd.pop(_UpperCAmelCase ) __snake_case = { "decoder.project_in_dim.weight": "decoder.project_in.weight", "decoder.project_out_dim.weight": "decoder.project_out.weight", "decoder.layer_norm.weight": "decoder.final_layer_norm.weight", "decoder.layer_norm.bias": "decoder.final_layer_norm.bias", } for old_key, new_key in keys_to_rename.items(): if old_key in sd: __snake_case = sd.pop(_UpperCAmelCase ) __snake_case = list(sd.keys() ) for key in keys: if ".qkv_proj." in key: __snake_case = sd[key] # We split QKV in separate Q,K,V __snake_case = key.replace(".qkv_proj." , ".q_proj." ) __snake_case = key.replace(".qkv_proj." , ".k_proj." ) __snake_case = key.replace(".qkv_proj." , ".v_proj." ) __snake_case = value.shape[0] assert depth % 3 == 0 # `SequeuceParallelTransformerBlock` has QKV weight is separated in K,V,Q despite the naming: # https://cs.github.com/facebookresearch/metaseq/blob/51871bd73cd04c038f239ea2a26db1d7f6b37927/metaseq/modules/sequence_parallel_transformer_layer.py#L97 __snake_case , __snake_case , __snake_case = torch.split(_UpperCAmelCase , depth // 3 , dim=0 ) __snake_case = q __snake_case = k __snake_case = v del sd[key] return sd @torch.no_grad() def __UpperCAmelCase ( _UpperCAmelCase : List[str] , _UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : int=None ) -> Any: __snake_case = load_checkpoint(_UpperCAmelCase ) if config is not None: __snake_case = OPTConfig.from_pretrained(_UpperCAmelCase ) else: __snake_case = OPTConfig() __snake_case = OPTModel(_UpperCAmelCase ).half().eval() model.load_state_dict(_UpperCAmelCase ) # Check results Path(_UpperCAmelCase ).mkdir(exist_ok=_UpperCAmelCase ) model.save_pretrained(_UpperCAmelCase ) if __name__ == "__main__": a : int = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--fairseq_path''', type=str, help=( '''path to fairseq checkpoint in correct format. You can find all checkpoints in the correct format here:''' ''' https://huggingface.co/models?other=opt_metasq''' ), ) parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--hf_config''', default=None, type=str, help='''Define HF config.''') a : Optional[int] = parser.parse_args() convert_opt_checkpoint(args.fairseq_path, args.pytorch_dump_folder_path, config=args.hf_config)
69
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : List[Any]=2_81_23 ) -> str: __snake_case = [1] * (limit + 1) for i in range(2 , int(limit**0.5 ) + 1 ): sum_divs[i * i] += i for k in range(i + 1 , limit // i + 1 ): sum_divs[k * i] += k + i __snake_case = set() __snake_case = 0 for n in range(1 , limit + 1 ): if sum_divs[n] > n: abundants.add(_UpperCAmelCase ) if not any((n - a in abundants) for a in abundants ): res += n return res if __name__ == "__main__": print(solution())
69
1
'''simple docstring''' def __UpperCAmelCase ( _UpperCAmelCase : str , _UpperCAmelCase : str ) -> list: __snake_case = len(_UpperCAmelCase ) __snake_case = [] for i in range(len(_UpperCAmelCase ) - pat_len + 1 ): __snake_case = True for j in range(_UpperCAmelCase ): if s[i + j] != pattern[j]: __snake_case = False break if match_found: position.append(_UpperCAmelCase ) return position if __name__ == "__main__": assert naive_pattern_search('''ABCDEFG''', '''DE''') == [3] print(naive_pattern_search('''ABAAABCDBBABCDDEBCABC''', '''ABC'''))
69
'''simple docstring''' import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class SCREAMING_SNAKE_CASE__ : def __init__( self : str , a_ : List[str] , a_ : Tuple=3 , a_ : Any=7 , a_ : Any=True , a_ : Union[str, Any]=True , a_ : Tuple=False , a_ : Optional[int]=True , a_ : Any=99 , a_ : Dict=32 , a_ : Dict=5 , a_ : List[Any]=4 , a_ : Any=37 , a_ : Any="gelu" , a_ : List[str]=0.1 , a_ : Dict=0.1 , a_ : Optional[Any]=512 , a_ : List[Any]=16 , a_ : Any=2 , a_ : str=0.02 , a_ : Any=3 , a_ : List[Any]=4 , a_ : List[str]=None , ): """simple docstring""" __snake_case = parent __snake_case = batch_size __snake_case = seq_length __snake_case = is_training __snake_case = use_input_mask __snake_case = use_token_type_ids __snake_case = use_labels __snake_case = vocab_size __snake_case = hidden_size __snake_case = num_hidden_layers __snake_case = num_attention_heads __snake_case = intermediate_size __snake_case = hidden_act __snake_case = hidden_dropout_prob __snake_case = attention_probs_dropout_prob __snake_case = max_position_embeddings __snake_case = type_vocab_size __snake_case = type_sequence_label_size __snake_case = initializer_range __snake_case = num_labels __snake_case = num_choices __snake_case = scope def A ( self : Any ): """simple docstring""" __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __snake_case = None if self.use_input_mask: __snake_case = random_attention_mask([self.batch_size, self.seq_length] ) __snake_case = None __snake_case = None __snake_case = None __snake_case = None if self.use_labels: __snake_case = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __snake_case = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __snake_case = ids_tensor([self.batch_size] , self.num_choices ) __snake_case = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A ( self : Optional[int] ): """simple docstring""" return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=a_ , ) def A ( self : List[str] , a_ : Dict , a_ : Tuple , a_ : Optional[Any] , a_ : Dict , a_ : Dict , a_ : Dict , a_ : Union[str, Any] ): """simple docstring""" __snake_case = FalconModel(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ ) __snake_case = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : List[Any] , a_ : List[Any] , a_ : Union[str, Any] , a_ : Optional[Any] , a_ : Any , a_ : List[Any] , a_ : Optional[Any] , a_ : Union[str, Any] , a_ : Tuple , a_ : Optional[int] , ): """simple docstring""" __snake_case = True __snake_case = FalconModel(a_ ) model.to(a_ ) model.eval() __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , ) __snake_case = model(a_ , attention_mask=a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A ( self : Optional[int] , a_ : int , a_ : int , a_ : List[Any] , a_ : str , a_ : List[str] , a_ : str , a_ : str , a_ : Union[str, Any] , a_ : Optional[int] , ): """simple docstring""" __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A ( self : List[Any] , a_ : Optional[int] , a_ : Optional[Any] , a_ : str , a_ : Tuple , a_ : str , a_ : List[Any] , a_ : Optional[Any] , a_ : Any , a_ : Dict , ): """simple docstring""" __snake_case = True __snake_case = True __snake_case = FalconForCausalLM(config=a_ ) model.to(a_ ) model.eval() # first forward pass __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , use_cache=a_ , ) __snake_case = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids __snake_case = ids_tensor((self.batch_size, 3) , config.vocab_size ) __snake_case = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and __snake_case = torch.cat([input_ids, next_tokens] , dim=-1 ) __snake_case = torch.cat([input_mask, next_mask] , dim=-1 ) __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , output_hidden_states=a_ , )["hidden_states"][0] __snake_case = model( a_ , attention_mask=a_ , encoder_hidden_states=a_ , encoder_attention_mask=a_ , past_key_values=a_ , output_hidden_states=a_ , )["hidden_states"][0] # select random slice __snake_case = ids_tensor((1,) , output_from_past.shape[-1] ).item() __snake_case = output_from_no_past[:, -3:, random_slice_idx].detach() __snake_case = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(a_ , a_ , atol=1e-3 ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case = self.prepare_config_and_inputs() ( ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ( __snake_case ) , ) = config_and_inputs __snake_case = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class SCREAMING_SNAKE_CASE__ ( _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , unittest.TestCase ): __SCREAMING_SNAKE_CASE = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __SCREAMING_SNAKE_CASE = (FalconForCausalLM,) if is_torch_available() else () __SCREAMING_SNAKE_CASE = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __SCREAMING_SNAKE_CASE = False __SCREAMING_SNAKE_CASE = False def A ( self : Optional[Any] ): """simple docstring""" __snake_case = FalconModelTester(self ) __snake_case = ConfigTester(self , config_class=a_ , hidden_size=37 ) def A ( self : Optional[Any] ): """simple docstring""" self.config_tester.run_common_tests() def A ( self : List[Any] ): """simple docstring""" __snake_case = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def A ( self : List[str] ): """simple docstring""" __snake_case , *__snake_case = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: __snake_case = alibi self.model_tester.create_and_check_model(a_ , *a_ ) def A ( self : Tuple ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Union[str, Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "single_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = input_dict["input_ids"] __snake_case = FalconForCausalLM(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , use_cache=a_ ) __snake_case = input_ids.shape[0] __snake_case = model._convert_to_rw_cache(result.past_key_values ) __snake_case = model._convert_cache_to_standard_format(a_ , a_ ) for layer in range(len(a_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def A ( self : Optional[Any] ): """simple docstring""" __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() __snake_case = 3 __snake_case = "multi_label_classification" __snake_case = input_dict["input_ids"] __snake_case = input_ids.ne(1 ).to(a_ ) __snake_case = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) __snake_case = FalconForSequenceClassification(a_ ) model.to(a_ ) model.eval() __snake_case = model(a_ , attention_mask=a_ , labels=a_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def A ( self : Dict ): """simple docstring""" for model_class in self.all_generative_model_classes: __snake_case , __snake_case = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(a_ , "use_cache" ): return __snake_case = model_class(a_ ).to(a_ ) if "use_cache" not in inputs: __snake_case = True __snake_case = model(**a_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return __snake_case = ( getattr(a_ , "decoder_layers" , a_ ) or getattr(a_ , "num_decoder_layers" , a_ ) or config.num_hidden_layers ) __snake_case = getattr(a_ , "num_kv_heads" , config.num_attention_heads ) __snake_case = getattr(a_ , "d_model" , config.hidden_size ) __snake_case = embed_dim // num_attention_heads __snake_case = outputs["past_key_values"] self.assertEqual(len(a_ ) , a_ ) __snake_case , __snake_case = inputs["input_ids"].shape for i in range(a_ ): if config.new_decoder_architecture: __snake_case = config.num_attention_heads elif config.multi_query: __snake_case = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): @slow def A ( self : Any ): """simple docstring""" __snake_case = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) __snake_case = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) __snake_case = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=19 ) __snake_case = tokenizer.batch_decode(a_ )[0] self.assertEqual(a_ , a_ ) @slow def A ( self : Optional[int] ): """simple docstring""" for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , do_sample=a_ , max_new_tokens=4 ) model.generate(**a_ , num_beams=2 , max_new_tokens=4 ) @slow def A ( self : Any ): """simple docstring""" with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: __snake_case = AutoTokenizer.from_pretrained(a_ ) __snake_case = FalconForCausalLM.from_pretrained(a_ ) model.eval() model.to(device=a_ ) __snake_case = tokenizer("My favorite food is" , return_tensors="pt" ).to(a_ ) # Test results are the same with and without cache __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) __snake_case = model.generate(**a_ , do_sample=a_ , max_new_tokens=20 , use_cache=a_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
69
1