code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
import os def _snake_case( ) -> str: '''simple docstring''' with open(os.path.dirname(_a ) + '/p022_names.txt' ) as file: A__ = str(file.readlines()[0] ) A__ = names.replace('\"' , '' ).split(',' ) names.sort() A__ = 0 A__ = 0 for i, name in enumerate(_a ): for letter in name: name_score += ord(_a ) - 64 total_score += (i + 1) * name_score A__ = 0 return total_score if __name__ == "__main__": print(solution())
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
from functools import lru_cache def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> List[str]: '''simple docstring''' UpperCAmelCase = 2 UpperCAmelCase = set() while i * i <= n: if n % i: i += 1 else: n //= i factors.add(_a ) if n > 1: factors.add(_a ) return factors @lru_cache def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Any: '''simple docstring''' return len(unique_prime_factors(_a ) ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> List[str]: '''simple docstring''' return len(set(_a ) ) in (0, 1) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> List[str]: '''simple docstring''' UpperCAmelCase = 2 while True: # Increment each value of a generated range UpperCAmelCase = [base + i for i in range(_a )] # Run elements through out unique_prime_factors function # Append our target number to the end. UpperCAmelCase = [upf_len(_a ) for x in group] checker.append(_a ) # If all numbers in the list are equal, return the group variable. if equality(_a ): return group # Increment our base variable by 1 base += 1 def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ = 4 ) -> Dict: '''simple docstring''' UpperCAmelCase = run(_a ) return results[0] if len(_a ) else None if __name__ == "__main__": print(solution())
273
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a_ = logging.getLogger(__name__) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = False def __UpperCamelCase ( self : str , a : str , a : Optional[int] , a : Any , a : str ) -> List[Any]: """simple docstring""" if not self.initialized: SCREAMING_SNAKE_CASE : List[str] = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[int] = True def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def __UpperCamelCase ( self : Optional[Any] , a : List[Any] , a : Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Dict = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : Tuple , a : Any , a : Tuple , a : Tuple , a : Tuple , a : List[Any]=None ) -> Optional[int]: """simple docstring""" if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def __UpperCamelCase ( self : Any ) -> Dict: """simple docstring""" logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __UpperCamelCase ( self : Tuple , a : Optional[int] , a : Any ) -> int: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. SCREAMING_SNAKE_CASE : Optional[Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = ray.get(random_worker.retrieve.remote(a , a ) ) else: SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Any = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def __UpperCamelCase ( cls : str , a : Optional[Any] , a : Any=None , **a : List[Any] ) -> str: """simple docstring""" return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def __UpperCamelCase ( cls : Union[str, Any] , a : int , a : Any , a : List[Any]=None , **a : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : str = kwargs.pop("config" , a ) or RagConfig.from_pretrained(a , **a ) SCREAMING_SNAKE_CASE : List[Any] = RagTokenizer.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.question_encoder SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.generator if indexed_dataset is not None: SCREAMING_SNAKE_CASE : str = "custom" SCREAMING_SNAKE_CASE : List[Any] = CustomHFIndex(config.retrieval_vector_size , a ) else: SCREAMING_SNAKE_CASE : List[str] = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
76
0
from __future__ import annotations def UpperCamelCase ( _A ): """simple docstring""" for i in range(1, len(matrix[0] ) ): matrix[0][i] += matrix[0][i - 1] # preprocessing the first column for i in range(1, len(_a ) ): matrix[i][0] += matrix[i - 1][0] # updating the path cost for current position for i in range(1, len(_a ) ): for j in range(1, len(matrix[0] ) ): matrix[i][j] += min(matrix[i - 1][j], matrix[i][j - 1] ) return matrix[-1][-1] if __name__ == "__main__": import doctest doctest.testmod()
342
from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : Dict , a : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : int = data SCREAMING_SNAKE_CASE : int = None def __repr__( self : str ) -> str: """simple docstring""" return F"Node({self.data})" class _UpperCamelCase : '''simple docstring''' def __init__( self : List[str] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = None def __iter__( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.head while node: yield node.data SCREAMING_SNAKE_CASE : List[str] = node.next def __len__( self : str ) -> int: """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return "->".join([str(a ) for item in self] ) def __getitem__( self : List[Any] , a : int ) -> Any: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Tuple , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) SCREAMING_SNAKE_CASE : str = self.head for _ in range(a ): SCREAMING_SNAKE_CASE : str = current.next SCREAMING_SNAKE_CASE : Any = data def __UpperCamelCase ( self : List[str] , a : Any ) -> None: """simple docstring""" self.insert_nth(len(self ) , a ) def __UpperCamelCase ( self : Union[str, Any] , a : Any ) -> None: """simple docstring""" self.insert_nth(0 , a ) def __UpperCamelCase ( self : Optional[Any] , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) SCREAMING_SNAKE_CASE : Any = Node(a ) if self.head is None: SCREAMING_SNAKE_CASE : Optional[int] = new_node elif index == 0: SCREAMING_SNAKE_CASE : Optional[int] = self.head # link new_node to head SCREAMING_SNAKE_CASE : List[Any] = new_node else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : int = new_node def __UpperCamelCase ( self : Optional[int] ) -> None: # print every node data """simple docstring""" print(self ) def __UpperCamelCase ( self : int ) -> Any: """simple docstring""" return self.delete_nth(0 ) def __UpperCamelCase ( self : Any ) -> Any: # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def __UpperCamelCase ( self : List[str] , a : int = 0 ) -> Any: """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) SCREAMING_SNAKE_CASE : Tuple = self.head # default first node if index == 0: SCREAMING_SNAKE_CASE : List[str] = self.head.next else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Any = temp.next SCREAMING_SNAKE_CASE : List[Any] = temp.next SCREAMING_SNAKE_CASE : List[str] = temp.next.next return delete_node.data def __UpperCamelCase ( self : List[Any] ) -> bool: """simple docstring""" return self.head is None def __UpperCamelCase ( self : Optional[int] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : str = self.head while current: # Store the current node's next node. SCREAMING_SNAKE_CASE : Any = current.next # Make the current node's next point backwards SCREAMING_SNAKE_CASE : List[Any] = prev # Make the previous node be the current node SCREAMING_SNAKE_CASE : Any = current # Make the current node the next node (to progress iteration) SCREAMING_SNAKE_CASE : str = next_node # Return prev in order to put the head at the end SCREAMING_SNAKE_CASE : Optional[Any] = prev def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_a) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10): assert len(_a) == i linked_list.insert_nth(_a , i + 1) assert str(_a) == "->".join(str(_a) for i in range(1 , 11)) linked_list.insert_head(0) linked_list.insert_tail(11) assert str(_a) == "->".join(str(_a) for i in range(0 , 12)) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9) == 10 assert linked_list.delete_tail() == 11 assert len(_a) == 9 assert str(_a) == "->".join(str(_a) for i in range(1 , 10)) assert all(linked_list[i] == i + 1 for i in range(0 , 9)) is True for i in range(0 , 9): SCREAMING_SNAKE_CASE : str = -i assert all(linked_list[i] == -i for i in range(0 , 9)) is True linked_list.reverse() assert str(_a) == "->".join(str(_a) for i in range(-8 , 1)) def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Optional[Any] = [ -9, 100, Node(77345112), "dlrow olleH", 7, 5555, 0, -192.5_5555, "Hello, world!", 77.9, Node(10), None, None, 12.20, ] SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in test_input: linked_list.insert_tail(_a) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_a) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head SCREAMING_SNAKE_CASE : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail SCREAMING_SNAKE_CASE : Any = linked_list.delete_tail() assert result == 12.2 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list SCREAMING_SNAKE_CASE : Any = linked_list.delete_nth(10) assert result is None assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!")) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_a) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_a) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowerCamelCase__ ( ): from doctest import testmod testmod() SCREAMING_SNAKE_CASE : Optional[int] = LinkedList() linked_list.insert_head(input("Inserting 1st at head ").strip()) linked_list.insert_head(input("Inserting 2nd at head ").strip()) print("\nPrint list:") linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail ").strip()) linked_list.insert_tail(input("Inserting 2nd at tail ").strip()) print("\nPrint list:") linked_list.print_list() print("\nDelete head") linked_list.delete_head() print("Delete tail") linked_list.delete_tail() print("\nPrint list:") linked_list.print_list() print("\nReverse linked list") linked_list.reverse() print("\nPrint list:") linked_list.print_list() print("\nString representation of linked list:") print(_a) print("\nReading/changing Node data using indexing:") print(f"Element at Position 1: {linked_list[1]}") SCREAMING_SNAKE_CASE : Dict = input("Enter New Value: ").strip() print("New list:") print(_a) print(f"length of linked_list is : {len(_a)}") if __name__ == "__main__": main()
76
0
'''simple docstring''' from ..utils import DummyObject, requires_backends class lowercase_ ( metaclass=__A ): __UpperCAmelCase = ['speech'] def __init__( self , *a , **a ): requires_backends(self , ["speech"] ) class lowercase_ ( metaclass=__A ): __UpperCAmelCase = ['speech'] def __init__( self , *a , **a ): requires_backends(self , ["speech"] )
80
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ = get_logger(__name__) class _UpperCamelCase ( enum.Enum ): '''simple docstring''' lowerCamelCase__ ='all_checks' lowerCamelCase__ ='basic_checks' lowerCamelCase__ ='no_checks' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a , _a=None): if expected_checksums is None: logger.info("Unable to verify checksums.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedDownloadedFile(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE : Tuple = " for " + verification_name if verification_name is not None else "" if len(_a) > 0: raise NonMatchingChecksumError( f"Checksums didn't match{for_verification_name}:\n" f"{bad_urls}\n" "Set `verification_mode='no_checks'` to skip checksums verification and ignore this error") logger.info("All the checksums matched successfully" + for_verification_name) class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a): if expected_splits is None: logger.info("Unable to verify splits sizes.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreSplits(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedSplits(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : List[str] = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_a) > 0: raise NonMatchingSplitsSizesError(str(_a)) logger.info("All the splits matched successfully.") def lowerCamelCase__ ( _a , _a = True): if record_checksum: SCREAMING_SNAKE_CASE : List[str] = shaaaa() with open(_a , "rb") as f: for chunk in iter(lambda: f.read(1 << 20) , b""): m.update(_a) SCREAMING_SNAKE_CASE : Optional[int] = m.hexdigest() else: SCREAMING_SNAKE_CASE : List[str] = None return {"num_bytes": os.path.getsize(_a), "checksum": checksum} def lowerCamelCase__ ( _a): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
76
0
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder __snake_case = """base_with_context""" def _A ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : str ): UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(weights['''token_embedder''']['''embedding'''] ) ) UpperCamelCase :Optional[Any] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=_a ) for lyr_num, lyr in enumerate(model.encoders ): UpperCamelCase :List[Any] = weights[F'''layers_{lyr_num}'''] UpperCamelCase :Any = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) UpperCamelCase :Tuple = ly_weight["attention"] UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) UpperCamelCase :int = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) UpperCamelCase :Any = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) UpperCamelCase :int = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) UpperCamelCase :List[str] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def _A ( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : List[str] ): UpperCamelCase :int = nn.Parameter(torch.FloatTensor(weights['''input_proj''']['''kernel'''].T ) ) UpperCamelCase :Any = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=_a ) for lyr_num, lyr in enumerate(model.encoders ): UpperCamelCase :Optional[Any] = weights[F'''layers_{lyr_num}'''] UpperCamelCase :Union[str, Any] = ly_weight["attention"] UpperCamelCase :Any = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) UpperCamelCase :List[str] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) UpperCamelCase :int = nn.Parameter( torch.FloatTensor(ly_weight['''pre_attention_layer_norm''']['''scale'''] ) ) UpperCamelCase :Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) UpperCamelCase :int = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) UpperCamelCase :str = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) UpperCamelCase :str = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(weights['''encoder_norm''']['''scale'''] ) ) return model def _A ( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Any ): UpperCamelCase :Optional[Any] = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense0''']['''kernel'''].T ) ) UpperCamelCase :Tuple = nn.Parameter(torch.FloatTensor(weights['''time_emb_dense1''']['''kernel'''].T ) ) UpperCamelCase :List[Any] = nn.Parameter( torch.FloatTensor(weights['''Embed_0''']['''embedding'''] ) , requires_grad=_a ) UpperCamelCase :Dict = nn.Parameter( torch.FloatTensor(weights['''continuous_inputs_projection''']['''kernel'''].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCamelCase :Optional[int] = weights[F'''layers_{lyr_num}'''] UpperCamelCase :Optional[Any] = nn.Parameter( torch.FloatTensor(ly_weight['''pre_self_attention_layer_norm''']['''scale'''] ) ) UpperCamelCase :Any = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_0''']['''DenseGeneral_0''']['''kernel'''].T ) ) UpperCamelCase :Optional[int] = ly_weight["self_attention"] UpperCamelCase :str = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) UpperCamelCase :Optional[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) UpperCamelCase :Optional[int] = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) UpperCamelCase :List[Any] = ly_weight["MultiHeadDotProductAttention_0"] UpperCamelCase :List[Any] = nn.Parameter(torch.FloatTensor(attention_weights['''query''']['''kernel'''].T ) ) UpperCamelCase :Tuple = nn.Parameter(torch.FloatTensor(attention_weights['''key''']['''kernel'''].T ) ) UpperCamelCase :Any = nn.Parameter(torch.FloatTensor(attention_weights['''value''']['''kernel'''].T ) ) UpperCamelCase :str = nn.Parameter(torch.FloatTensor(attention_weights['''out''']['''kernel'''].T ) ) UpperCamelCase :Dict = nn.Parameter( torch.FloatTensor(ly_weight['''pre_cross_attention_layer_norm''']['''scale'''] ) ) UpperCamelCase :int = nn.Parameter(torch.FloatTensor(ly_weight['''pre_mlp_layer_norm''']['''scale'''] ) ) UpperCamelCase :str = nn.Parameter( torch.FloatTensor(ly_weight['''FiLMLayer_1''']['''DenseGeneral_0''']['''kernel'''].T ) ) UpperCamelCase :Tuple = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_0''']['''kernel'''].T ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wi_1''']['''kernel'''].T ) ) UpperCamelCase :Optional[Any] = nn.Parameter(torch.FloatTensor(ly_weight['''mlp''']['''wo''']['''kernel'''].T ) ) UpperCamelCase :Any = nn.Parameter(torch.FloatTensor(weights['''decoder_norm''']['''scale'''] ) ) UpperCamelCase :Union[str, Any] = nn.Parameter(torch.FloatTensor(weights['''spec_out_dense''']['''kernel'''].T ) ) return model def _A ( SCREAMING_SNAKE_CASE__ : List[Any] ): UpperCamelCase :Any = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCamelCase :Optional[int] = jnp.tree_util.tree_map(onp.array , _a ) UpperCamelCase :Dict = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] UpperCamelCase :int = os.path.join(args.checkpoint_path , '''..''' , '''config.gin''' ) UpperCamelCase :List[str] = inference.parse_training_gin_file(_a , _a ) UpperCamelCase :str = inference.InferenceModel(args.checkpoint_path , _a ) UpperCamelCase :Any = DDPMScheduler(beta_schedule='''squaredcos_cap_v2''' , variance_type='''fixed_large''' ) UpperCamelCase :List[Any] = SpectrogramNotesEncoder( max_length=synth_model.sequence_length['''inputs'''] , vocab_size=synth_model.model.module.config.vocab_size , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) UpperCamelCase :Dict = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims , targets_context_length=synth_model.sequence_length['''targets_context'''] , d_model=synth_model.model.module.config.emb_dim , dropout_rate=synth_model.model.module.config.dropout_rate , num_layers=synth_model.model.module.config.num_encoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , feed_forward_proj='''gated-gelu''' , ) UpperCamelCase :Union[str, Any] = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims , targets_length=synth_model.sequence_length['''targets_context'''] , max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time , d_model=synth_model.model.module.config.emb_dim , num_layers=synth_model.model.module.config.num_decoder_layers , num_heads=synth_model.model.module.config.num_heads , d_kv=synth_model.model.module.config.head_dim , d_ff=synth_model.model.module.config.mlp_dim , dropout_rate=synth_model.model.module.config.dropout_rate , ) UpperCamelCase :Optional[int] = load_notes_encoder(ta_checkpoint['''target''']['''token_encoder'''] , _a ) UpperCamelCase :List[str] = load_continuous_encoder(ta_checkpoint['''target''']['''continuous_encoder'''] , _a ) UpperCamelCase :Union[str, Any] = load_decoder(ta_checkpoint['''target''']['''decoder'''] , _a ) UpperCamelCase :Tuple = OnnxRuntimeModel.from_pretrained('''kashif/soundstream_mel_decoder''' ) UpperCamelCase :Dict = SpectrogramDiffusionPipeline( notes_encoder=_a , continuous_encoder=_a , decoder=_a , scheduler=_a , melgan=_a , ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() parser.add_argument("""--output_path""", default=None, type=str, required=True, help="""Path to the converted model.""") parser.add_argument( """--save""", default=True, type=bool, required=False, help="""Whether to save the converted model or not.""" ) parser.add_argument( """--checkpoint_path""", default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help="""Path to the original jax model checkpoint.""", ) __snake_case = parser.parse_args() main(args)
259
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( _a , _a): # Load checkpoint SCREAMING_SNAKE_CASE : int = torch.load(_a , map_location="cpu") SCREAMING_SNAKE_CASE : Dict = chkpt["model"] # We have the base model one level deeper than the original XLM repository SCREAMING_SNAKE_CASE : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: SCREAMING_SNAKE_CASE : List[str] = v else: SCREAMING_SNAKE_CASE : int = v SCREAMING_SNAKE_CASE : int = chkpt["params"] SCREAMING_SNAKE_CASE : Union[str, Any] = {n: v for n, v in config.items() if not isinstance(_a , (torch.FloatTensor, numpy.ndarray))} SCREAMING_SNAKE_CASE : List[Any] = chkpt["dico_word2id"] SCREAMING_SNAKE_CASE : List[Any] = {s + "</w>" if s.find("@@") == -1 and i > 13 else s.replace("@@" , ""): i for s, i in vocab.items()} # Save pytorch-model SCREAMING_SNAKE_CASE : Tuple = pytorch_dump_folder_path + "/" + WEIGHTS_NAME SCREAMING_SNAKE_CASE : Any = pytorch_dump_folder_path + "/" + CONFIG_NAME SCREAMING_SNAKE_CASE : Optional[int] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(f"Save PyTorch model to {pytorch_weights_dump_path}") torch.save(_a , _a) print(f"Save configuration file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") print(f"Save vocab file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) a_ = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
76
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __A ={ '''configuration_clipseg''': [ '''CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''CLIPSegConfig''', '''CLIPSegTextConfig''', '''CLIPSegVisionConfig''', ], '''processing_clipseg''': ['''CLIPSegProcessor'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __A =[ '''CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST''', '''CLIPSegModel''', '''CLIPSegPreTrainedModel''', '''CLIPSegTextModel''', '''CLIPSegVisionModel''', '''CLIPSegForImageSegmentation''', ] if TYPE_CHECKING: from .configuration_clipseg import ( CLIPSEG_PRETRAINED_CONFIG_ARCHIVE_MAP, CLIPSegConfig, CLIPSegTextConfig, CLIPSegVisionConfig, ) from .processing_clipseg import CLIPSegProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_clipseg import ( CLIPSEG_PRETRAINED_MODEL_ARCHIVE_LIST, CLIPSegForImageSegmentation, CLIPSegModel, CLIPSegPreTrainedModel, CLIPSegTextModel, CLIPSegVisionModel, ) else: import sys __A =_LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
19
def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(a - b) for a, b in zip(_a , _a))) def lowerCamelCase__ ( _a): if point: if isinstance(_a , _a): for item in point: if not isinstance(_a , (int, float)): SCREAMING_SNAKE_CASE : List[Any] = ( "Expected a list of numbers as input, found " f"{type(_a).__name__}" ) raise TypeError(_a) else: SCREAMING_SNAKE_CASE : List[Any] = f"Expected a list of numbers as input, found {type(_a).__name__}" raise TypeError(_a) else: raise ValueError("Missing an input") def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(x - y) for x, y in zip(_a , _a))) if __name__ == "__main__": import doctest doctest.testmod()
76
0
from collections.abc import Callable import numpy as np def UpperCamelCase__ ( A__ , A__ , A__ , A__ , A__ ) -> Union[str, Any]: snake_case__ : Dict = int(np.ceil((x_end - xa) / step_size ) ) snake_case__ : Tuple = np.zeros((n + 1,) ) snake_case__ : int = ya snake_case__ : int = xa for k in range(_a ): snake_case__ : Any = y[k] + step_size * ode_func(_a , y[k] ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
143
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ ='vit_msn' def __init__( self : str , a : Tuple=768 , a : Tuple=12 , a : Any=12 , a : int=3072 , a : List[Any]="gelu" , a : Dict=0.0 , a : int=0.0 , a : str=0.02 , a : List[str]=1e-06 , a : List[Any]=224 , a : Union[str, Any]=16 , a : Union[str, Any]=3 , a : Tuple=True , **a : Dict , ) -> List[Any]: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Dict = hidden_size SCREAMING_SNAKE_CASE : Optional[Any] = num_hidden_layers SCREAMING_SNAKE_CASE : Optional[Any] = num_attention_heads SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : Any = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = initializer_range SCREAMING_SNAKE_CASE : int = layer_norm_eps SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : Tuple = patch_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : List[str] = qkv_bias
76
0
from typing import TYPE_CHECKING from ....utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ :int = {'''configuration_mmbt''': ['''MMBTConfig''']} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ :List[Any] = ['''MMBTForClassification''', '''MMBTModel''', '''ModalEmbeddings'''] if TYPE_CHECKING: from .configuration_mmbt import MMBTConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mmbt import MMBTForClassification, MMBTModel, ModalEmbeddings else: import sys A_ :Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
71
import baseaa def lowerCamelCase__ ( _a): return baseaa.aaaencode(string.encode("utf-8")) def lowerCamelCase__ ( _a): return baseaa.aaadecode(_a).decode("utf-8") if __name__ == "__main__": import doctest doctest.testmod()
76
0
"""simple docstring""" import inspect import unittest from transformers import DPTConfig from transformers.file_utils import is_torch_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, _config_zero_init, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import MODEL_MAPPING, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTModel from transformers.models.dpt.modeling_dpt import DPT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class __lowerCAmelCase : '''simple docstring''' def __init__( self , _a , _a=2 , _a=32 , _a=16 , _a=3 , _a=True , _a=True , _a=32 , _a=4 , _a=[0, 1, 2, 3] , _a=4 , _a=37 , _a="gelu" , _a=0.1 , _a=0.1 , _a=0.02 , _a=3 , _a=[1, 384, 24, 24] , _a=True , _a=None , ): __a = parent __a = batch_size __a = image_size __a = patch_size __a = num_channels __a = is_training __a = use_labels __a = hidden_size __a = num_hidden_layers __a = backbone_out_indices __a = num_attention_heads __a = intermediate_size __a = hidden_act __a = hidden_dropout_prob __a = attention_probs_dropout_prob __a = initializer_range __a = num_labels __a = backbone_featmap_shape __a = scope __a = is_hybrid # sequence length of DPT = num_patches + 1 (we add 1 for the [CLS] token) __a = (image_size // patch_size) ** 2 __a = num_patches + 1 def __UpperCAmelCase ( self ): __a = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __a = None if self.use_labels: __a = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __a = self.get_config() return config, pixel_values, labels def __UpperCAmelCase ( self ): __a = { "global_padding": "same", "layer_type": "bottleneck", "depths": [3, 4, 9], "out_features": ["stage1", "stage2", "stage3"], "embedding_dynamic_padding": True, "hidden_sizes": [96, 192, 384, 768], "num_groups": 2, } return DPTConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , backbone_out_indices=self.backbone_out_indices , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_a , initializer_range=self.initializer_range , is_hybrid=self.is_hybrid , backbone_config=_a , backbone_featmap_shape=self.backbone_featmap_shape , ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = DPTModel(config=_a ) model.to(_a ) model.eval() __a = model(_a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = self.num_labels __a = DPTForDepthEstimation(_a ) model.to(_a ) model.eval() __a = model(_a ) self.parent.assertEqual(result.predicted_depth.shape , (self.batch_size, self.image_size, self.image_size) ) def __UpperCAmelCase ( self , _a , _a , _a ): __a = self.num_labels __a = DPTForSemanticSegmentation(_a ) model.to(_a ) model.eval() __a = model(_a , labels=_a ) self.parent.assertEqual( result.logits.shape , (self.batch_size, self.num_labels, self.image_size, self.image_size) ) def __UpperCAmelCase ( self ): __a = self.prepare_config_and_inputs() __a = config_and_inputs __a = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class __lowerCAmelCase ( __A , __A , unittest.TestCase ): '''simple docstring''' __UpperCAmelCase : Any = (DPTModel, DPTForDepthEstimation, DPTForSemanticSegmentation) if is_torch_available() else () __UpperCAmelCase : str = ( { 'depth-estimation': DPTForDepthEstimation, 'feature-extraction': DPTModel, 'image-segmentation': DPTForSemanticSegmentation, } if is_torch_available() else {} ) __UpperCAmelCase : List[str] = False __UpperCAmelCase : int = False __UpperCAmelCase : Optional[Any] = False def __UpperCAmelCase ( self ): __a = DPTModelTester(self ) __a = ConfigTester(self , config_class=_a , has_text_modality=_a , hidden_size=37 ) def __UpperCAmelCase ( self ): self.config_tester.run_common_tests() @unittest.skip(reason='''DPT does not use inputs_embeds''' ) def __UpperCAmelCase ( self ): pass def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) __a = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_a , nn.Linear ) ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __a = model_class(_a ) __a = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __a = [*signature.parameters.keys()] __a = ["pixel_values"] self.assertListEqual(arg_names[:1] , _a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_depth_estimation(*_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*_a ) def __UpperCAmelCase ( self ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __a = self.model_tester.prepare_config_and_inputs_for_common() __a = True if model_class in get_values(_a ): continue __a = model_class(_a ) model.to(_a ) model.train() __a = self._prepare_for_class(_a , _a , return_labels=_a ) __a = model(**_a ).loss loss.backward() def __UpperCAmelCase ( self ): for model_class in self.all_model_classes: if model_class.__name__ == "DPTForDepthEstimation": continue __a = self.model_tester.prepare_config_and_inputs_for_common() __a = False __a = True if model_class in get_values(_a ) or not model_class.supports_gradient_checkpointing: continue __a = model_class(_a ) model.to(_a ) model.gradient_checkpointing_enable() model.train() __a = self._prepare_for_class(_a , _a , return_labels=_a ) __a = model(**_a ).loss loss.backward() def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() __a = _config_zero_init(_a ) for model_class in self.all_model_classes: __a = model_class(config=_a ) # Skip the check for the backbone __a = [] for name, module in model.named_modules(): if module.__class__.__name__ == "DPTViTHybridEmbeddings": __a = [f'''{name}.{key}''' for key in module.state_dict().keys()] break for name, param in model.named_parameters(): if param.requires_grad: if name in backbone_params: continue self.assertIn( ((param.data.mean() * 1E9).round() / 1E9).item() , [0.0, 1.0] , msg=f'''Parameter {name} of model {model_class} seems not properly initialized''' , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def __UpperCAmelCase ( self ): pass @slow def __UpperCAmelCase ( self ): for model_name in DPT_PRETRAINED_MODEL_ARCHIVE_LIST[1:]: __a = DPTModel.from_pretrained(_a ) self.assertIsNotNone(_a ) def __UpperCAmelCase ( self ): __a = self.model_tester.prepare_config_and_inputs_for_common() __a = "add" with self.assertRaises(_a ): __a = DPTForDepthEstimation(_a ) def lowercase ( ) -> str: __a = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision @slow class __lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' def __UpperCAmelCase ( self ): __a = DPTImageProcessor.from_pretrained('''Intel/dpt-hybrid-midas''' ) __a = DPTForDepthEstimation.from_pretrained('''Intel/dpt-hybrid-midas''' ).to(_a ) __a = prepare_img() __a = image_processor(images=_a , return_tensors='''pt''' ).to(_a ) # forward pass with torch.no_grad(): __a = model(**_a ) __a = outputs.predicted_depth # verify the predicted depth __a = torch.Size((1, 384, 384) ) self.assertEqual(predicted_depth.shape , _a ) __a = torch.tensor( [[[5.6437, 5.6146, 5.6511], [5.4371, 5.5649, 5.5958], [5.5215, 5.5184, 5.5293]]] ).to(_a ) self.assertTrue(torch.allclose(outputs.predicted_depth[:3, :3, :3] / 100 , _a , atol=1E-4 ) )
45
from datetime import datetime as dt import os from github import Github a_ = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : int = Github(os.environ["GITHUB_TOKEN"]) SCREAMING_SNAKE_CASE : List[str] = g.get_repo("huggingface/transformers") SCREAMING_SNAKE_CASE : Optional[int] = repo.get_issues(state="open") for issue in open_issues: SCREAMING_SNAKE_CASE : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda _a: i.created_at , reverse=_a) SCREAMING_SNAKE_CASE : str = comments[0] if len(_a) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="closed") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") if __name__ == "__main__": main()
76
0
def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Optional[Any]: """simple docstring""" return int(input_a == input_a == 0 ) def __lowerCamelCase ( ) -> List[str]: """simple docstring""" print("""Truth Table of NOR Gate:""" ) print("""| Input 1 | Input 2 | Output |""" ) print(F'| 0 | 0 | {nor_gate(0 ,0 )} |' ) print(F'| 0 | 1 | {nor_gate(0 ,1 )} |' ) print(F'| 1 | 0 | {nor_gate(1 ,0 )} |' ) print(F'| 1 | 1 | {nor_gate(1 ,1 )} |' ) if __name__ == "__main__": import doctest doctest.testmod() main()
306
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): if isinstance(_a , (list, tuple)) and isinstance(videos[0] , (list, tuple)) and is_valid_image(videos[0][0]): return videos elif isinstance(_a , (list, tuple)) and is_valid_image(videos[0]): return [videos] elif is_valid_image(_a): return [[videos]] raise ValueError(f"Could not make batched video from {videos}") class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =['pixel_values'] def __init__( self : Optional[Any] , a : bool = True , a : Dict[str, int] = None , a : PILImageResampling = PILImageResampling.BILINEAR , a : bool = True , a : Dict[str, int] = None , a : bool = True , a : Union[int, float] = 1 / 255 , a : bool = True , a : bool = True , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , **a : Tuple , ) -> None: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Tuple = size if size is not None else {"shortest_edge": 256} SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE : str = get_size_dict(a , param_name="crop_size" ) SCREAMING_SNAKE_CASE : Dict = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop SCREAMING_SNAKE_CASE : int = crop_size SCREAMING_SNAKE_CASE : int = resample SCREAMING_SNAKE_CASE : Any = do_rescale SCREAMING_SNAKE_CASE : int = rescale_factor SCREAMING_SNAKE_CASE : Tuple = offset SCREAMING_SNAKE_CASE : str = do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __UpperCamelCase ( self : Optional[Any] , a : np.ndarray , a : Dict[str, int] , a : PILImageResampling = PILImageResampling.BILINEAR , a : Optional[Union[str, ChannelDimension]] = None , **a : Union[str, Any] , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) if "shortest_edge" in size: SCREAMING_SNAKE_CASE : str = get_resize_output_image_size(a , size["shortest_edge"] , default_to_square=a ) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE : Dict = (size["height"], size["width"]) else: raise ValueError(F"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(a , size=a , resample=a , data_format=a , **a ) def __UpperCamelCase ( self : List[str] , a : np.ndarray , a : Dict[str, int] , a : Optional[Union[str, ChannelDimension]] = None , **a : str , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : str = get_size_dict(a ) if "height" not in size or "width" not in size: raise ValueError(F"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(a , size=(size["height"], size["width"]) , data_format=a , **a ) def __UpperCamelCase ( self : List[Any] , a : np.ndarray , a : Union[int, float] , a : bool = True , a : Optional[Union[str, ChannelDimension]] = None , **a : Tuple , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : int = image.astype(np.floataa ) if offset: SCREAMING_SNAKE_CASE : Union[str, Any] = image - (scale / 2) return rescale(a , scale=a , data_format=a , **a ) def __UpperCamelCase ( self : int , a : np.ndarray , a : Union[float, List[float]] , a : Union[float, List[float]] , a : Optional[Union[str, ChannelDimension]] = None , **a : List[str] , ) -> np.ndarray: """simple docstring""" return normalize(a , mean=a , std=a , data_format=a , **a ) def __UpperCamelCase ( self : Tuple , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: """simple docstring""" if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) if offset and not do_rescale: raise ValueError("For offset, do_rescale must also be set to True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : List[str] = to_numpy_array(a ) if do_resize: SCREAMING_SNAKE_CASE : Optional[Any] = self.resize(image=a , size=a , resample=a ) if do_center_crop: SCREAMING_SNAKE_CASE : Union[str, Any] = self.center_crop(a , size=a ) if do_rescale: SCREAMING_SNAKE_CASE : Any = self.rescale(image=a , scale=a , offset=a ) if do_normalize: SCREAMING_SNAKE_CASE : Tuple = self.normalize(image=a , mean=a , std=a ) SCREAMING_SNAKE_CASE : Optional[int] = to_channel_dimension_format(a , a ) return image def __UpperCamelCase ( self : Dict , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[str, TensorType]] = None , a : ChannelDimension = ChannelDimension.FIRST , **a : Tuple , ) -> PIL.Image.Image: """simple docstring""" SCREAMING_SNAKE_CASE : str = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Union[str, Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : int = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : str = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : Optional[Any] = offset if offset is not None else self.offset SCREAMING_SNAKE_CASE : str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : int = size if size is not None else self.size SCREAMING_SNAKE_CASE : List[Any] = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = get_size_dict(a , param_name="crop_size" ) if not valid_images(a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) SCREAMING_SNAKE_CASE : Optional[int] = make_batched(a ) SCREAMING_SNAKE_CASE : List[Any] = [ [ self._preprocess_image( image=a , do_resize=a , size=a , resample=a , do_center_crop=a , crop_size=a , do_rescale=a , rescale_factor=a , offset=a , do_normalize=a , image_mean=a , image_std=a , data_format=a , ) for img in video ] for video in videos ] SCREAMING_SNAKE_CASE : Optional[int] = {"pixel_values": videos} return BatchFeature(data=a , tensor_type=a )
76
0
import argparse import json from pathlib import Path import torch import torchaudio from datasets import load_dataset from huggingface_hub import hf_hub_download from transformers import ASTConfig, ASTFeatureExtractor, ASTForAudioClassification from transformers.utils import logging logging.set_verbosity_info() __lowerCAmelCase : Union[str, Any] = logging.get_logger(__name__) def UpperCAmelCase_ ( __lowerCAmelCase ) -> Any: __lowercase : List[str] = ASTConfig() if "10-10" in model_name: pass elif "speech-commands" in model_name: __lowercase : Dict = 128 elif "12-12" in model_name: __lowercase : List[str] = 12 __lowercase : Optional[int] = 12 elif "14-14" in model_name: __lowercase : Union[str, Any] = 14 __lowercase : Tuple = 14 elif "16-16" in model_name: __lowercase : Optional[Any] = 16 __lowercase : str = 16 else: raise ValueError('''Model not supported''' ) __lowercase : Optional[Any] = "huggingface/label-files" if "speech-commands" in model_name: __lowercase : Union[str, Any] = 35 __lowercase : Tuple = "speech-commands-v2-id2label.json" else: __lowercase : Union[str, Any] = 527 __lowercase : str = "audioset-id2label.json" __lowercase : List[str] = json.load(open(hf_hub_download(_a , _a , repo_type='''dataset''' ) , '''r''' ) ) __lowercase : Union[str, Any] = {int(_a ): v for k, v in idalabel.items()} __lowercase : Optional[Any] = idalabel __lowercase : Tuple = {v: k for k, v in idalabel.items()} return config def UpperCAmelCase_ ( __lowerCAmelCase ) -> int: if "module.v" in name: __lowercase : int = name.replace('''module.v''' , '''audio_spectrogram_transformer''' ) if "cls_token" in name: __lowercase : List[Any] = name.replace('''cls_token''' , '''embeddings.cls_token''' ) if "dist_token" in name: __lowercase : List[str] = name.replace('''dist_token''' , '''embeddings.distillation_token''' ) if "pos_embed" in name: __lowercase : Union[str, Any] = name.replace('''pos_embed''' , '''embeddings.position_embeddings''' ) if "patch_embed.proj" in name: __lowercase : str = name.replace('''patch_embed.proj''' , '''embeddings.patch_embeddings.projection''' ) # transformer blocks if "blocks" in name: __lowercase : List[str] = name.replace('''blocks''' , '''encoder.layer''' ) if "attn.proj" in name: __lowercase : Union[str, Any] = name.replace('''attn.proj''' , '''attention.output.dense''' ) if "attn" in name: __lowercase : List[Any] = name.replace('''attn''' , '''attention.self''' ) if "norm1" in name: __lowercase : Union[str, Any] = name.replace('''norm1''' , '''layernorm_before''' ) if "norm2" in name: __lowercase : Optional[Any] = name.replace('''norm2''' , '''layernorm_after''' ) if "mlp.fc1" in name: __lowercase : List[str] = name.replace('''mlp.fc1''' , '''intermediate.dense''' ) if "mlp.fc2" in name: __lowercase : Tuple = name.replace('''mlp.fc2''' , '''output.dense''' ) # final layernorm if "audio_spectrogram_transformer.norm" in name: __lowercase : Optional[Any] = name.replace('''audio_spectrogram_transformer.norm''' , '''audio_spectrogram_transformer.layernorm''' ) # classifier head if "module.mlp_head.0" in name: __lowercase : str = name.replace('''module.mlp_head.0''' , '''classifier.layernorm''' ) if "module.mlp_head.1" in name: __lowercase : Dict = name.replace('''module.mlp_head.1''' , '''classifier.dense''' ) return name def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase ) -> Dict: for key in orig_state_dict.copy().keys(): __lowercase : List[Any] = orig_state_dict.pop(_a ) if "qkv" in key: __lowercase : str = key.split('''.''' ) __lowercase : List[str] = int(key_split[3] ) __lowercase : Union[str, Any] = config.hidden_size if "weight" in key: __lowercase : str = val[:dim, :] __lowercase : str = val[dim : dim * 2, :] __lowercase : List[str] = val[-dim:, :] else: __lowercase : Optional[int] = val[:dim] __lowercase : Any = val[dim : dim * 2] __lowercase : List[str] = val[-dim:] else: __lowercase : Dict = val return orig_state_dict def UpperCAmelCase_ ( __lowerCAmelCase ) -> Union[str, Any]: __lowercase : List[Any] = [ "module.v.head.weight", "module.v.head.bias", "module.v.head_dist.weight", "module.v.head_dist.bias", ] for k in ignore_keys: state_dict.pop(_a , _a ) @torch.no_grad() def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=False ) -> List[str]: __lowercase : int = get_audio_spectrogram_transformer_config(_a ) __lowercase : Optional[Any] = { "ast-finetuned-audioset-10-10-0.4593": ( "https://www.dropbox.com/s/ca0b1v2nlxzyeb4/audioset_10_10_0.4593.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.450": ( "https://www.dropbox.com/s/1tv0hovue1bxupk/audioset_10_10_0.4495.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448": ( "https://www.dropbox.com/s/6u5sikl4b9wo4u5/audioset_10_10_0.4483.pth?dl=1" ), "ast-finetuned-audioset-10-10-0.448-v2": ( "https://www.dropbox.com/s/kt6i0v9fvfm1mbq/audioset_10_10_0.4475.pth?dl=1" ), "ast-finetuned-audioset-12-12-0.447": ( "https://www.dropbox.com/s/snfhx3tizr4nuc8/audioset_12_12_0.4467.pth?dl=1" ), "ast-finetuned-audioset-14-14-0.443": ( "https://www.dropbox.com/s/z18s6pemtnxm4k7/audioset_14_14_0.4431.pth?dl=1" ), "ast-finetuned-audioset-16-16-0.442": ( "https://www.dropbox.com/s/mdsa4t1xmcimia6/audioset_16_16_0.4422.pth?dl=1" ), "ast-finetuned-speech-commands-v2": ( "https://www.dropbox.com/s/q0tbqpwv44pquwy/speechcommands_10_10_0.9812.pth?dl=1" ), } # load original state_dict __lowercase : List[str] = model_name_to_url[model_name] __lowercase : List[Any] = torch.hub.load_state_dict_from_url(_a , map_location='''cpu''' ) # remove some keys remove_keys(_a ) # rename some keys __lowercase : List[str] = convert_state_dict(_a , _a ) # load 🤗 model __lowercase : Dict = ASTForAudioClassification(_a ) model.eval() model.load_state_dict(_a ) # verify outputs on dummy input # source: https://github.com/YuanGongND/ast/blob/79e873b8a54d0a3b330dd522584ff2b9926cd581/src/run.py#L62 __lowercase : List[str] = -4.2677393 if "speech-commands" not in model_name else -6.845978 __lowercase : Tuple = 4.5689974 if "speech-commands" not in model_name else 5.5654526 __lowercase : List[Any] = 1_024 if "speech-commands" not in model_name else 128 __lowercase : Optional[int] = ASTFeatureExtractor(mean=_a , std=_a , max_length=_a ) if "speech-commands" in model_name: __lowercase : Union[str, Any] = load_dataset('''speech_commands''' , '''v0.02''' , split='''validation''' ) __lowercase : Any = dataset[0]["audio"]["array"] else: __lowercase : Dict = hf_hub_download( repo_id='''nielsr/audio-spectogram-transformer-checkpoint''' , filename='''sample_audio.flac''' , repo_type='''dataset''' , ) __lowercase : List[str] = torchaudio.load(_a ) __lowercase : Any = waveform.squeeze().numpy() __lowercase : Optional[Any] = feature_extractor(_a , sampling_rate=16_000 , return_tensors='''pt''' ) # forward pass __lowercase : Optional[Any] = model(**_a ) __lowercase : int = outputs.logits if model_name == "ast-finetuned-audioset-10-10-0.4593": __lowercase : Any = torch.tensor([-0.8760, -7.0042, -8.6602] ) elif model_name == "ast-finetuned-audioset-10-10-0.450": __lowercase : Any = torch.tensor([-1.1986, -7.0903, -8.2718] ) elif model_name == "ast-finetuned-audioset-10-10-0.448": __lowercase : str = torch.tensor([-2.6128, -8.0080, -9.4344] ) elif model_name == "ast-finetuned-audioset-10-10-0.448-v2": __lowercase : str = torch.tensor([-1.5080, -7.4534, -8.8917] ) elif model_name == "ast-finetuned-audioset-12-12-0.447": __lowercase : str = torch.tensor([-0.5050, -6.5833, -8.0843] ) elif model_name == "ast-finetuned-audioset-14-14-0.443": __lowercase : Union[str, Any] = torch.tensor([-0.3826, -7.0336, -8.2413] ) elif model_name == "ast-finetuned-audioset-16-16-0.442": __lowercase : Any = torch.tensor([-1.2113, -6.9101, -8.3470] ) elif model_name == "ast-finetuned-speech-commands-v2": __lowercase : Union[str, Any] = torch.tensor([6.1589, -8.0566, -8.7984] ) else: raise ValueError('''Unknown model name''' ) if not torch.allclose(logits[0, :3] , _a , atol=1E-4 ): raise ValueError('''Logits don\'t match''' ) print('''Looks ok!''' ) if pytorch_dump_folder_path is not None: Path(_a ).mkdir(exist_ok=_a ) print(F'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(_a ) print(F'Saving feature extractor to {pytorch_dump_folder_path}' ) feature_extractor.save_pretrained(_a ) if push_to_hub: print('''Pushing model and feature extractor to the hub...''' ) model.push_to_hub(F'MIT/{model_name}' ) feature_extractor.push_to_hub(F'MIT/{model_name}' ) if __name__ == "__main__": __lowerCAmelCase : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( "--model_name", default="ast-finetuned-audioset-10-10-0.4593", type=str, help="Name of the Audio Spectrogram Transformer model you\'d like to convert.", ) parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to the output PyTorch model directory." ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model to the 🤗 hub." ) __lowerCAmelCase : Optional[Any] = parser.parse_args() convert_audio_spectrogram_transformer_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
156
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a_ = logging.get_logger(__name__) a_ = {'vocab_file': 'vocab.txt'} a_ = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a_ = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } a_ = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =ConvBertTokenizer def __init__( self : List[str] , a : Union[str, Any]=None , a : Optional[int]=None , a : int=True , a : Tuple="[UNK]" , a : Dict="[SEP]" , a : Dict="[PAD]" , a : List[Any]="[CLS]" , a : Tuple="[MASK]" , a : Dict=True , a : Optional[Any]=None , **a : str , ) -> Dict: """simple docstring""" super().__init__( a , tokenizer_file=a , do_lower_case=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , tokenize_chinese_chars=a , strip_accents=a , **a , ) SCREAMING_SNAKE_CASE : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , a ) != do_lower_case or normalizer_state.get("strip_accents" , a ) != strip_accents or normalizer_state.get("handle_chinese_chars" , a ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE : List[str] = getattr(a , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE : Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE : Any = strip_accents SCREAMING_SNAKE_CASE : Optional[int] = tokenize_chinese_chars SCREAMING_SNAKE_CASE : List[str] = normalizer_class(**a ) SCREAMING_SNAKE_CASE : str = do_lower_case def __UpperCamelCase ( self : Union[str, Any] , a : List[Any] , a : int=None ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : Tuple , a : str , a : Optional[str] = None ) -> Tuple[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self._tokenizer.model.save(a , name=a ) return tuple(a )
76
0
import os from itertools import chain from random import randrange, shuffle import pytest from .sola import PokerHand lowercase_ = ( "4S 3H 2C 7S 5H", "9D 8H 2C 6S 7H", "2D 6D 9D TH 7D", "TC 8C 2S JH 6C", "JH 8S TH AH QH", "TS KS 5S 9S AC", "KD 6S 9D TH AD", "KS 8D 4D 9S 4S", # pair "8C 4S KH JS 4D", # pair "QH 8H KD JH 8S", # pair "KC 4H KS 2H 8D", # pair "KD 4S KC 3H 8S", # pair "AH 8S AS KC JH", # pair "3H 4C 4H 3S 2H", # 2 pairs "5S 5D 2C KH KH", # 2 pairs "3C KH 5D 5S KH", # 2 pairs "AS 3C KH AD KH", # 2 pairs "7C 7S 3S 7H 5S", # 3 of a kind "7C 7S KH 2H 7H", # 3 of a kind "AC KH QH AH AS", # 3 of a kind "2H 4D 3C AS 5S", # straight (low ace) "3C 5C 4C 2C 6H", # straight "6S 8S 7S 5H 9H", # straight "JS QS 9H TS KH", # straight "QC KH TS JS AH", # straight (high ace) "8C 9C 5C 3C TC", # flush "3S 8S 9S 5S KS", # flush "4C 5C 9C 8C KC", # flush "JH 8H AH KH QH", # flush "3D 2H 3H 2C 2D", # full house "2H 2C 3S 3H 3D", # full house "KH KC 3S 3H 3D", # full house "JC 6H JS JD JH", # 4 of a kind "JC 7H JS JD JH", # 4 of a kind "JC KH JS JD JH", # 4 of a kind "2S AS 4S 5S 3S", # straight flush (low ace) "2D 6D 3D 4D 5D", # straight flush "5C 6C 3C 7C 4C", # straight flush "JH 9H TH KH QH", # straight flush "JH AH TH KH QH", # royal flush (high ace straight flush) ) lowercase_ = ( ("2H 3H 4H 5H 6H", "KS AS TS QS JS", "Loss"), ("2H 3H 4H 5H 6H", "AS AD AC AH JD", "Win"), ("AS AH 2H AD AC", "JS JD JC JH 3D", "Win"), ("2S AH 2H AS AC", "JS JD JC JH AD", "Loss"), ("2S AH 2H AS AC", "2H 3H 5H 6H 7H", "Win"), ("AS 3S 4S 8S 2S", "2H 3H 5H 6H 7H", "Win"), ("2H 3H 5H 6H 7H", "2S 3H 4H 5S 6C", "Win"), ("2S 3H 4H 5S 6C", "3D 4C 5H 6H 2S", "Tie"), ("2S 3H 4H 5S 6C", "AH AC 5H 6H AS", "Win"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H AS", "Loss"), ("2S 2H 4H 5S 4C", "AH AC 5H 6H 7S", "Win"), ("6S AD 7H 4S AS", "AH AC 5H 6H 7S", "Loss"), ("2S AH 4H 5S KC", "AH AC 5H 6H 7S", "Loss"), ("2S 3H 6H 7S 9C", "7H 3C TH 6H 9S", "Loss"), ("4S 5H 6H TS AC", "3S 5H 6H TS AC", "Win"), ("2S AH 4H 5S 6C", "AD 4C 5H 6H 2C", "Tie"), ("AS AH 3H AD AC", "AS AH 2H AD AC", "Win"), ("AH AC 5H 5C QS", "AH AC 5H 5C KS", "Loss"), ("AH AC 5H 5C QS", "KH KC 5H 5C QS", "Win"), ("7C 7S KH 2H 7H", "3C 3S AH 2H 3H", "Win"), ("3C 3S AH 2H 3H", "7C 7S KH 2H 7H", "Loss"), ("6H 5H 4H 3H 2H", "5H 4H 3H 2H AH", "Win"), ("5H 4H 3H 2H AH", "5H 4H 3H 2H AH", "Tie"), ("5H 4H 3H 2H AH", "6H 5H 4H 3H 2H", "Loss"), ("AH AD KS KC AC", "AH KD KH AC KC", "Win"), ("2H 4D 3C AS 5S", "2H 4D 3C 6S 5S", "Loss"), ("2H 3S 3C 3H 2S", "3S 3C 2S 2H 2D", "Win"), ("4D 6D 5D 2D JH", "3S 8S 3H TC KH", "Loss"), ("4S 6C 8S 3S 7S", "AD KS 2D 7D 7C", "Loss"), ("6S 4C 7H 8C 3H", "5H JC AH 9D 9C", "Loss"), ("9D 9H JH TC QH", "3C 2S JS 5C 7H", "Win"), ("2H TC 8S AD 9S", "4H TS 7H 2C 5C", "Win"), ("9D 3S 2C 7S 7C", "JC TD 3C TC 9H", "Loss"), ) lowercase_ = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", True), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", False), ("AS 3S 4S 8S 2S", True), ) lowercase_ = ( ("2H 3H 4H 5H 6H", True), ("AS AH 2H AD AC", False), ("2H 3H 5H 6H 7H", False), ("KS AS TS QS JS", True), ("8H 9H QS JS TH", True), ) lowercase_ = ( ("2H 4D 3C AS 5S", True, [5, 4, 3, 2, 14]), ("2H 5D 3C AS 5S", False, [14, 5, 5, 3, 2]), ("JH QD KC AS TS", False, [14, 13, 12, 11, 10]), ("9D 3S 2C 7S 7C", False, [9, 7, 7, 3, 2]), ) lowercase_ = ( ("JH AH TH KH QH", 0), ("JH 9H TH KH QH", 0), ("JC KH JS JD JH", 7), ("KH KC 3S 3H 3D", 6), ("8C 9C 5C 3C TC", 0), ("JS QS 9H TS KH", 0), ("7C 7S KH 2H 7H", 3), ("3C KH 5D 5S KH", 2), ("QH 8H KD JH 8S", 1), ("2D 6D 9D TH 7D", 0), ) lowercase_ = ( ("JH AH TH KH QH", 23), ("JH 9H TH KH QH", 22), ("JC KH JS JD JH", 21), ("KH KC 3S 3H 3D", 20), ("8C 9C 5C 3C TC", 19), ("JS QS 9H TS KH", 18), ("7C 7S KH 2H 7H", 17), ("3C KH 5D 5S KH", 16), ("QH 8H KD JH 8S", 15), ("2D 6D 9D TH 7D", 14), ) def _snake_case( ) -> str: '''simple docstring''' A__ = randrange(len(_a ) ), randrange(len(_a ) ) A__ = ["Loss", "Tie", "Win"][(play >= oppo) + (play > oppo)] A__ = SORTED_HANDS[play], SORTED_HANDS[oppo] return hand, other, expected def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[Any] = 100 ) -> Tuple: '''simple docstring''' return (generate_random_hand() for _ in range(_a )) @pytest.mark.parametrize('hand, expected' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : str ) -> Union[str, Any]: '''simple docstring''' assert PokerHand(_a )._is_flush() == expected @pytest.mark.parametrize('hand, expected' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : Dict , SCREAMING_SNAKE_CASE__ : Tuple ) -> Optional[Any]: '''simple docstring''' assert PokerHand(_a )._is_straight() == expected @pytest.mark.parametrize('hand, expected, card_values' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : List[Any] ) -> Optional[Any]: '''simple docstring''' A__ = PokerHand(_a ) assert player._is_five_high_straight() == expected assert player._card_values == card_values @pytest.mark.parametrize('hand, expected' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' assert PokerHand(_a )._is_same_kind() == expected @pytest.mark.parametrize('hand, expected' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[int] ) -> str: '''simple docstring''' assert PokerHand(_a )._hand_type == expected @pytest.mark.parametrize('hand, other, expected' , _a ) def _snake_case( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[int] , SCREAMING_SNAKE_CASE__ : Dict ) -> Optional[Any]: '''simple docstring''' assert PokerHand(_a ).compare_with(PokerHand(_a ) ) == expected @pytest.mark.parametrize('hand, other, expected' , generate_random_hands() ) def _snake_case( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : int ) -> Optional[Any]: '''simple docstring''' assert PokerHand(_a ).compare_with(PokerHand(_a ) ) == expected def _snake_case( ) -> List[str]: '''simple docstring''' A__ = [PokerHand(_a ) for hand in SORTED_HANDS] A__ = poker_hands.copy() shuffle(_a ) A__ = chain(sorted(_a ) ) for index, hand in enumerate(_a ): assert hand == poker_hands[index] def _snake_case( ) -> List[Any]: '''simple docstring''' A__ = [PokerHand('2D AC 3H 4H 5S' ), PokerHand('2S 3H 4H 5S 6C' )] pokerhands.sort(reverse=_a ) assert pokerhands[0].__str__() == "2S 3H 4H 5S 6C" def _snake_case( ) -> Any: '''simple docstring''' A__ = PokerHand('2C 4S AS 3D 5C' ) A__ = True A__ = [5, 4, 3, 2, 14] for _ in range(10 ): assert pokerhand._is_five_high_straight() == expected assert pokerhand._card_values == expected_card_values def _snake_case( ) -> Tuple: '''simple docstring''' A__ = 0 A__ = os.path.abspath(os.path.dirname(_a ) ) A__ = os.path.join(_a , 'poker_hands.txt' ) with open(_a ) as file_hand: for line in file_hand: A__ = line[:14].strip() A__ = line[15:].strip() A__ = PokerHand(_a ), PokerHand(_a ) A__ = player.compare_with(_a ) if output == "Win": answer += 1 assert answer == 376
7
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. a_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_a) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE : Union[str, Any] = terminalreporter.config.getoption("--make-reports") if make_reports: pytest_terminal_summary_main(_a , id=_a)
76
0
import importlib import inspect import json import os import re import shutil import sys from pathlib import Path from typing import Dict, Optional, Union from urllib import request from huggingface_hub import HfFolder, cached_download, hf_hub_download, model_info from packaging import version from .. import __version__ from . import DIFFUSERS_DYNAMIC_MODULE_NAME, HF_MODULES_CACHE, logging __A : str = ( "https://raw.githubusercontent.com/huggingface/diffusers/{revision}/examples/community/{pipeline}.py" ) __A : Optional[Any] = logging.get_logger(__name__) # pylint: disable=invalid-name def __SCREAMING_SNAKE_CASE ( ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = "https://pypi.org/pypi/diffusers/json" UpperCAmelCase = json.loads(request.urlopen(_a ).read() )["releases"].keys() return sorted(_a , key=lambda UpperCamelCase__ : version.Version(_a ) ) def __SCREAMING_SNAKE_CASE ( ) -> Optional[int]: '''simple docstring''' if HF_MODULES_CACHE in sys.path: return sys.path.append(_a ) os.makedirs(_a , exist_ok=_a ) UpperCAmelCase = Path(_a ) / "__init__.py" if not init_path.exists(): init_path.touch() def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' init_hf_modules() UpperCAmelCase = Path(_a ) / name # If the parent module does not exist yet, recursively create it. if not dynamic_module_path.parent.exists(): create_dynamic_module(dynamic_module_path.parent ) os.makedirs(_a , exist_ok=_a ) UpperCAmelCase = dynamic_module_path / "__init__.py" if not init_path.exists(): init_path.touch() def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> List[Any]: '''simple docstring''' with open(_a , '''r''' , encoding='''utf-8''' ) as f: UpperCAmelCase = f.read() # Imports of the form `import .xxx` UpperCAmelCase = re.findall('''^\s*import\s+\.(\S+)\s*$''' , _a , flags=re.MULTILINE ) # Imports of the form `from .xxx import yyy` relative_imports += re.findall('''^\s*from\s+\.(\S+)\s+import''' , _a , flags=re.MULTILINE ) # Unique-ify return list(set(_a ) ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase = False UpperCAmelCase = [module_file] UpperCAmelCase = [] # Let's recurse through all relative imports while not no_change: UpperCAmelCase = [] for f in files_to_check: new_imports.extend(get_relative_imports(_a ) ) UpperCAmelCase = Path(_a ).parent UpperCAmelCase = [str(module_path / m ) for m in new_imports] UpperCAmelCase = [f for f in new_import_files if f not in all_relative_imports] UpperCAmelCase = [F"""{f}.py""" for f in new_import_files] UpperCAmelCase = len(_a ) == 0 all_relative_imports.extend(_a ) return all_relative_imports def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> int: '''simple docstring''' with open(_a , '''r''' , encoding='''utf-8''' ) as f: UpperCAmelCase = f.read() # Imports of the form `import xxx` UpperCAmelCase = re.findall('''^\s*import\s+(\S+)\s*$''' , _a , flags=re.MULTILINE ) # Imports of the form `from xxx import yyy` imports += re.findall('''^\s*from\s+(\S+)\s+import''' , _a , flags=re.MULTILINE ) # Only keep the top-level module UpperCAmelCase = [imp.split('''.''' )[0] for imp in imports if not imp.startswith('''.''' )] # Unique-ify and test we got them all UpperCAmelCase = list(set(_a ) ) UpperCAmelCase = [] for imp in imports: try: importlib.import_module(_a ) except ImportError: missing_packages.append(_a ) if len(_a ) > 0: raise ImportError( '''This modeling file requires the following packages that were not found in your environment: ''' F"""{", ".join(_a )}. Run `pip install {" ".join(_a )}`""" ) return get_relative_imports(_a ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = module_path.replace(os.path.sep , '''.''' ) UpperCAmelCase = importlib.import_module(_a ) if class_name is None: return find_pipeline_class(_a ) return getattr(_a , _a ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> int: '''simple docstring''' from ..pipelines import DiffusionPipeline UpperCAmelCase = dict(inspect.getmembers(_a , inspect.isclass ) ) UpperCAmelCase = None for cls_name, cls in cls_members.items(): if ( cls_name != DiffusionPipeline.__name__ and issubclass(cls , _a ) and cls.__module__.split('''.''' )[0] != "diffusers" ): if pipeline_class is not None: raise ValueError( F"""Multiple classes that inherit from {DiffusionPipeline.__name__} have been found:""" F""" {pipeline_class.__name__}, and {cls_name}. Please make sure to define only one in""" F""" {loaded_module}.""" ) UpperCAmelCase = cls return pipeline_class def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = False , UpperCamelCase__ = False , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = False , ) -> Dict: '''simple docstring''' UpperCAmelCase = str(_a ) UpperCAmelCase = os.path.join(_a , _a ) if os.path.isfile(_a ): UpperCAmelCase = module_file_or_url UpperCAmelCase = "local" elif pretrained_model_name_or_path.count('''/''' ) == 0: UpperCAmelCase = get_diffusers_versions() # cut ".dev0" UpperCAmelCase = "v" + ".".join(__version__.split('''.''' )[:3] ) # retrieve github version that matches if revision is None: UpperCAmelCase = latest_version if latest_version[1:] in available_versions else "main" logger.info(F"""Defaulting to latest_version: {revision}.""" ) elif revision in available_versions: UpperCAmelCase = F"""v{revision}""" elif revision == "main": UpperCAmelCase = revision else: raise ValueError( F"""`custom_revision`: {revision} does not exist. Please make sure to choose one of""" F""" {", ".join(available_versions + ["main"] )}.""" ) # community pipeline on GitHub UpperCAmelCase = COMMUNITY_PIPELINES_URL.format(revision=_a , pipeline=_a ) try: UpperCAmelCase = cached_download( _a , cache_dir=_a , force_download=_a , proxies=_a , resume_download=_a , local_files_only=_a , use_auth_token=_a , ) UpperCAmelCase = "git" UpperCAmelCase = pretrained_model_name_or_path + ".py" except EnvironmentError: logger.error(F"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise else: try: # Load from URL or cache if already cached UpperCAmelCase = hf_hub_download( _a , _a , cache_dir=_a , force_download=_a , proxies=_a , resume_download=_a , local_files_only=_a , use_auth_token=_a , ) UpperCAmelCase = os.path.join('''local''' , '''--'''.join(pretrained_model_name_or_path.split('''/''' ) ) ) except EnvironmentError: logger.error(F"""Could not locate the {module_file} inside {pretrained_model_name_or_path}.""" ) raise # Check we have all the requirements in our environment UpperCAmelCase = check_imports(_a ) # Now we move the module inside our cached dynamic modules. UpperCAmelCase = DIFFUSERS_DYNAMIC_MODULE_NAME + os.path.sep + submodule create_dynamic_module(_a ) UpperCAmelCase = Path(_a ) / full_submodule if submodule == "local" or submodule == "git": # We always copy local files (we could hash the file to see if there was a change, and give them the name of # that hash, to only copy when there is a modification but it seems overkill for now). # The only reason we do the copy is to avoid putting too many folders in sys.path. shutil.copy(_a , submodule_path / module_file ) for module_needed in modules_needed: UpperCAmelCase = F"""{module_needed}.py""" shutil.copy(os.path.join(_a , _a ) , submodule_path / module_needed ) else: # Get the commit hash # TODO: we will get this info in the etag soon, so retrieve it from there and not here. if isinstance(_a , _a ): UpperCAmelCase = use_auth_token elif use_auth_token is True: UpperCAmelCase = HfFolder.get_token() else: UpperCAmelCase = None UpperCAmelCase = model_info(_a , revision=_a , token=_a ).sha # The module file will end up being placed in a subfolder with the git hash of the repo. This way we get the # benefit of versioning. UpperCAmelCase = submodule_path / commit_hash UpperCAmelCase = full_submodule + os.path.sep + commit_hash create_dynamic_module(_a ) if not (submodule_path / module_file).exists(): shutil.copy(_a , submodule_path / module_file ) # Make sure we also have every file with relative for module_needed in modules_needed: if not (submodule_path / module_needed).exists(): get_cached_module_file( _a , F"""{module_needed}.py""" , cache_dir=_a , force_download=_a , resume_download=_a , proxies=_a , use_auth_token=_a , revision=_a , local_files_only=_a , ) return os.path.join(_a , _a ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = False , UpperCamelCase__ = False , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = None , UpperCamelCase__ = False , **UpperCamelCase__ , ) -> Optional[int]: '''simple docstring''' UpperCAmelCase = get_cached_module_file( _a , _a , cache_dir=_a , force_download=_a , resume_download=_a , proxies=_a , use_auth_token=_a , revision=_a , local_files_only=_a , ) return get_class_in_module(_a , final_module.replace('''.py''' , '''''' ) )
273
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , a : int , a : Optional[int]=13 , a : Optional[int]=3 , a : int=224 , a : Optional[int]=30 , a : int=400 , a : Union[str, Any]=True , a : int=None , a : Tuple=True , a : Tuple=[0.5, 0.5, 0.5] , a : Optional[int]=[0.5, 0.5, 0.5] , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = size if size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE : Union[str, Any] = parent SCREAMING_SNAKE_CASE : int = batch_size SCREAMING_SNAKE_CASE : int = num_channels SCREAMING_SNAKE_CASE : Any = image_size SCREAMING_SNAKE_CASE : Tuple = min_resolution SCREAMING_SNAKE_CASE : str = max_resolution SCREAMING_SNAKE_CASE : int = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : int = do_normalize SCREAMING_SNAKE_CASE : Tuple = image_mean SCREAMING_SNAKE_CASE : Tuple = image_std def __UpperCamelCase ( self : Any ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCamelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =ViTImageProcessor if is_vision_available() else None def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = EfficientFormerImageProcessorTester(self ) @property def __UpperCamelCase ( self : Any ) -> List[str]: """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def __UpperCamelCase ( self : List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , "image_mean" ) ) self.assertTrue(hasattr(a , "image_std" ) ) self.assertTrue(hasattr(a , "do_normalize" ) ) self.assertTrue(hasattr(a , "do_resize" ) ) self.assertTrue(hasattr(a , "size" ) ) def __UpperCamelCase ( self : int ) -> str: """simple docstring""" pass def __UpperCamelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE : List[str] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE : int = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Any = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
76
0
import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging __magic_name__: Dict = logging.get_logger(__name__) def UpperCamelCase ( _A, _A, _A, _A=False ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise if not is_sharded: __magic_name__ : str = os.path.abspath(_a ) logger.info(f'Loading PyTorch weights from {pt_path}' ) __magic_name__ : Union[str, Any] = torch.load(_a, map_location="""cpu""" ) logger.info(f'PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.' ) __magic_name__ : str = convert_pytorch_state_dict_to_flax(_a, _a ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files __magic_name__ : List[str] = convert_pytorch_sharded_state_dict_to_flax(_a, _a ) return flax_state_dict def UpperCamelCase ( _A, _A, _A, _A, ): """simple docstring""" def is_key_or_prefix_key_in_dict(_A ) -> bool: return len(set(_a ) & {key, (model_prefix,) + key} ) > 0 # layer norm __magic_name__ : Dict = pt_tuple_key[:-1] + ("scale",) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(_a ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean __magic_name__ : Tuple = pt_tuple_key[:-1] + ("mean",) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(_a ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var __magic_name__ : List[str] = pt_tuple_key[:-1] + ("var",) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(_a ): return renamed_pt_tuple_key, pt_tensor # embedding __magic_name__ : int = pt_tuple_key[:-1] + ("embedding",) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(_a ): return renamed_pt_tuple_key, pt_tensor # conv layer __magic_name__ : Any = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(_a ): __magic_name__ : Optional[int] = pt_tensor.transpose(2, 3, 1, 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer __magic_name__ : Union[str, Any] = pt_tuple_key[:-1] + ("kernel",) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(_a ): __magic_name__ : Tuple = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight __magic_name__ : Dict = pt_tuple_key[:-1] + ("weight",) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias __magic_name__ : List[str] = pt_tuple_key[:-1] + ("bias",) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 __magic_name__ : Optional[int] = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): __magic_name__ : List[str] = pt_tuple_key[-2] + "_g" elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): __magic_name__ : Any = pt_tuple_key[-2] + "_v" if name is not None: __magic_name__ : str = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def UpperCamelCase ( _A, _A ): """simple docstring""" __magic_name__ : int = {k: v.numpy() for k, v in pt_state_dict.items()} __magic_name__ : List[Any] = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: __magic_name__ : Dict = flax_model.params["params"] else: __magic_name__ : Union[str, Any] = flax_model.params __magic_name__ : Optional[Any] = flatten_dict(_a ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __magic_name__ : int = flatten_dict(flax_model.params["""batch_stats"""] ) random_flax_state_dict.update(_a ) __magic_name__ : Dict = {} __magic_name__ : int = (model_prefix not in flax_model_params) and ( model_prefix in {k.split(""".""" )[0] for k in pt_state_dict.keys()} ) __magic_name__ : Union[str, Any] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split(""".""" )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __magic_name__ : Optional[int] = tuple(pt_key.split(""".""" ) ) # remove base model prefix if necessary __magic_name__ : Dict = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __magic_name__ : Dict = pt_tuple_key[1:] # Correctly rename weight parameters __magic_name__ : int = rename_key_and_reshape_tensor( _a, _a, _a, _a ) # add model prefix if necessary __magic_name__ : Optional[Any] = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __magic_name__ : Any = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' f'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: __magic_name__ : Dict = jnp.asarray(_a ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_a, _a ) continue # also add unexpected weight so that warning is thrown __magic_name__ : List[Any] = jnp.asarray(_a ) else: # also add unexpected weight so that warning is thrown __magic_name__ : str = jnp.asarray(_a ) return unflatten_dict(_a ) def UpperCamelCase ( _A, _A ): """simple docstring""" import torch # Load the index __magic_name__ : int = {} for shard_file in shard_filenames: # load using msgpack utils __magic_name__ : Union[str, Any] = torch.load(_a ) __magic_name__ : Tuple = {k: v.numpy() for k, v in pt_state_dict.items()} __magic_name__ : Tuple = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: __magic_name__ : str = flax_model.params["params"] __magic_name__ : Union[str, Any] = flatten_dict(_a ) random_flax_state_dict.update(flatten_dict(flax_model.params["""batch_stats"""] ) ) else: __magic_name__ : Any = flax_model.params __magic_name__ : Tuple = flatten_dict(_a ) __magic_name__ : List[Any] = (model_prefix not in flax_model_params) and ( model_prefix in {k.split(""".""" )[0] for k in pt_state_dict.keys()} ) __magic_name__ : Union[str, Any] = (model_prefix in flax_model_params) and ( model_prefix not in {k.split(""".""" )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): __magic_name__ : int = tuple(pt_key.split(""".""" ) ) # remove base model prefix if necessary __magic_name__ : Optional[int] = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: __magic_name__ : Tuple = pt_tuple_key[1:] # Correctly rename weight parameters __magic_name__ : Optional[int] = rename_key_and_reshape_tensor( _a, _a, _a, _a ) # add model prefix if necessary __magic_name__ : Optional[int] = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: __magic_name__ : Optional[Any] = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( f'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' f'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: __magic_name__ : Dict = jnp.asarray(_a ) continue if "var" in flax_key[-1]: __magic_name__ : Tuple = jnp.asarray(_a ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_a, _a ) continue # also add unexpected weight so that warning is thrown __magic_name__ : int = jnp.asarray(_a ) else: # also add unexpected weight so that warning is thrown __magic_name__ : str = jnp.asarray(_a ) return unflatten_dict(_a ) def UpperCamelCase ( _A, _A ): """simple docstring""" __magic_name__ : str = os.path.abspath(_a ) logger.info(f'Loading Flax weights from {flax_checkpoint_path}' ) # import correct flax class __magic_name__ : int = getattr(_a, """Flax""" + model.__class__.__name__ ) # load flax weight dict with open(_a, """rb""" ) as state_f: try: __magic_name__ : str = from_bytes(_a, state_f.read() ) except UnpicklingError: raise EnvironmentError(f'Unable to convert {flax_checkpoint_path} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(_a, _a ) def UpperCamelCase ( _A, _A ): """simple docstring""" try: import torch # noqa: F401 except ImportError: logger.error( """Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see""" """ https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation""" """ instructions.""" ) raise # check if we have bf16 weights __magic_name__ : int = flatten_dict(jax.tree_util.tree_map(lambda _A : x.dtype == jnp.bfloataa, _a ) ).values() if any(_a ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( """Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` """ """before loading those in PyTorch model.""" ) __magic_name__ : Optional[Any] = jax.tree_util.tree_map( lambda _A : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params, _a ) __magic_name__ : Optional[Any] = flatten_dict(_a ) __magic_name__ : Optional[int] = pt_model.state_dict() __magic_name__ : Union[str, Any] = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split(""".""" )[0] for k in pt_model_dict.keys()} ) __magic_name__ : Dict = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split(""".""" )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys __magic_name__ : int = [] __magic_name__ : int = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): __magic_name__ : Dict = flax_key_tuple[0] == pt_model.base_model_prefix __magic_name__ : Union[str, Any] = ".".join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: __magic_name__ : Union[str, Any] = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: __magic_name__ : Tuple = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(_a ) not in pt_model_dict: # conv layer __magic_name__ : List[str] = flax_key_tuple[:-1] + ("weight",) __magic_name__ : Tuple = jnp.transpose(_a, (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(_a ) not in pt_model_dict: # linear layer __magic_name__ : int = flax_key_tuple[:-1] + ("weight",) __magic_name__ : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: __magic_name__ : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: __magic_name__ : Union[str, Any] = flax_key_tuple[:-1] + ("running_mean",) elif "var" in flax_key_tuple[-1]: __magic_name__ : Union[str, Any] = flax_key_tuple[:-1] + ("running_var",) if "batch_stats" in flax_state: __magic_name__ : Dict = ".".join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: __magic_name__ : List[str] = ".".join(_a ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. __magic_name__ : Optional[int] = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: __magic_name__ : int = key.split(""".""" ) __magic_name__ : int = None if key_components[-3::2] == ["parametrizations", "original0"]: __magic_name__ : Union[str, Any] = key_components[-2] + "_g" elif key_components[-3::2] == ["parametrizations", "original1"]: __magic_name__ : Dict = key_components[-2] + "_v" if name is not None: __magic_name__ : Optional[int] = key_components[:-3] + [name] __magic_name__ : List[str] = ".".join(_a ) __magic_name__ : str = key if flax_key in special_pt_names: __magic_name__ : Optional[Any] = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( f'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' f'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict __magic_name__ : Dict = np.asarray(_a ) if not isinstance(_a, np.ndarray ) else flax_tensor __magic_name__ : int = torch.from_numpy(_a ) # remove from missing keys missing_keys.remove(_a ) else: # weight is not expected by PyTorch model unexpected_keys.append(_a ) pt_model.load_state_dict(_a ) # re-transform missing_keys to list __magic_name__ : str = list(_a ) if len(_a ) > 0: logger.warning( """Some weights of the Flax model were not used when initializing the PyTorch model""" f' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' f' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' """ (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This""" f' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' """ to be exactly identical (e.g. initializing a BertForSequenceClassification model from a""" """ FlaxBertForSequenceClassification model).""" ) else: logger.warning(f'All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n' ) if len(_a ) > 0: logger.warning( f'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' f' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' """ use it for predictions and inference.""" ) else: logger.warning( f'All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n' """If your task is similar to the task the model of the checkpoint was trained on, """ f'you can already use {pt_model.__class__.__name__} for predictions without further training.' ) return pt_model
342
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Any = tokenizer(example["content"] , truncation=_a)["input_ids"] SCREAMING_SNAKE_CASE : Dict = len(example["content"]) / len(output["input_ids"]) return output a_ = HfArgumentParser(PretokenizationArguments) a_ = parser.parse_args() if args.num_workers is None: a_ = multiprocessing.cpu_count() a_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) a_ = time.time() a_ = load_dataset(args.dataset_name, split='train') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a_ = time.time() a_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ 'repo_name', 'path', 'copies', 'size', 'content', 'license', 'hash', 'line_mean', 'line_max', 'alpha_frac', 'autogenerated', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
76
0
'''simple docstring''' from __future__ import annotations from collections import Counter from random import random class lowercase_ : def __init__( self ): UpperCamelCase__ = {} def __a ( self , a ): UpperCamelCase__ = {} def __a ( self , a , a , a ): if nodea not in self.connections: self.add_node(a ) if nodea not in self.connections: self.add_node(a ) UpperCamelCase__ = probability def __a ( self ): return list(self.connections ) def __a ( self , a ): UpperCamelCase__ = 0 UpperCamelCase__ = random() for dest in self.connections[node]: current_probability += self.connections[node][dest] if current_probability > random_value: return dest return "" def _UpperCamelCase ( __A , __A , __A ) -> Union[str, Any]: '''simple docstring''' UpperCamelCase__ = MarkovChainGraphUndirectedUnweighted() for nodea, nodea, probability in transitions: graph.add_transition_probability(_a , _a , _a ) UpperCamelCase__ = Counter(graph.get_nodes() ) UpperCamelCase__ = start for _ in range(_a ): UpperCamelCase__ = graph.transition(_a ) visited[node] += 1 return visited if __name__ == "__main__": import doctest doctest.testmod()
80
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): # initialize config if "resnet-50" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-50") elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-101") else: raise ValueError("Model name should include either resnet50 or resnet101") SCREAMING_SNAKE_CASE : str = DetrConfig(use_timm_backbone=_a , backbone_config=_a) # set label attributes SCREAMING_SNAKE_CASE : List[str] = "panoptic" in model_name if is_panoptic: SCREAMING_SNAKE_CASE : Union[str, Any] = 250 else: SCREAMING_SNAKE_CASE : Union[str, Any] = 91 SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : Union[str, Any] = "coco-detection-id2label.json" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(_a , _a , repo_type="dataset") , "r")) SCREAMING_SNAKE_CASE : int = {int(_a): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCamelCase__ ( _a): # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE : Union[str, Any] = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight")) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight")) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias")) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean")) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var")) # stages for stage_idx in range(len(config.backbone_config.depths)): for layer_idx in range(config.backbone_config.depths[stage_idx]): # shortcut if layer_idx == 0: rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var", )) # 3 convs for i in range(3): rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var", )) # fmt: on for i in range(config.encoder_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", )) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", )) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ]) return rename_keys def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : str = state_dict.pop(_a) SCREAMING_SNAKE_CASE : int = val def lowerCamelCase__ ( _a , _a=False): SCREAMING_SNAKE_CASE : Optional[Any] = "" if is_panoptic: SCREAMING_SNAKE_CASE : Optional[int] = "detr." # first: transformer encoder for i in range(6): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : int = in_proj_bias[:256] SCREAMING_SNAKE_CASE : Tuple = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : str = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : str = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[:256] SCREAMING_SNAKE_CASE : List[Any] = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : Any = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight") SCREAMING_SNAKE_CASE : int = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias") # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE : Tuple = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[-256:] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Union[str, Any] = Image.open(requests.get(_a , stream=_a).raw) return im @torch.no_grad() def lowerCamelCase__ ( _a , _a=None , _a=False): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Optional[int] = get_detr_config(_a) # load original model from torch hub SCREAMING_SNAKE_CASE : Union[str, Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f"Converting model {model_name}...") SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=_a).eval() SCREAMING_SNAKE_CASE : Tuple = detr.state_dict() # rename keys for src, dest in create_rename_keys(_a): if is_panoptic: SCREAMING_SNAKE_CASE : List[str] = "detr." + src rename_key(_a , _a , _a) # query, key and value matrices need special treatment read_in_q_k_v(_a , is_panoptic=_a) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE : List[Any] = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr") and not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor") ): SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Optional[int] = val elif key.startswith("bbox_attention") or key.startswith("mask_head"): continue else: SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : List[Any] = val else: if not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor"): SCREAMING_SNAKE_CASE : Any = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Any = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE : int = DetrForSegmentation(_a) if is_panoptic else DetrForObjectDetection(_a) model.load_state_dict(_a) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE : int = "coco_panoptic" if is_panoptic else "coco_detection" SCREAMING_SNAKE_CASE : Optional[int] = DetrImageProcessor(format=_a) SCREAMING_SNAKE_CASE : List[str] = processor(images=prepare_img() , return_tensors="pt") SCREAMING_SNAKE_CASE : Any = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = detr(_a) SCREAMING_SNAKE_CASE : Any = model(_a) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4) print("Looks ok!") if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}...") Path(_a).mkdir(exist_ok=_a) model.save_pretrained(_a) processor.save_pretrained(_a) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub...") model.push_to_hub(f"nielsr/{model_name}") processor.push_to_hub(f"nielsr/{model_name}") if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') a_ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
76
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_torch_available __snake_case = {"""configuration_speech_encoder_decoder""": ["""SpeechEncoderDecoderConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""SpeechEncoderDecoderModel"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __snake_case = ["""FlaxSpeechEncoderDecoderModel"""] if TYPE_CHECKING: from .configuration_speech_encoder_decoder import SpeechEncoderDecoderConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_encoder_decoder import SpeechEncoderDecoderModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_speech_encoder_decoder import FlaxSpeechEncoderDecoderModel else: import sys __snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
259
import os def lowerCamelCase__ ( ): with open(os.path.dirname(_a) + "/p022_names.txt") as file: SCREAMING_SNAKE_CASE : List[str] = str(file.readlines()[0]) SCREAMING_SNAKE_CASE : List[Any] = names.replace("\"" , "").split(",") names.sort() SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Dict = 0 for i, name in enumerate(_a): for letter in name: name_score += ord(_a) - 64 total_score += (i + 1) * name_score SCREAMING_SNAKE_CASE : str = 0 return total_score if __name__ == "__main__": print(solution())
76
0
def lowerCamelCase_ ( lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): if isinstance(_a , _a ) and isinstance(_a , _a ): lowerCamelCase_ = len(set_a.intersection(_a ) ) if alternative_union: lowerCamelCase_ = len(_a ) + len(_a ) else: lowerCamelCase_ = len(set_a.union(_a ) ) return intersection / union if isinstance(_a , (list, tuple) ) and isinstance(_a , (list, tuple) ): lowerCamelCase_ = [element for element in set_a if element in set_b] if alternative_union: lowerCamelCase_ = len(_a ) + len(_a ) return len(_a ) / union else: lowerCamelCase_ = set_a + [element for element in set_b if element not in set_a] return len(_a ) / len(_a ) return len(_a ) / len(_a ) return None if __name__ == "__main__": __A ={'''a''', '''b''', '''c''', '''d''', '''e'''} __A ={'''c''', '''d''', '''e''', '''f''', '''h''', '''i'''} print(jaccard_similarity(set_a, set_b))
19
from collections.abc import Callable import numpy as np def lowerCamelCase__ ( _a , _a , _a , _a , _a): SCREAMING_SNAKE_CASE : Dict = int(np.ceil((x_end - xa) / step_size)) SCREAMING_SNAKE_CASE : Tuple = np.zeros((n + 1,)) SCREAMING_SNAKE_CASE : int = ya SCREAMING_SNAKE_CASE : int = xa for k in range(_a): SCREAMING_SNAKE_CASE : Any = y[k] + step_size * ode_func(_a , y[k]) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
76
0
import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class __snake_case ( __A ,__A ): __lowerCamelCase = 1 @register_to_config def __init__( self , __UpperCamelCase = 1000 , __UpperCamelCase = None ) -> Any: '''simple docstring''' self.set_timesteps(__UpperCamelCase ) # standard deviation of the initial noise distribution snake_case__ : int = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. snake_case__ : str = 4 # running values snake_case__ : Optional[int] = [] def __a ( self , __UpperCamelCase , __UpperCamelCase = None ) -> Optional[int]: '''simple docstring''' snake_case__ : Union[str, Any] = num_inference_steps snake_case__ : Tuple = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] snake_case__ : Dict = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: snake_case__ : Optional[Any] = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: snake_case__ : Tuple = torch.sin(steps * math.pi / 2 ) ** 2 snake_case__ : Any = (1.0 - self.betas**2) ** 0.5 snake_case__ : List[Any] = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] snake_case__ : Optional[Any] = timesteps.to(__UpperCamelCase ) snake_case__ : str = [] def __a ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase = True , ) -> Union[SchedulerOutput, Tuple]: '''simple docstring''' if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) snake_case__ : Dict = (self.timesteps == timestep).nonzero().item() snake_case__ : Tuple = timestep_index + 1 snake_case__ : Union[str, Any] = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(__UpperCamelCase ) if len(self.ets ) == 1: snake_case__ : List[Any] = self.ets[-1] elif len(self.ets ) == 2: snake_case__ : Tuple = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: snake_case__ : int = (23 * self.ets[-1] - 16 * self.ets[-2] + 5 * self.ets[-3]) / 12 else: snake_case__ : Optional[int] = (1 / 24) * (55 * self.ets[-1] - 59 * self.ets[-2] + 37 * self.ets[-3] - 9 * self.ets[-4]) snake_case__ : List[str] = self._get_prev_sample(__UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=__UpperCamelCase ) def __a ( self , __UpperCamelCase , *__UpperCamelCase , **__UpperCamelCase ) -> torch.FloatTensor: '''simple docstring''' return sample def __a ( self , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase , __UpperCamelCase ) -> Any: '''simple docstring''' snake_case__ : Dict = self.alphas[timestep_index] snake_case__ : Any = self.betas[timestep_index] snake_case__ : str = self.alphas[prev_timestep_index] snake_case__ : str = self.betas[prev_timestep_index] snake_case__ : List[Any] = (sample - sigma * ets) / max(__UpperCamelCase , 1E-8 ) snake_case__ : Optional[int] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ) -> Union[str, Any]: '''simple docstring''' return self.config.num_train_timesteps
143
def lowerCamelCase__ ( _a , _a): return int((input_a, input_a).count(1) != 0) def lowerCamelCase__ ( ): assert or_gate(0 , 0) == 0 assert or_gate(0 , 1) == 1 assert or_gate(1 , 0) == 1 assert or_gate(1 , 1) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
76
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer A_ :Optional[Any] = logging.get_logger(__name__) A_ :Union[str, Any] = {'''vocab_file''': '''vocab.txt'''} A_ :str = { '''vocab_file''': { '''YituTech/conv-bert-base''': '''https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt''', '''YituTech/conv-bert-medium-small''': ( '''https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt''' ), '''YituTech/conv-bert-small''': '''https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt''', } } A_ :List[Any] = { '''YituTech/conv-bert-base''': 512, '''YituTech/conv-bert-medium-small''': 512, '''YituTech/conv-bert-small''': 512, } A_ :Optional[int] = { '''YituTech/conv-bert-base''': {'''do_lower_case''': True}, '''YituTech/conv-bert-medium-small''': {'''do_lower_case''': True}, '''YituTech/conv-bert-small''': {'''do_lower_case''': True}, } class __A ( __A ): """simple docstring""" UpperCamelCase__ : List[Any] =VOCAB_FILES_NAMES UpperCamelCase__ : Optional[Any] =PRETRAINED_VOCAB_FILES_MAP UpperCamelCase__ : List[str] =PRETRAINED_INIT_CONFIGURATION UpperCamelCase__ : List[str] =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase__ : List[Any] =ConvBertTokenizer def __init__( self , lowerCamelCase__=None , lowerCamelCase__=None , lowerCamelCase__=True , lowerCamelCase__="[UNK]" , lowerCamelCase__="[SEP]" , lowerCamelCase__="[PAD]" , lowerCamelCase__="[CLS]" , lowerCamelCase__="[MASK]" , lowerCamelCase__=True , lowerCamelCase__=None , **lowerCamelCase__ , ): """simple docstring""" super().__init__( lowerCamelCase__ , tokenizer_file=lowerCamelCase__ , do_lower_case=lowerCamelCase__ , unk_token=lowerCamelCase__ , sep_token=lowerCamelCase__ , pad_token=lowerCamelCase__ , cls_token=lowerCamelCase__ , mask_token=lowerCamelCase__ , tokenize_chinese_chars=lowerCamelCase__ , strip_accents=lowerCamelCase__ , **lowerCamelCase__ , ) __UpperCamelCase : Optional[Any] =json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get('lowercase' , lowerCamelCase__ ) != do_lower_case or normalizer_state.get('strip_accents' , lowerCamelCase__ ) != strip_accents or normalizer_state.get('handle_chinese_chars' , lowerCamelCase__ ) != tokenize_chinese_chars ): __UpperCamelCase : List[str] =getattr(lowerCamelCase__ , normalizer_state.pop('type' ) ) __UpperCamelCase : Optional[Any] =do_lower_case __UpperCamelCase : Any =strip_accents __UpperCamelCase : Optional[int] =tokenize_chinese_chars __UpperCamelCase : List[str] =normalizer_class(**lowerCamelCase__ ) __UpperCamelCase : str =do_lower_case def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__=None ): """simple docstring""" __UpperCamelCase : str =[self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : List[str] =[self.sep_token_id] __UpperCamelCase : Union[str, Any] =[self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ = None ): """simple docstring""" __UpperCamelCase : Dict =self._tokenizer.model.save(lowerCamelCase__ , name=lowerCamelCase__ ) return tuple(lowerCamelCase__ )
71
a_ = 8.314_4598 def lowerCamelCase__ ( _a , _a): if temperature < 0: raise Exception("Temperature cannot be less than 0 K") if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol") else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a_ = 300 a_ = 28 a_ = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
76
0
"""simple docstring""" # Lint as: python3 import dataclasses import re from dataclasses import dataclass from functools import total_ordering from typing import Optional, Union lowercase_ = re.compile(r"^(?P<major>\d+)" r"\.(?P<minor>\d+)" r"\.(?P<patch>\d+)$") @total_ordering @dataclass class __lowerCAmelCase : '''simple docstring''' __UpperCAmelCase : Union[str, Any] = 4_2 __UpperCAmelCase : Union[str, Any] = None __UpperCAmelCase : List[str] = None __UpperCAmelCase : Optional[int] = None __UpperCAmelCase : int = None def __UpperCAmelCase ( self ): __a = _str_to_version_tuple(self.version_str ) def __repr__( self ): return f'''{self.tuple[0]}.{self.tuple[1]}.{self.tuple[2]}''' @property def __UpperCAmelCase ( self ): return self.major, self.minor, self.patch def __UpperCAmelCase ( self , _a ): if isinstance(_a , _a ): return Version(_a ) elif isinstance(_a , _a ): return other raise TypeError(f'''{other} (type {type(_a )}) cannot be compared to version.''' ) def __eq__( self , _a ): try: __a = self._validate_operand(_a ) except (TypeError, ValueError): return False else: return self.tuple == other.tuple def __lt__( self , _a ): __a = self._validate_operand(_a ) return self.tuple < other.tuple def __hash__( self ): return hash(_version_tuple_to_str(self.tuple ) ) @classmethod def __UpperCAmelCase ( cls , _a ): __a = {f.name for f in dataclasses.fields(cls )} return cls(**{k: v for k, v in dic.items() if k in field_names} ) def __UpperCAmelCase ( self ): return self.version_str def lowercase ( lowerCAmelCase__ : Dict ) -> str: __a = _VERSION_REG.match(_a ) if not res: raise ValueError(f'''Invalid version \'{version_str}\'. Format should be x.y.z with {{x,y,z}} being digits.''' ) return tuple(int(_a ) for v in [res.group('''major''' ), res.group('''minor''' ), res.group('''patch''' )] ) def lowercase ( lowerCAmelCase__ : List[str] ) -> List[Any]: return ".".join(str(_a ) for v in version_tuple )
45
a_ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : int = set() # keep track of all the paths to be checked SCREAMING_SNAKE_CASE : int = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) # get the last node from the path SCREAMING_SNAKE_CASE : Union[str, Any] = path[-1] if node not in explored: SCREAMING_SNAKE_CASE : List[str] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: SCREAMING_SNAKE_CASE : List[Any] = list(_a) new_path.append(_a) queue.append(_a) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_a) # in case there's no path between the 2 nodes return [] def lowerCamelCase__ ( _a , _a , _a): if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 SCREAMING_SNAKE_CASE : str = [start] SCREAMING_SNAKE_CASE : Optional[Any] = set(_a) # Keep tab on distances from `start` node. SCREAMING_SNAKE_CASE : Union[str, Any] = {start: 0, target: -1} while queue: SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) if node == target: SCREAMING_SNAKE_CASE : Union[str, Any] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node]) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_a) queue.append(_a) SCREAMING_SNAKE_CASE : Optional[Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
76
0
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def __lowerCamelCase ( snake_case__ ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = tokenizer(example["""content"""] ,truncation=_a )["input_ids"] _SCREAMING_SNAKE_CASE = len(example["""content"""] ) / len(output["""input_ids"""] ) return output UpperCamelCase = HfArgumentParser(PretokenizationArguments) UpperCamelCase = parser.parse_args() if args.num_workers is None: UpperCamelCase = multiprocessing.cpu_count() UpperCamelCase = AutoTokenizer.from_pretrained(args.tokenizer_dir) UpperCamelCase = time.time() UpperCamelCase = load_dataset(args.dataset_name, split='''train''') print(f"Dataset loaded in {time.time()-t_start:.2f}s") UpperCamelCase = time.time() UpperCamelCase = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ '''repo_name''', '''path''', '''copies''', '''size''', '''content''', '''license''', '''hash''', '''line_mean''', '''line_max''', '''alpha_frac''', '''autogenerated''', ], ) print(f"Dataset tokenized in {time.time()-t_start:.2f}s") UpperCamelCase = time.time() ds.push_to_hub(args.tokenized_data_repo) print(f"Data pushed to the hub in {time.time()-t_start:.2f}s")
306
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCamelCase ( self : str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : List[str] = AutoTokenizer.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : Tuple = tokenizer("Hello there" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer("Hi I am" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : str = model(a , labels=a ).loss SCREAMING_SNAKE_CASE : Any = -tf.math.reduce_mean(a ).numpy() SCREAMING_SNAKE_CASE : Union[str, Any] = -21.22_8168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
76
0
from typing import Optional, Tuple import jax import jax.numpy as jnp from flax import linen as nn from flax.core.frozen_dict import FrozenDict from transformers import CLIPConfig, FlaxPreTrainedModel from transformers.models.clip.modeling_flax_clip import FlaxCLIPVisionModule def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase=1E-12 ) -> Tuple: __lowercase : Optional[Any] = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(_a , axis=1 ) , a_min=_a ) ).T __lowercase : Any = jnp.divide(emb_a.T , jnp.clip(jnp.linalg.norm(_a , axis=1 ) , a_min=_a ) ).T return jnp.matmul(_a , norm_emb_a.T ) class __lowerCAmelCase ( nn.Module ): """simple docstring""" A__ : int = 4_2 A__ : str = jnp.floataa def snake_case_ ( self : List[Any] ): __lowercase : Dict = FlaxCLIPVisionModule(self.config.vision_config ) __lowercase : Any = nn.Dense(self.config.projection_dim , use_bias=_snake_case , dtype=self.dtype ) __lowercase : str = self.param('''concept_embeds''' , jax.nn.initializers.ones , (17, self.config.projection_dim) ) __lowercase : Optional[int] = self.param( '''special_care_embeds''' , jax.nn.initializers.ones , (3, self.config.projection_dim) ) __lowercase : List[Any] = self.param('''concept_embeds_weights''' , jax.nn.initializers.ones , (17,) ) __lowercase : Any = self.param('''special_care_embeds_weights''' , jax.nn.initializers.ones , (3,) ) def __call__( self : Optional[Any] , _snake_case : List[str] ): __lowercase : Union[str, Any] = self.vision_model(_snake_case )[1] __lowercase : Optional[Any] = self.visual_projection(_snake_case ) __lowercase : List[Any] = jax_cosine_distance(_snake_case , self.special_care_embeds ) __lowercase : str = jax_cosine_distance(_snake_case , self.concept_embeds ) # increase this value to create a stronger `nfsw` filter # at the cost of increasing the possibility of filtering benign image inputs __lowercase : Optional[int] = 0.0 __lowercase : List[str] = special_cos_dist - self.special_care_embeds_weights[None, :] + adjustment __lowercase : str = jnp.round(_snake_case , 3 ) __lowercase : Optional[Any] = jnp.any(special_scores > 0 , axis=1 , keepdims=_snake_case ) # Use a lower threshold if an image has any special care concept __lowercase : Union[str, Any] = is_special_care * 0.01 __lowercase : Union[str, Any] = cos_dist - self.concept_embeds_weights[None, :] + special_adjustment __lowercase : Dict = jnp.round(_snake_case , 3 ) __lowercase : int = jnp.any(concept_scores > 0 , axis=1 ) return has_nsfw_concepts class __lowerCAmelCase ( __A ): """simple docstring""" A__ : Any = CLIPConfig A__ : int = '''clip_input''' A__ : Tuple = FlaxStableDiffusionSafetyCheckerModule def __init__( self : str , _snake_case : CLIPConfig , _snake_case : Optional[Tuple] = None , _snake_case : int = 0 , _snake_case : jnp.dtype = jnp.floataa , _snake_case : bool = True , **_snake_case : str , ): if input_shape is None: __lowercase : List[Any] = (1, 224, 224, 3) __lowercase : Optional[int] = self.module_class(config=_snake_case , dtype=_snake_case , **_snake_case ) super().__init__(_snake_case , _snake_case , input_shape=_snake_case , seed=_snake_case , dtype=_snake_case , _do_init=_do_init ) def snake_case_ ( self : str , _snake_case : jax.random.KeyArray , _snake_case : Tuple , _snake_case : FrozenDict = None ): __lowercase : Union[str, Any] = jax.random.normal(_snake_case , _snake_case ) __lowercase : Tuple = jax.random.split(_snake_case ) __lowercase : Union[str, Any] = {"params": params_rng, "dropout": dropout_rng} __lowercase : int = self.module.init(_snake_case , _snake_case )["params"] return random_params def __call__( self : List[Any] , _snake_case : Optional[int] , _snake_case : dict = None , ): __lowercase : Any = jnp.transpose(_snake_case , (0, 2, 3, 1) ) return self.module.apply( {'''params''': params or self.params} , jnp.array(_snake_case , dtype=jnp.floataa ) , rngs={} , )
156
from math import factorial def lowerCamelCase__ ( _a , _a , _a): if successes > trials: raise ValueError("successes must be lower or equal to trials") if trials < 0 or successes < 0: raise ValueError("the function is defined for non-negative integers") if not isinstance(_a , _a) or not isinstance(_a , _a): raise ValueError("the function is defined for non-negative integers") if not 0 < prob < 1: raise ValueError("prob has to be in range of 1 - 0") SCREAMING_SNAKE_CASE : int = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! SCREAMING_SNAKE_CASE : List[Any] = float(factorial(_a)) coefficient /= factorial(_a) * factorial(trials - successes) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('Probability of 2 successes out of 4 trails') print('with probability of 0.75 is:', end=' ') print(binomial_distribution(2, 4, 0.75))
76
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class A ( unittest.TestCase ): """simple docstring""" def __init__( self : Tuple,lowercase_ : int,lowercase_ : Optional[int]=1_3,lowercase_ : Optional[int]=3,lowercase_ : int=2_2_4,lowercase_ : Optional[int]=3_0,lowercase_ : int=4_0_0,lowercase_ : Union[str, Any]=True,lowercase_ : int=None,lowercase_ : Tuple=True,lowercase_ : Tuple=[0.5, 0.5, 0.5],lowercase_ : Optional[int]=[0.5, 0.5, 0.5],)-> Optional[Any]: '''simple docstring''' A__ = size if size is not None else {"height": 1_8, "width": 1_8} A__ = parent A__ = batch_size A__ = num_channels A__ = image_size A__ = min_resolution A__ = max_resolution A__ = do_resize A__ = size A__ = do_normalize A__ = image_mean A__ = image_std def snake_case__ ( self : Any )-> Optional[int]: '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class A ( __A , unittest.TestCase ): """simple docstring""" lowerCamelCase = ViTImageProcessor if is_vision_available() else None def snake_case__ ( self : Union[str, Any] )-> List[Any]: '''simple docstring''' A__ = EfficientFormerImageProcessorTester(self ) @property def snake_case__ ( self : Any )-> List[str]: '''simple docstring''' return self.image_proc_tester.prepare_image_processor_dict() def snake_case__ ( self : List[Any] )-> Dict: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_,'image_mean' ) ) self.assertTrue(hasattr(lowercase_,'image_std' ) ) self.assertTrue(hasattr(lowercase_,'do_normalize' ) ) self.assertTrue(hasattr(lowercase_,'do_resize' ) ) self.assertTrue(hasattr(lowercase_,'size' ) ) def snake_case__ ( self : int )-> str: '''simple docstring''' pass def snake_case__ ( self : Optional[Any] )-> List[Any]: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ = prepare_image_inputs(self.image_proc_tester,equal_resolution=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,Image.Image ) # Test not batched input A__ = image_processor(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),) # Test batched A__ = image_processor(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),) def snake_case__ ( self : List[str] )-> Any: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ = prepare_image_inputs(self.image_proc_tester,equal_resolution=lowercase_,numpify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,np.ndarray ) # Test not batched input A__ = image_processor(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),) # Test batched A__ = image_processor(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),) def snake_case__ ( self : List[str] )-> List[str]: '''simple docstring''' A__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ = prepare_image_inputs(self.image_proc_tester,equal_resolution=lowercase_,torchify=lowercase_ ) for image in image_inputs: self.assertIsInstance(lowercase_,torch.Tensor ) # Test not batched input A__ = image_processor(image_inputs[0],return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),) # Test batched A__ = image_processor(lowercase_,return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape,( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size['height'], self.image_proc_tester.size['width'], ),)
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
# This script creates a super tiny model that is useful inside tests, when we just want to test that # the machinery works, without needing to the check the quality of the outcomes. # # This version creates a tiny model through reduction of a normal pre-trained model, but keeping the # full vocab, merges file, and thus also resulting in a larger model due to a large vocab size. # This gives ~3MB in total for all files. # # If you want a 50 times smaller than this see `fsmt-make-super-tiny-model.py`, which is slightly more complicated # # # It will be used then as "stas/tiny-wmt19-en-de" # Build from transformers import FSMTTokenizer, FSMTConfig, FSMTForConditionalGeneration __A : Optional[Any] = "facebook/wmt19-en-de" __A : Union[str, Any] = FSMTTokenizer.from_pretrained(mname) # get the correct vocab sizes, etc. from the master model __A : Union[str, Any] = FSMTConfig.from_pretrained(mname) config.update( dict( d_model=4, encoder_layers=1, decoder_layers=1, encoder_ffn_dim=4, decoder_ffn_dim=4, encoder_attention_heads=1, decoder_attention_heads=1, ) ) __A : int = FSMTForConditionalGeneration(config) print(F'num of params {tiny_model.num_parameters()}') # Test __A : List[Any] = tokenizer(["Making tiny model"], return_tensors="pt") __A : List[str] = tiny_model(**batch) print("test output:", len(outputs.logits[0])) # Save __A : List[str] = "tiny-wmt19-en-de" tiny_model.half() # makes it smaller tiny_model.save_pretrained(mname_tiny) tokenizer.save_pretrained(mname_tiny) print(F'Generated {mname_tiny}') # Upload # transformers-cli upload tiny-wmt19-en-de
273
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a_ = logging.getLogger(__name__) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = False def __UpperCamelCase ( self : str , a : str , a : Optional[int] , a : Any , a : str ) -> List[Any]: """simple docstring""" if not self.initialized: SCREAMING_SNAKE_CASE : List[str] = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[int] = True def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def __UpperCamelCase ( self : Optional[Any] , a : List[Any] , a : Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Dict = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : Tuple , a : Any , a : Tuple , a : Tuple , a : Tuple , a : List[Any]=None ) -> Optional[int]: """simple docstring""" if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def __UpperCamelCase ( self : Any ) -> Dict: """simple docstring""" logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __UpperCamelCase ( self : Tuple , a : Optional[int] , a : Any ) -> int: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. SCREAMING_SNAKE_CASE : Optional[Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = ray.get(random_worker.retrieve.remote(a , a ) ) else: SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Any = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def __UpperCamelCase ( cls : str , a : Optional[Any] , a : Any=None , **a : List[Any] ) -> str: """simple docstring""" return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def __UpperCamelCase ( cls : Union[str, Any] , a : int , a : Any , a : List[Any]=None , **a : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : str = kwargs.pop("config" , a ) or RagConfig.from_pretrained(a , **a ) SCREAMING_SNAKE_CASE : List[Any] = RagTokenizer.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.question_encoder SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.generator if indexed_dataset is not None: SCREAMING_SNAKE_CASE : str = "custom" SCREAMING_SNAKE_CASE : List[Any] = CustomHFIndex(config.retrieval_vector_size , a ) else: SCREAMING_SNAKE_CASE : List[str] = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
76
0
import warnings from ...utils import logging from .image_processing_flava import FlavaImageProcessor __magic_name__: Optional[int] = logging.get_logger(__name__) class snake_case__ ( __A ): def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> None: warnings.warn( """The class FlavaFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please""" """ use FlavaImageProcessor instead.""" , lowerCAmelCase__ , ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ )
342
from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : Dict , a : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : int = data SCREAMING_SNAKE_CASE : int = None def __repr__( self : str ) -> str: """simple docstring""" return F"Node({self.data})" class _UpperCamelCase : '''simple docstring''' def __init__( self : List[str] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = None def __iter__( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.head while node: yield node.data SCREAMING_SNAKE_CASE : List[str] = node.next def __len__( self : str ) -> int: """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return "->".join([str(a ) for item in self] ) def __getitem__( self : List[Any] , a : int ) -> Any: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Tuple , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) SCREAMING_SNAKE_CASE : str = self.head for _ in range(a ): SCREAMING_SNAKE_CASE : str = current.next SCREAMING_SNAKE_CASE : Any = data def __UpperCamelCase ( self : List[str] , a : Any ) -> None: """simple docstring""" self.insert_nth(len(self ) , a ) def __UpperCamelCase ( self : Union[str, Any] , a : Any ) -> None: """simple docstring""" self.insert_nth(0 , a ) def __UpperCamelCase ( self : Optional[Any] , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) SCREAMING_SNAKE_CASE : Any = Node(a ) if self.head is None: SCREAMING_SNAKE_CASE : Optional[int] = new_node elif index == 0: SCREAMING_SNAKE_CASE : Optional[int] = self.head # link new_node to head SCREAMING_SNAKE_CASE : List[Any] = new_node else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : int = new_node def __UpperCamelCase ( self : Optional[int] ) -> None: # print every node data """simple docstring""" print(self ) def __UpperCamelCase ( self : int ) -> Any: """simple docstring""" return self.delete_nth(0 ) def __UpperCamelCase ( self : Any ) -> Any: # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def __UpperCamelCase ( self : List[str] , a : int = 0 ) -> Any: """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) SCREAMING_SNAKE_CASE : Tuple = self.head # default first node if index == 0: SCREAMING_SNAKE_CASE : List[str] = self.head.next else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Any = temp.next SCREAMING_SNAKE_CASE : List[Any] = temp.next SCREAMING_SNAKE_CASE : List[str] = temp.next.next return delete_node.data def __UpperCamelCase ( self : List[Any] ) -> bool: """simple docstring""" return self.head is None def __UpperCamelCase ( self : Optional[int] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : str = self.head while current: # Store the current node's next node. SCREAMING_SNAKE_CASE : Any = current.next # Make the current node's next point backwards SCREAMING_SNAKE_CASE : List[Any] = prev # Make the previous node be the current node SCREAMING_SNAKE_CASE : Any = current # Make the current node the next node (to progress iteration) SCREAMING_SNAKE_CASE : str = next_node # Return prev in order to put the head at the end SCREAMING_SNAKE_CASE : Optional[Any] = prev def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_a) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10): assert len(_a) == i linked_list.insert_nth(_a , i + 1) assert str(_a) == "->".join(str(_a) for i in range(1 , 11)) linked_list.insert_head(0) linked_list.insert_tail(11) assert str(_a) == "->".join(str(_a) for i in range(0 , 12)) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9) == 10 assert linked_list.delete_tail() == 11 assert len(_a) == 9 assert str(_a) == "->".join(str(_a) for i in range(1 , 10)) assert all(linked_list[i] == i + 1 for i in range(0 , 9)) is True for i in range(0 , 9): SCREAMING_SNAKE_CASE : str = -i assert all(linked_list[i] == -i for i in range(0 , 9)) is True linked_list.reverse() assert str(_a) == "->".join(str(_a) for i in range(-8 , 1)) def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Optional[Any] = [ -9, 100, Node(77345112), "dlrow olleH", 7, 5555, 0, -192.5_5555, "Hello, world!", 77.9, Node(10), None, None, 12.20, ] SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in test_input: linked_list.insert_tail(_a) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_a) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head SCREAMING_SNAKE_CASE : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail SCREAMING_SNAKE_CASE : Any = linked_list.delete_tail() assert result == 12.2 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list SCREAMING_SNAKE_CASE : Any = linked_list.delete_nth(10) assert result is None assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!")) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_a) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_a) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowerCamelCase__ ( ): from doctest import testmod testmod() SCREAMING_SNAKE_CASE : Optional[int] = LinkedList() linked_list.insert_head(input("Inserting 1st at head ").strip()) linked_list.insert_head(input("Inserting 2nd at head ").strip()) print("\nPrint list:") linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail ").strip()) linked_list.insert_tail(input("Inserting 2nd at tail ").strip()) print("\nPrint list:") linked_list.print_list() print("\nDelete head") linked_list.delete_head() print("Delete tail") linked_list.delete_tail() print("\nPrint list:") linked_list.print_list() print("\nReverse linked list") linked_list.reverse() print("\nPrint list:") linked_list.print_list() print("\nString representation of linked list:") print(_a) print("\nReading/changing Node data using indexing:") print(f"Element at Position 1: {linked_list[1]}") SCREAMING_SNAKE_CASE : Dict = input("Enter New Value: ").strip() print("New list:") print(_a) print(f"length of linked_list is : {len(_a)}") if __name__ == "__main__": main()
76
0
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) a__ : Any = { 'configuration_roberta': ['ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RobertaConfig', 'RobertaOnnxConfig'], 'tokenization_roberta': ['RobertaTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = ['RobertaTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Optional[Any] = [ 'ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'RobertaForCausalLM', 'RobertaForMaskedLM', 'RobertaForMultipleChoice', 'RobertaForQuestionAnswering', 'RobertaForSequenceClassification', 'RobertaForTokenClassification', 'RobertaModel', 'RobertaPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : Any = [ 'TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRobertaForCausalLM', 'TFRobertaForMaskedLM', 'TFRobertaForMultipleChoice', 'TFRobertaForQuestionAnswering', 'TFRobertaForSequenceClassification', 'TFRobertaForTokenClassification', 'TFRobertaMainLayer', 'TFRobertaModel', 'TFRobertaPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a__ : int = [ 'FlaxRobertaForCausalLM', 'FlaxRobertaForMaskedLM', 'FlaxRobertaForMultipleChoice', 'FlaxRobertaForQuestionAnswering', 'FlaxRobertaForSequenceClassification', 'FlaxRobertaForTokenClassification', 'FlaxRobertaModel', 'FlaxRobertaPreTrainedModel', ] if TYPE_CHECKING: from .configuration_roberta import ROBERTA_PRETRAINED_CONFIG_ARCHIVE_MAP, RobertaConfig, RobertaOnnxConfig from .tokenization_roberta import RobertaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_roberta_fast import RobertaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_roberta import ( ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, RobertaForCausalLM, RobertaForMaskedLM, RobertaForMultipleChoice, RobertaForQuestionAnswering, RobertaForSequenceClassification, RobertaForTokenClassification, RobertaModel, RobertaPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_roberta import ( TF_ROBERTA_PRETRAINED_MODEL_ARCHIVE_LIST, TFRobertaForCausalLM, TFRobertaForMaskedLM, TFRobertaForMultipleChoice, TFRobertaForQuestionAnswering, TFRobertaForSequenceClassification, TFRobertaForTokenClassification, TFRobertaMainLayer, TFRobertaModel, TFRobertaPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_roberta import ( FlaxRobertaForCausalLM, FlaxRobertaForMaskedLM, FlaxRobertaForMultipleChoice, FlaxRobertaForQuestionAnswering, FlaxRobertaForSequenceClassification, FlaxRobertaForTokenClassification, FlaxRobertaModel, FlaxRobertaPreTrainedModel, ) else: import sys a__ : Dict = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
80
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ = get_logger(__name__) class _UpperCamelCase ( enum.Enum ): '''simple docstring''' lowerCamelCase__ ='all_checks' lowerCamelCase__ ='basic_checks' lowerCamelCase__ ='no_checks' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a , _a=None): if expected_checksums is None: logger.info("Unable to verify checksums.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedDownloadedFile(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE : Tuple = " for " + verification_name if verification_name is not None else "" if len(_a) > 0: raise NonMatchingChecksumError( f"Checksums didn't match{for_verification_name}:\n" f"{bad_urls}\n" "Set `verification_mode='no_checks'` to skip checksums verification and ignore this error") logger.info("All the checksums matched successfully" + for_verification_name) class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a): if expected_splits is None: logger.info("Unable to verify splits sizes.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreSplits(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedSplits(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : List[str] = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_a) > 0: raise NonMatchingSplitsSizesError(str(_a)) logger.info("All the splits matched successfully.") def lowerCamelCase__ ( _a , _a = True): if record_checksum: SCREAMING_SNAKE_CASE : List[str] = shaaaa() with open(_a , "rb") as f: for chunk in iter(lambda: f.read(1 << 20) , b""): m.update(_a) SCREAMING_SNAKE_CASE : Optional[int] = m.hexdigest() else: SCREAMING_SNAKE_CASE : List[str] = None return {"num_bytes": os.path.getsize(_a), "checksum": checksum} def lowerCamelCase__ ( _a): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
76
0
import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets __snake_case = """\\n@inproceedings{lin-2004-rouge,\n title = \"{ROUGE}: A Package for Automatic Evaluation of Summaries\",\n author = \"Lin, Chin-Yew\",\n booktitle = \"Text Summarization Branches Out\",\n month = jul,\n year = \"2004\",\n address = \"Barcelona, Spain\",\n publisher = \"Association for Computational Linguistics\",\n url = \"https://www.aclweb.org/anthology/W04-1013\",\n pages = \"74--81\",\n}\n""" __snake_case = """\\nROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for\nevaluating automatic summarization and machine translation software in natural language processing.\nThe metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation.\n\nNote that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters.\n\nThis metrics is a wrapper around Google Research reimplementation of ROUGE:\nhttps://github.com/google-research/google-research/tree/master/rouge\n""" __snake_case = """\nCalculates average rouge scores for a list of hypotheses and references\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n rouge_types: A list of rouge types to calculate.\n Valid names:\n `\"rouge{n}\"` (e.g. `\"rouge1\"`, `\"rouge2\"`) where: {n} is the n-gram based scoring,\n `\"rougeL\"`: Longest common subsequence based scoring.\n `\"rougeLSum\"`: rougeLsum splits text using `\"\n\"`.\n See details in https://github.com/huggingface/datasets/issues/617\n use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes.\n use_aggregator: Return aggregates if this is set to True\nReturns:\n rouge1: rouge_1 (precision, recall, f1),\n rouge2: rouge_2 (precision, recall, f1),\n rougeL: rouge_l (precision, recall, f1),\n rougeLsum: rouge_lsum (precision, recall, f1)\nExamples:\n\n >>> rouge = datasets.load_metric(\'rouge\')\n >>> predictions = [\"hello there\", \"general kenobi\"]\n >>> references = [\"hello there\", \"general kenobi\"]\n >>> results = rouge.compute(predictions=predictions, references=references)\n >>> print(list(results.keys()))\n [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\']\n >>> print(results[\"rouge1\"])\n AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0))\n >>> print(results[\"rouge1\"].mid.fmeasure)\n 1.0\n""" @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class UpperCAmelCase_ ( datasets.Metric ): """simple docstring""" def UpperCAmelCase ( self ) -> List[str]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/google-research/google-research/tree/master/rouge'''] , reference_urls=[ '''https://en.wikipedia.org/wiki/ROUGE_(metric)''', '''https://github.com/google-research/google-research/tree/master/rouge''', ] , ) def UpperCAmelCase ( self , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_=None , SCREAMING_SNAKE_CASE_=True , SCREAMING_SNAKE_CASE_=False ) -> List[str]: if rouge_types is None: UpperCamelCase :Any = ["rouge1", "rouge2", "rougeL", "rougeLsum"] UpperCamelCase :List[str] = rouge_scorer.RougeScorer(rouge_types=SCREAMING_SNAKE_CASE_ , use_stemmer=SCREAMING_SNAKE_CASE_ ) if use_aggregator: UpperCamelCase :Union[str, Any] = scoring.BootstrapAggregator() else: UpperCamelCase :Union[str, Any] = [] for ref, pred in zip(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ): UpperCamelCase :int = scorer.score(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) if use_aggregator: aggregator.add_scores(SCREAMING_SNAKE_CASE_ ) else: scores.append(SCREAMING_SNAKE_CASE_ ) if use_aggregator: UpperCamelCase :Optional[int] = aggregator.aggregate() else: UpperCamelCase :Tuple = {} for key in scores[0]: UpperCamelCase :List[Any] = [score[key] for score in scores] return result
259
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( _a , _a): # Load checkpoint SCREAMING_SNAKE_CASE : int = torch.load(_a , map_location="cpu") SCREAMING_SNAKE_CASE : Dict = chkpt["model"] # We have the base model one level deeper than the original XLM repository SCREAMING_SNAKE_CASE : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: SCREAMING_SNAKE_CASE : List[str] = v else: SCREAMING_SNAKE_CASE : int = v SCREAMING_SNAKE_CASE : int = chkpt["params"] SCREAMING_SNAKE_CASE : Union[str, Any] = {n: v for n, v in config.items() if not isinstance(_a , (torch.FloatTensor, numpy.ndarray))} SCREAMING_SNAKE_CASE : List[Any] = chkpt["dico_word2id"] SCREAMING_SNAKE_CASE : List[Any] = {s + "</w>" if s.find("@@") == -1 and i > 13 else s.replace("@@" , ""): i for s, i in vocab.items()} # Save pytorch-model SCREAMING_SNAKE_CASE : Tuple = pytorch_dump_folder_path + "/" + WEIGHTS_NAME SCREAMING_SNAKE_CASE : Any = pytorch_dump_folder_path + "/" + CONFIG_NAME SCREAMING_SNAKE_CASE : Optional[int] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(f"Save PyTorch model to {pytorch_weights_dump_path}") torch.save(_a , _a) print(f"Save configuration file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") print(f"Save vocab file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) a_ = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
76
0
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class _SCREAMING_SNAKE_CASE ( __A ): lowerCAmelCase__ = 42 lowerCAmelCase__ = 42 def __init__( self , lowercase , lowercase ) -> Tuple: super().__init__() self.register_modules(unet=lowercase , scheduler=lowercase ) @torch.no_grad() def __call__( self , lowercase = 1 , lowercase = 50 , lowercase = None , lowercase = "pil" , lowercase = True , **lowercase , ) -> Union[Tuple, ImagePipelineOutput]: lowerCamelCase_ = self.unet.config.sample_size lowerCamelCase_ = (batch_size, 3, img_size, img_size) lowerCamelCase_ = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) lowerCamelCase_ = randn_tensor(lowercase , generator=lowercase , device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(lowercase ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper lowerCamelCase_ = self.scheduler.schedule[t] lowerCamelCase_ = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat lowerCamelCase_ = self.scheduler.add_noise_to_input(lowercase , lowercase , generator=lowercase ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. lowerCamelCase_ = (sigma_hat / 2) * model((sample_hat + 1) / 2 , sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev lowerCamelCase_ = self.scheduler.step(lowercase , lowercase , lowercase , lowercase ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. lowerCamelCase_ = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 , sigma_prev / 2 ).sample lowerCamelCase_ = self.scheduler.step_correct( lowercase , lowercase , lowercase , lowercase , step_output.prev_sample , step_output["derivative"] , ) lowerCamelCase_ = step_output.prev_sample lowerCamelCase_ = (sample / 2 + 0.5).clamp(0 , 1 ) lowerCamelCase_ = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": lowerCamelCase_ = self.numpy_to_pil(lowercase ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowercase )
19
def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(a - b) for a, b in zip(_a , _a))) def lowerCamelCase__ ( _a): if point: if isinstance(_a , _a): for item in point: if not isinstance(_a , (int, float)): SCREAMING_SNAKE_CASE : List[Any] = ( "Expected a list of numbers as input, found " f"{type(_a).__name__}" ) raise TypeError(_a) else: SCREAMING_SNAKE_CASE : List[Any] = f"Expected a list of numbers as input, found {type(_a).__name__}" raise TypeError(_a) else: raise ValueError("Missing an input") def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(x - y) for x, y in zip(_a , _a))) if __name__ == "__main__": import doctest doctest.testmod()
76
0
import argparse import os import re import tensorflow as tf import torch from transformers import BertConfig, BertModel from transformers.utils import logging logging.set_verbosity_info() lowerCAmelCase__ : Union[str, Any] = logging.get_logger(__name__) def UpperCamelCase__ ( A__ , A__ , A__ ) -> Tuple: snake_case__ : Dict = os.path.abspath(_a ) logger.info(F"""Converting TensorFlow checkpoint from {tf_path}""" ) # Load weights from TF model snake_case__ : Any = tf.train.list_variables(_a ) snake_case__ : str = [] snake_case__ : Tuple = [] snake_case__ : Optional[Any] = [] for full_name, shape in init_vars: # logger.info(f"Loading TF weight {name} with shape {shape}") snake_case__ : Tuple = full_name.split('/' ) if full_name == "_CHECKPOINTABLE_OBJECT_GRAPH" or name[0] in ["global_step", "save_counter"]: logger.info(F"""Skipping non-model layer {full_name}""" ) continue if "optimizer" in full_name: logger.info(F"""Skipping optimization layer {full_name}""" ) continue if name[0] == "model": # ignore initial 'model' snake_case__ : Optional[Any] = name[1:] # figure out how many levels deep the name is snake_case__ : Dict = 0 for _name in name: if _name.startswith('layer_with_weights' ): depth += 1 else: break layer_depth.append(_a ) # read data snake_case__ : Optional[int] = tf.train.load_variable(_a , _a ) names.append('/'.join(_a ) ) arrays.append(_a ) logger.info(F"""Read a total of {len(_a ):,} layers""" ) # Sanity check if len(set(_a ) ) != 1: raise ValueError(F"""Found layer names with different depths (layer depth {list(set(_a ) )})""" ) snake_case__ : int = list(set(_a ) )[0] if layer_depth != 1: raise ValueError( 'The model contains more than just the embedding/encoder layers. This script does not handle MLM/NSP' ' heads.' ) # convert layers logger.info('Converting weights...' ) for full_name, array in zip(_a , _a ): snake_case__ : int = full_name.split('/' ) snake_case__ : List[str] = model snake_case__ : List[str] = [] for i, m_name in enumerate(_a ): if m_name == ".ATTRIBUTES": # variable names end with .ATTRIBUTES/VARIABLE_VALUE break if m_name.startswith('layer_with_weights' ): snake_case__ : List[str] = int(m_name.split('-' )[-1] ) if layer_num <= 2: # embedding layers # layer_num 0: word_embeddings # layer_num 1: position_embeddings # layer_num 2: token_type_embeddings continue elif layer_num == 3: # embedding LayerNorm trace.extend(['embeddings', 'LayerNorm'] ) snake_case__ : List[Any] = getattr(_a , 'embeddings' ) snake_case__ : Any = getattr(_a , 'LayerNorm' ) elif layer_num > 3 and layer_num < config.num_hidden_layers + 4: # encoder layers trace.extend(['encoder', 'layer', str(layer_num - 4 )] ) snake_case__ : Optional[int] = getattr(_a , 'encoder' ) snake_case__ : Union[str, Any] = getattr(_a , 'layer' ) snake_case__ : str = pointer[layer_num - 4] elif layer_num == config.num_hidden_layers + 4: # pooler layer trace.extend(['pooler', 'dense'] ) snake_case__ : Dict = getattr(_a , 'pooler' ) snake_case__ : Union[str, Any] = getattr(_a , 'dense' ) elif m_name == "embeddings": trace.append('embeddings' ) snake_case__ : Optional[Any] = getattr(_a , 'embeddings' ) if layer_num == 0: trace.append('word_embeddings' ) snake_case__ : str = getattr(_a , 'word_embeddings' ) elif layer_num == 1: trace.append('position_embeddings' ) snake_case__ : Optional[int] = getattr(_a , 'position_embeddings' ) elif layer_num == 2: trace.append('token_type_embeddings' ) snake_case__ : Optional[int] = getattr(_a , 'token_type_embeddings' ) else: raise ValueError(F"""Unknown embedding layer with name {full_name}""" ) trace.append('weight' ) snake_case__ : str = getattr(_a , 'weight' ) elif m_name == "_attention_layer": # self-attention layer trace.extend(['attention', 'self'] ) snake_case__ : List[str] = getattr(_a , 'attention' ) snake_case__ : Union[str, Any] = getattr(_a , 'self' ) elif m_name == "_attention_layer_norm": # output attention norm trace.extend(['attention', 'output', 'LayerNorm'] ) snake_case__ : str = getattr(_a , 'attention' ) snake_case__ : Optional[Any] = getattr(_a , 'output' ) snake_case__ : Tuple = getattr(_a , 'LayerNorm' ) elif m_name == "_attention_output_dense": # output attention dense trace.extend(['attention', 'output', 'dense'] ) snake_case__ : int = getattr(_a , 'attention' ) snake_case__ : List[str] = getattr(_a , 'output' ) snake_case__ : int = getattr(_a , 'dense' ) elif m_name == "_output_dense": # output dense trace.extend(['output', 'dense'] ) snake_case__ : Union[str, Any] = getattr(_a , 'output' ) snake_case__ : str = getattr(_a , 'dense' ) elif m_name == "_output_layer_norm": # output dense trace.extend(['output', 'LayerNorm'] ) snake_case__ : Dict = getattr(_a , 'output' ) snake_case__ : List[str] = getattr(_a , 'LayerNorm' ) elif m_name == "_key_dense": # attention key trace.append('key' ) snake_case__ : Optional[int] = getattr(_a , 'key' ) elif m_name == "_query_dense": # attention query trace.append('query' ) snake_case__ : int = getattr(_a , 'query' ) elif m_name == "_value_dense": # attention value trace.append('value' ) snake_case__ : Union[str, Any] = getattr(_a , 'value' ) elif m_name == "_intermediate_dense": # attention intermediate dense trace.extend(['intermediate', 'dense'] ) snake_case__ : Optional[Any] = getattr(_a , 'intermediate' ) snake_case__ : Any = getattr(_a , 'dense' ) elif m_name == "_output_layer_norm": # output layer norm trace.append('output' ) snake_case__ : Any = getattr(_a , 'output' ) # weights & biases elif m_name in ["bias", "beta"]: trace.append('bias' ) snake_case__ : Optional[int] = getattr(_a , 'bias' ) elif m_name in ["kernel", "gamma"]: trace.append('weight' ) snake_case__ : Optional[int] = getattr(_a , 'weight' ) else: logger.warning(F"""Ignored {m_name}""" ) # for certain layers reshape is necessary snake_case__ : List[str] = ".".join(_a ) if re.match(r'(\S+)\.attention\.self\.(key|value|query)\.(bias|weight)' , _a ) or re.match( r'(\S+)\.attention\.output\.dense\.weight' , _a ): snake_case__ : Optional[Any] = array.reshape(pointer.data.shape ) if "kernel" in full_name: snake_case__ : int = array.transpose() if pointer.shape == array.shape: snake_case__ : Optional[Any] = torch.from_numpy(_a ) else: raise ValueError( F"""Shape mismatch in layer {full_name}: Model expects shape {pointer.shape} but layer contains shape:""" F""" {array.shape}""" ) logger.info(F"""Successfully set variable {full_name} to PyTorch layer {trace}""" ) return model def UpperCamelCase__ ( A__ , A__ , A__ ) -> Union[str, Any]: # Instantiate model logger.info(F"""Loading model based on config from {config_path}...""" ) snake_case__ : Optional[int] = BertConfig.from_json_file(_a ) snake_case__ : Dict = BertModel(_a ) # Load weights from checkpoint logger.info(F"""Loading weights from checkpoint {tf_checkpoint_path}...""" ) load_tfa_weights_in_bert(_a , _a , _a ) # Save pytorch-model logger.info(F"""Saving PyTorch model to {pytorch_dump_path}...""" ) torch.save(model.state_dict() , _a ) if __name__ == "__main__": lowerCAmelCase__ : Optional[int] = argparse.ArgumentParser() parser.add_argument( '''--tf_checkpoint_path''', type=str, required=True, help='''Path to the TensorFlow 2.x checkpoint path.''' ) parser.add_argument( '''--bert_config_file''', type=str, required=True, help='''The config json file corresponding to the BERT model. This specifies the model architecture.''', ) parser.add_argument( '''--pytorch_dump_path''', type=str, required=True, help='''Path to the output PyTorch model (must include filename).''', ) lowerCAmelCase__ : Union[str, Any] = parser.parse_args() convert_tfa_checkpoint_to_pytorch(args.tf_checkpoint_path, args.bert_config_file, args.pytorch_dump_path)
143
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ ='vit_msn' def __init__( self : str , a : Tuple=768 , a : Tuple=12 , a : Any=12 , a : int=3072 , a : List[Any]="gelu" , a : Dict=0.0 , a : int=0.0 , a : str=0.02 , a : List[str]=1e-06 , a : List[Any]=224 , a : Union[str, Any]=16 , a : Union[str, Any]=3 , a : Tuple=True , **a : Dict , ) -> List[Any]: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Dict = hidden_size SCREAMING_SNAKE_CASE : Optional[Any] = num_hidden_layers SCREAMING_SNAKE_CASE : Optional[Any] = num_attention_heads SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : Any = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = initializer_range SCREAMING_SNAKE_CASE : int = layer_norm_eps SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : Tuple = patch_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : List[str] = qkv_bias
76
0
import gc import random import tempfile import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel from diffusers.pipelines.stable_diffusion_safe import StableDiffusionPipelineSafe as StableDiffusionPipeline from diffusers.utils import floats_tensor, nightly, torch_device from diffusers.utils.testing_utils import require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def __lowercase ( self ): """simple docstring""" __UpperCamelCase : int =1 __UpperCamelCase : Any =3 __UpperCamelCase : Union[str, Any] =(32, 32) __UpperCamelCase : Any =floats_tensor((batch_size, num_channels) + sizes , rng=random.Random(0 ) ).to(lowerCamelCase__ ) return image @property def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Tuple =UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=('DownBlock2D', 'CrossAttnDownBlock2D') , up_block_types=('CrossAttnUpBlock2D', 'UpBlock2D') , cross_attention_dim=32 , ) return model @property def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Optional[Any] =AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=['DownEncoderBlock2D', 'DownEncoderBlock2D'] , up_block_types=['UpDecoderBlock2D', 'UpDecoderBlock2D'] , latent_channels=4 , ) return model @property def __lowercase ( self ): """simple docstring""" torch.manual_seed(0 ) __UpperCamelCase : Tuple =CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , ) return CLIPTextModel(lowerCamelCase__ ) @property def __lowercase ( self ): """simple docstring""" def extract(*lowerCamelCase__ , **lowerCamelCase__ ): class __A : """simple docstring""" def __init__( self ): """simple docstring""" __UpperCamelCase : str =torch.ones([0] ) def __lowercase ( self , lowerCamelCase__ ): """simple docstring""" self.pixel_values.to(lowerCamelCase__ ) return self return Out() return extract def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict ="cpu" # ensure determinism for the device-dependent torch.Generator __UpperCamelCase : List[Any] =self.dummy_cond_unet __UpperCamelCase : List[Any] =DDIMScheduler( beta_start=0.00_085 , beta_end=0.012 , beta_schedule='scaled_linear' , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) __UpperCamelCase : Any =self.dummy_vae __UpperCamelCase : Optional[int] =self.dummy_text_encoder __UpperCamelCase : Dict =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # make sure here that pndm scheduler skips prk __UpperCamelCase : Dict =StableDiffusionPipeline( unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , safety_checker=lowerCamelCase__ , feature_extractor=self.dummy_extractor , ) __UpperCamelCase : Union[str, Any] =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : str ="A painting of a squirrel eating a burger" __UpperCamelCase : Optional[Any] =torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __UpperCamelCase : Any =sd_pipe([prompt] , generator=lowerCamelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' ) __UpperCamelCase : Optional[Any] =output.images __UpperCamelCase : List[Any] =torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __UpperCamelCase : str =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=lowerCamelCase__ , )[0] __UpperCamelCase : int =image[0, -3:, -3:, -1] __UpperCamelCase : int =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __UpperCamelCase : List[Any] =np.array([0.5_756, 0.6_118, 0.5_005, 0.5_041, 0.5_471, 0.4_726, 0.4_976, 0.4_865, 0.4_864] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] ="cpu" # ensure determinism for the device-dependent torch.Generator __UpperCamelCase : Optional[Any] =self.dummy_cond_unet __UpperCamelCase : Tuple =PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.dummy_vae __UpperCamelCase : Optional[Any] =self.dummy_text_encoder __UpperCamelCase : int =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # make sure here that pndm scheduler skips prk __UpperCamelCase : Any =StableDiffusionPipeline( unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , safety_checker=lowerCamelCase__ , feature_extractor=self.dummy_extractor , ) __UpperCamelCase : Tuple =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] ="A painting of a squirrel eating a burger" __UpperCamelCase : str =torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __UpperCamelCase : Optional[Any] =sd_pipe([prompt] , generator=lowerCamelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' ) __UpperCamelCase : Union[str, Any] =output.images __UpperCamelCase : Any =torch.Generator(device=lowerCamelCase__ ).manual_seed(0 ) __UpperCamelCase : int =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=6.0 , num_inference_steps=2 , output_type='np' , return_dict=lowerCamelCase__ , )[0] __UpperCamelCase : Union[str, Any] =image[0, -3:, -3:, -1] __UpperCamelCase : List[str] =image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) __UpperCamelCase : Optional[Any] =np.array([0.5_125, 0.5_716, 0.4_828, 0.5_060, 0.5_650, 0.4_768, 0.5_185, 0.4_895, 0.4_993] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =StableDiffusionPipeline.from_pretrained( 'hf-internal-testing/tiny-stable-diffusion-lms-pipe' , safety_checker=lowerCamelCase__ ) assert isinstance(lowerCamelCase__ , lowerCamelCase__ ) assert isinstance(pipe.scheduler , lowerCamelCase__ ) assert pipe.safety_checker is None __UpperCamelCase : List[Any] =pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None # check that there's no error when saving a pipeline with one of the models being None with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCamelCase__ ) __UpperCamelCase : Tuple =StableDiffusionPipeline.from_pretrained(lowerCamelCase__ ) # sanity check that the pipeline still works assert pipe.safety_checker is None __UpperCamelCase : Optional[Any] =pipe('example prompt' , num_inference_steps=2 ).images[0] assert image is not None @unittest.skipIf(torch_device != 'cuda' , 'This test requires a GPU' ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.dummy_cond_unet __UpperCamelCase : List[str] =PNDMScheduler(skip_prk_steps=lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =self.dummy_vae __UpperCamelCase : List[str] =self.dummy_text_encoder __UpperCamelCase : List[str] =CLIPTokenizer.from_pretrained('hf-internal-testing/tiny-random-clip' ) # put models in fp16 __UpperCamelCase : Union[str, Any] =unet.half() __UpperCamelCase : str =vae.half() __UpperCamelCase : Dict =bert.half() # make sure here that pndm scheduler skips prk __UpperCamelCase : Optional[Any] =StableDiffusionPipeline( unet=lowerCamelCase__ , scheduler=lowerCamelCase__ , vae=lowerCamelCase__ , text_encoder=lowerCamelCase__ , tokenizer=lowerCamelCase__ , safety_checker=lowerCamelCase__ , feature_extractor=self.dummy_extractor , ) __UpperCamelCase : Any =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : str ="A painting of a squirrel eating a burger" __UpperCamelCase : Any =sd_pipe([prompt] , num_inference_steps=2 , output_type='np' ).images assert image.shape == (1, 64, 64, 3) @nightly @require_torch_gpu class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=lowerCamelCase__ ) __UpperCamelCase : Any =LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) __UpperCamelCase : int =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =( "portrait of girl with smokey eyes makeup in abandoned hotel, grange clothes, redshift, wide high angle" " coloured polaroid photograph with flash, kodak film, hyper real, stunning moody cinematography, with" " anamorphic lenses, by maripol, fallen angels by wong kar - wai, style of suspiria and neon demon and" " children from bahnhof zoo, detailed " ) __UpperCamelCase : Tuple =4003660346 __UpperCamelCase : Dict =7 # without safety guidance (sld_guidance_scale = 0) __UpperCamelCase : Optional[Any] =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : int =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=0 , ) __UpperCamelCase : Any =output.images __UpperCamelCase : str =image[0, -3:, -3:, -1] __UpperCamelCase : List[str] =[0.2_278, 0.2_231, 0.2_249, 0.2_333, 0.2_303, 0.1_885, 0.2_273, 0.2_144, 0.2_176] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 # without safety guidance (strong configuration) __UpperCamelCase : int =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase : str =output.images __UpperCamelCase : Dict =image[0, -3:, -3:, -1] __UpperCamelCase : Any =[0.2_383, 0.2_276, 0.236, 0.2_192, 0.2_186, 0.2_053, 0.1_971, 0.1_901, 0.1_719] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' , safety_checker=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =LMSDiscreteScheduler.from_config(sd_pipe.scheduler.config ) __UpperCamelCase : Tuple =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : str ="padme amidala taking a bath artwork, safe for work, no nudity" __UpperCamelCase : List[Any] =2734971755 __UpperCamelCase : List[Any] =7 __UpperCamelCase : List[Any] =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : Dict =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=0 , ) __UpperCamelCase : Dict =output.images __UpperCamelCase : List[str] =image[0, -3:, -3:, -1] __UpperCamelCase : List[str] =[0.3_502, 0.3_622, 0.3_396, 0.3_642, 0.3_478, 0.3_318, 0.35, 0.3_348, 0.3_297] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 __UpperCamelCase : int =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : str =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase : Tuple =output.images __UpperCamelCase : Any =image[0, -3:, -3:, -1] __UpperCamelCase : Tuple =[0.5_531, 0.5_206, 0.4_895, 0.5_156, 0.5_182, 0.4_751, 0.4_802, 0.4_803, 0.4_443] assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =StableDiffusionPipeline.from_pretrained('runwayml/stable-diffusion-v1-5' ) __UpperCamelCase : Tuple =sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) __UpperCamelCase : int =( "the four horsewomen of the apocalypse, painting by tom of finland, gaston bussiere, craig mullins, j. c." " leyendecker" ) __UpperCamelCase : int =1044355234 __UpperCamelCase : Tuple =12 __UpperCamelCase : Any =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : Union[str, Any] =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=0 , ) __UpperCamelCase : int =output.images __UpperCamelCase : List[str] =image[0, -3:, -3:, -1] __UpperCamelCase : Dict =np.array([0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0, 0.0] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-7 __UpperCamelCase : int =torch.manual_seed(lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =sd_pipe( [prompt] , generator=lowerCamelCase__ , guidance_scale=lowerCamelCase__ , num_inference_steps=50 , output_type='np' , width=512 , height=512 , sld_guidance_scale=2000 , sld_warmup_steps=7 , sld_threshold=0.025 , sld_momentum_scale=0.5 , sld_mom_beta=0.7 , ) __UpperCamelCase : Tuple =output.images __UpperCamelCase : Tuple =image[0, -3:, -3:, -1] __UpperCamelCase : Union[str, Any] =np.array([0.5_818, 0.6_285, 0.6_835, 0.6_019, 0.625, 0.6_754, 0.6_096, 0.6_334, 0.6_561] ) assert image.shape == (1, 512, 512, 3) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
71
import baseaa def lowerCamelCase__ ( _a): return baseaa.aaaencode(string.encode("utf-8")) def lowerCamelCase__ ( _a): return baseaa.aaadecode(_a).decode("utf-8") if __name__ == "__main__": import doctest doctest.testmod()
76
0
"""simple docstring""" import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all feature extractors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...feature_extraction_utils import FeatureExtractionMixin from ...utils import CONFIG_NAME, FEATURE_EXTRACTOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) lowercase_ = logging.get_logger(__name__) lowercase_ = OrderedDict( [ ("audio-spectrogram-transformer", "ASTFeatureExtractor"), ("beit", "BeitFeatureExtractor"), ("chinese_clip", "ChineseCLIPFeatureExtractor"), ("clap", "ClapFeatureExtractor"), ("clip", "CLIPFeatureExtractor"), ("clipseg", "ViTFeatureExtractor"), ("conditional_detr", "ConditionalDetrFeatureExtractor"), ("convnext", "ConvNextFeatureExtractor"), ("cvt", "ConvNextFeatureExtractor"), ("data2vec-audio", "Wav2Vec2FeatureExtractor"), ("data2vec-vision", "BeitFeatureExtractor"), ("deformable_detr", "DeformableDetrFeatureExtractor"), ("deit", "DeiTFeatureExtractor"), ("detr", "DetrFeatureExtractor"), ("dinat", "ViTFeatureExtractor"), ("donut-swin", "DonutFeatureExtractor"), ("dpt", "DPTFeatureExtractor"), ("encodec", "EncodecFeatureExtractor"), ("flava", "FlavaFeatureExtractor"), ("glpn", "GLPNFeatureExtractor"), ("groupvit", "CLIPFeatureExtractor"), ("hubert", "Wav2Vec2FeatureExtractor"), ("imagegpt", "ImageGPTFeatureExtractor"), ("layoutlmv2", "LayoutLMv2FeatureExtractor"), ("layoutlmv3", "LayoutLMv3FeatureExtractor"), ("levit", "LevitFeatureExtractor"), ("maskformer", "MaskFormerFeatureExtractor"), ("mctct", "MCTCTFeatureExtractor"), ("mobilenet_v1", "MobileNetV1FeatureExtractor"), ("mobilenet_v2", "MobileNetV2FeatureExtractor"), ("mobilevit", "MobileViTFeatureExtractor"), ("nat", "ViTFeatureExtractor"), ("owlvit", "OwlViTFeatureExtractor"), ("perceiver", "PerceiverFeatureExtractor"), ("poolformer", "PoolFormerFeatureExtractor"), ("regnet", "ConvNextFeatureExtractor"), ("resnet", "ConvNextFeatureExtractor"), ("segformer", "SegformerFeatureExtractor"), ("sew", "Wav2Vec2FeatureExtractor"), ("sew-d", "Wav2Vec2FeatureExtractor"), ("speech_to_text", "Speech2TextFeatureExtractor"), ("speecht5", "SpeechT5FeatureExtractor"), ("swiftformer", "ViTFeatureExtractor"), ("swin", "ViTFeatureExtractor"), ("swinv2", "ViTFeatureExtractor"), ("table-transformer", "DetrFeatureExtractor"), ("timesformer", "VideoMAEFeatureExtractor"), ("tvlt", "TvltFeatureExtractor"), ("unispeech", "Wav2Vec2FeatureExtractor"), ("unispeech-sat", "Wav2Vec2FeatureExtractor"), ("van", "ConvNextFeatureExtractor"), ("videomae", "VideoMAEFeatureExtractor"), ("vilt", "ViltFeatureExtractor"), ("vit", "ViTFeatureExtractor"), ("vit_mae", "ViTFeatureExtractor"), ("vit_msn", "ViTFeatureExtractor"), ("wav2vec2", "Wav2Vec2FeatureExtractor"), ("wav2vec2-conformer", "Wav2Vec2FeatureExtractor"), ("wavlm", "Wav2Vec2FeatureExtractor"), ("whisper", "WhisperFeatureExtractor"), ("xclip", "CLIPFeatureExtractor"), ("yolos", "YolosFeatureExtractor"), ] ) lowercase_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FEATURE_EXTRACTOR_MAPPING_NAMES) def lowercase ( lowerCAmelCase__ : Dict ) -> Union[str, Any]: for module_name, extractors in FEATURE_EXTRACTOR_MAPPING_NAMES.items(): if class_name in extractors: __a = model_type_to_module_name(_a ) __a = importlib.import_module(f'''.{module_name}''' , '''transformers.models''' ) try: return getattr(_a , _a ) except AttributeError: continue for _, extractor in FEATURE_EXTRACTOR_MAPPING._extra_content.items(): if getattr(_a , '''__name__''' , _a ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. __a = importlib.import_module('''transformers''' ) if hasattr(_a , _a ): return getattr(_a , _a ) return None def lowercase ( lowerCAmelCase__ : Any , lowerCAmelCase__ : str = None , lowerCAmelCase__ : Optional[Any] = False , lowerCAmelCase__ : Tuple = False , lowerCAmelCase__ : Dict = None , lowerCAmelCase__ : Any = None , lowerCAmelCase__ : List[Any] = None , lowerCAmelCase__ : Any = False , **lowerCAmelCase__ : Union[str, Any] , ) -> str: __a = get_file_from_repo( _a , _a , cache_dir=_a , force_download=_a , resume_download=_a , proxies=_a , use_auth_token=_a , revision=_a , local_files_only=_a , ) if resolved_config_file is None: logger.info( '''Could not locate the feature extractor configuration file, will try to use the model config instead.''' ) return {} with open(_a , encoding='''utf-8''' ) as reader: return json.load(_a ) class __lowerCAmelCase : '''simple docstring''' def __init__( self ): raise EnvironmentError( '''AutoFeatureExtractor is designed to be instantiated ''' '''using the `AutoFeatureExtractor.from_pretrained(pretrained_model_name_or_path)` method.''' ) @classmethod @replace_list_option_in_docstrings(_a ) def __UpperCAmelCase ( cls , _a , **_a ): __a = kwargs.pop('''config''' , _a ) __a = kwargs.pop('''trust_remote_code''' , _a ) __a = True __a = FeatureExtractionMixin.get_feature_extractor_dict(_a , **_a ) __a = config_dict.get('''feature_extractor_type''' , _a ) __a = None if "AutoFeatureExtractor" in config_dict.get('''auto_map''' , {} ): __a = config_dict["auto_map"]["AutoFeatureExtractor"] # If we don't find the feature extractor class in the feature extractor config, let's try the model config. if feature_extractor_class is None and feature_extractor_auto_map is None: if not isinstance(_a , _a ): __a = AutoConfig.from_pretrained(_a , **_a ) # It could be in `config.feature_extractor_type`` __a = getattr(_a , '''feature_extractor_type''' , _a ) if hasattr(_a , '''auto_map''' ) and "AutoFeatureExtractor" in config.auto_map: __a = config.auto_map["AutoFeatureExtractor"] if feature_extractor_class is not None: __a = feature_extractor_class_from_name(_a ) __a = feature_extractor_auto_map is not None __a = feature_extractor_class is not None or type(_a ) in FEATURE_EXTRACTOR_MAPPING __a = resolve_trust_remote_code( _a , _a , _a , _a ) if has_remote_code and trust_remote_code: __a = get_class_from_dynamic_module( _a , _a , **_a ) __a = kwargs.pop('''code_revision''' , _a ) if os.path.isdir(_a ): feature_extractor_class.register_for_auto_class() return feature_extractor_class.from_dict(_a , **_a ) elif feature_extractor_class is not None: return feature_extractor_class.from_dict(_a , **_a ) # Last try: we use the FEATURE_EXTRACTOR_MAPPING. elif type(_a ) in FEATURE_EXTRACTOR_MAPPING: __a = FEATURE_EXTRACTOR_MAPPING[type(_a )] return feature_extractor_class.from_dict(_a , **_a ) raise ValueError( f'''Unrecognized feature extractor in {pretrained_model_name_or_path}. Should have a ''' f'''`feature_extractor_type` key in its {FEATURE_EXTRACTOR_NAME} of {CONFIG_NAME}, or one of the following ''' f'''`model_type` keys in its {CONFIG_NAME}: {', '.join(c for c in FEATURE_EXTRACTOR_MAPPING_NAMES.keys() )}''' ) @staticmethod def __UpperCAmelCase ( _a , _a ): FEATURE_EXTRACTOR_MAPPING.register(_a , _a )
45
from datetime import datetime as dt import os from github import Github a_ = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : int = Github(os.environ["GITHUB_TOKEN"]) SCREAMING_SNAKE_CASE : List[str] = g.get_repo("huggingface/transformers") SCREAMING_SNAKE_CASE : Optional[int] = repo.get_issues(state="open") for issue in open_issues: SCREAMING_SNAKE_CASE : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda _a: i.created_at , reverse=_a) SCREAMING_SNAKE_CASE : str = comments[0] if len(_a) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="closed") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") if __name__ == "__main__": main()
76
0
from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = Dict[str, Any] UpperCamelCase = List[Prediction] @add_end_docstrings(__A ) class __UpperCAmelCase (__A ): def __init__( self: List[str] , *UpperCAmelCase_: Optional[Any] , **UpperCAmelCase_: Tuple ): '''simple docstring''' super().__init__(*UpperCAmelCase_ , **UpperCAmelCase_ ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def UpperCamelCase ( self: Any , **UpperCAmelCase_: Optional[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} if "threshold" in kwargs: _SCREAMING_SNAKE_CASE = kwargs["threshold"] return {}, {}, postprocess_kwargs def __call__( self: List[Any] , *UpperCAmelCase_: List[Any] , **UpperCAmelCase_: Optional[int] ): '''simple docstring''' return super().__call__(*UpperCAmelCase_ , **UpperCAmelCase_ ) def UpperCamelCase ( self: Dict , UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = load_image(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = torch.IntTensor([[image.height, image.width]] ) _SCREAMING_SNAKE_CASE = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: _SCREAMING_SNAKE_CASE = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) _SCREAMING_SNAKE_CASE = target_size return inputs def UpperCamelCase ( self: str , UpperCAmelCase_: Union[str, Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_inputs.pop("""target_size""" ) _SCREAMING_SNAKE_CASE = self.model(**UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: _SCREAMING_SNAKE_CASE = model_inputs["bbox"] return model_outputs def UpperCamelCase ( self: str , UpperCAmelCase_: Optional[Any] , UpperCAmelCase_: Union[str, Any]=0.9 ): '''simple docstring''' _SCREAMING_SNAKE_CASE = model_outputs["target_size"] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. _SCREAMING_SNAKE_CASE = target_size[0].tolist() def unnormalize(UpperCAmelCase_: Optional[Any] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1_000), (height * bbox[1] / 1_000), (width * bbox[2] / 1_000), (height * bbox[3] / 1_000), ] ) ) _SCREAMING_SNAKE_CASE = model_outputs["logits"].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) _SCREAMING_SNAKE_CASE = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] _SCREAMING_SNAKE_CASE = [unnormalize(UpperCAmelCase_ ) for bbox in model_outputs["bbox"].squeeze(0 )] _SCREAMING_SNAKE_CASE = ["score", "label", "box"] _SCREAMING_SNAKE_CASE = [dict(zip(UpperCAmelCase_ , UpperCAmelCase_ ) ) for vals in zip(scores.tolist() , UpperCAmelCase_ , UpperCAmelCase_ ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel _SCREAMING_SNAKE_CASE = self.image_processor.post_process_object_detection(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = raw_annotations[0] _SCREAMING_SNAKE_CASE = raw_annotation["scores"] _SCREAMING_SNAKE_CASE = raw_annotation["labels"] _SCREAMING_SNAKE_CASE = raw_annotation["boxes"] _SCREAMING_SNAKE_CASE = scores.tolist() _SCREAMING_SNAKE_CASE = [self.model.config.idalabel[label.item()] for label in labels] _SCREAMING_SNAKE_CASE = [self._get_bounding_box(UpperCAmelCase_ ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] _SCREAMING_SNAKE_CASE = ["score", "label", "box"] _SCREAMING_SNAKE_CASE = [ dict(zip(UpperCAmelCase_ , UpperCAmelCase_ ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def UpperCamelCase ( self: Union[str, Any] , UpperCAmelCase_: "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) _SCREAMING_SNAKE_CASE = box.int().tolist() _SCREAMING_SNAKE_CASE = { "xmin": xmin, "ymin": ymin, "xmax": xmax, "ymax": ymax, } return bbox
306
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): if isinstance(_a , (list, tuple)) and isinstance(videos[0] , (list, tuple)) and is_valid_image(videos[0][0]): return videos elif isinstance(_a , (list, tuple)) and is_valid_image(videos[0]): return [videos] elif is_valid_image(_a): return [[videos]] raise ValueError(f"Could not make batched video from {videos}") class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =['pixel_values'] def __init__( self : Optional[Any] , a : bool = True , a : Dict[str, int] = None , a : PILImageResampling = PILImageResampling.BILINEAR , a : bool = True , a : Dict[str, int] = None , a : bool = True , a : Union[int, float] = 1 / 255 , a : bool = True , a : bool = True , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , **a : Tuple , ) -> None: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Tuple = size if size is not None else {"shortest_edge": 256} SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE : str = get_size_dict(a , param_name="crop_size" ) SCREAMING_SNAKE_CASE : Dict = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop SCREAMING_SNAKE_CASE : int = crop_size SCREAMING_SNAKE_CASE : int = resample SCREAMING_SNAKE_CASE : Any = do_rescale SCREAMING_SNAKE_CASE : int = rescale_factor SCREAMING_SNAKE_CASE : Tuple = offset SCREAMING_SNAKE_CASE : str = do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __UpperCamelCase ( self : Optional[Any] , a : np.ndarray , a : Dict[str, int] , a : PILImageResampling = PILImageResampling.BILINEAR , a : Optional[Union[str, ChannelDimension]] = None , **a : Union[str, Any] , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) if "shortest_edge" in size: SCREAMING_SNAKE_CASE : str = get_resize_output_image_size(a , size["shortest_edge"] , default_to_square=a ) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE : Dict = (size["height"], size["width"]) else: raise ValueError(F"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(a , size=a , resample=a , data_format=a , **a ) def __UpperCamelCase ( self : List[str] , a : np.ndarray , a : Dict[str, int] , a : Optional[Union[str, ChannelDimension]] = None , **a : str , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : str = get_size_dict(a ) if "height" not in size or "width" not in size: raise ValueError(F"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(a , size=(size["height"], size["width"]) , data_format=a , **a ) def __UpperCamelCase ( self : List[Any] , a : np.ndarray , a : Union[int, float] , a : bool = True , a : Optional[Union[str, ChannelDimension]] = None , **a : Tuple , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : int = image.astype(np.floataa ) if offset: SCREAMING_SNAKE_CASE : Union[str, Any] = image - (scale / 2) return rescale(a , scale=a , data_format=a , **a ) def __UpperCamelCase ( self : int , a : np.ndarray , a : Union[float, List[float]] , a : Union[float, List[float]] , a : Optional[Union[str, ChannelDimension]] = None , **a : List[str] , ) -> np.ndarray: """simple docstring""" return normalize(a , mean=a , std=a , data_format=a , **a ) def __UpperCamelCase ( self : Tuple , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: """simple docstring""" if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) if offset and not do_rescale: raise ValueError("For offset, do_rescale must also be set to True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : List[str] = to_numpy_array(a ) if do_resize: SCREAMING_SNAKE_CASE : Optional[Any] = self.resize(image=a , size=a , resample=a ) if do_center_crop: SCREAMING_SNAKE_CASE : Union[str, Any] = self.center_crop(a , size=a ) if do_rescale: SCREAMING_SNAKE_CASE : Any = self.rescale(image=a , scale=a , offset=a ) if do_normalize: SCREAMING_SNAKE_CASE : Tuple = self.normalize(image=a , mean=a , std=a ) SCREAMING_SNAKE_CASE : Optional[int] = to_channel_dimension_format(a , a ) return image def __UpperCamelCase ( self : Dict , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[str, TensorType]] = None , a : ChannelDimension = ChannelDimension.FIRST , **a : Tuple , ) -> PIL.Image.Image: """simple docstring""" SCREAMING_SNAKE_CASE : str = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Union[str, Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : int = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : str = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : Optional[Any] = offset if offset is not None else self.offset SCREAMING_SNAKE_CASE : str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : int = size if size is not None else self.size SCREAMING_SNAKE_CASE : List[Any] = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = get_size_dict(a , param_name="crop_size" ) if not valid_images(a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) SCREAMING_SNAKE_CASE : Optional[int] = make_batched(a ) SCREAMING_SNAKE_CASE : List[Any] = [ [ self._preprocess_image( image=a , do_resize=a , size=a , resample=a , do_center_crop=a , crop_size=a , do_rescale=a , rescale_factor=a , offset=a , do_normalize=a , image_mean=a , image_std=a , data_format=a , ) for img in video ] for video in videos ] SCREAMING_SNAKE_CASE : Optional[int] = {"pixel_values": videos} return BatchFeature(data=a , tensor_type=a )
76
0
from torch import nn def UpperCAmelCase_ ( __lowerCAmelCase ) -> Union[str, Any]: if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(F'Unsupported activation function: {act_fn}' )
156
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a_ = logging.get_logger(__name__) a_ = {'vocab_file': 'vocab.txt'} a_ = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a_ = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } a_ = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =ConvBertTokenizer def __init__( self : List[str] , a : Union[str, Any]=None , a : Optional[int]=None , a : int=True , a : Tuple="[UNK]" , a : Dict="[SEP]" , a : Dict="[PAD]" , a : List[Any]="[CLS]" , a : Tuple="[MASK]" , a : Dict=True , a : Optional[Any]=None , **a : str , ) -> Dict: """simple docstring""" super().__init__( a , tokenizer_file=a , do_lower_case=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , tokenize_chinese_chars=a , strip_accents=a , **a , ) SCREAMING_SNAKE_CASE : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , a ) != do_lower_case or normalizer_state.get("strip_accents" , a ) != strip_accents or normalizer_state.get("handle_chinese_chars" , a ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE : List[str] = getattr(a , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE : Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE : Any = strip_accents SCREAMING_SNAKE_CASE : Optional[int] = tokenize_chinese_chars SCREAMING_SNAKE_CASE : List[str] = normalizer_class(**a ) SCREAMING_SNAKE_CASE : str = do_lower_case def __UpperCamelCase ( self : Union[str, Any] , a : List[Any] , a : int=None ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : Tuple , a : str , a : Optional[str] = None ) -> Tuple[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self._tokenizer.model.save(a , name=a ) return tuple(a )
76
0
from collections import OrderedDict from typing import TYPE_CHECKING, Any, Mapping, Optional from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...utils import logging if TYPE_CHECKING: from ...processing_utils import ProcessorMixin from ...utils import TensorType lowercase_ = logging.get_logger(__name__) lowercase_ = { "microsoft/layoutlmv3-base": "https://huggingface.co/microsoft/layoutlmv3-base/resolve/main/config.json", } class A ( __A ): """simple docstring""" lowerCamelCase = 'layoutlmv3' def __init__( self : List[str],lowercase_ : Union[str, Any]=5_0_2_6_5,lowercase_ : Optional[int]=7_6_8,lowercase_ : Optional[Any]=1_2,lowercase_ : Any=1_2,lowercase_ : int=3_0_7_2,lowercase_ : Any="gelu",lowercase_ : List[Any]=0.1,lowercase_ : Dict=0.1,lowercase_ : Optional[int]=5_1_2,lowercase_ : Optional[Any]=2,lowercase_ : Union[str, Any]=0.02,lowercase_ : Optional[int]=1E-5,lowercase_ : Tuple=1,lowercase_ : Dict=0,lowercase_ : List[Any]=2,lowercase_ : Optional[Any]=1_0_2_4,lowercase_ : Optional[int]=1_2_8,lowercase_ : Any=1_2_8,lowercase_ : Union[str, Any]=True,lowercase_ : int=3_2,lowercase_ : int=1_2_8,lowercase_ : Dict=6_4,lowercase_ : List[str]=2_5_6,lowercase_ : Tuple=True,lowercase_ : int=True,lowercase_ : str=True,lowercase_ : Optional[int]=2_2_4,lowercase_ : int=3,lowercase_ : Tuple=1_6,lowercase_ : Dict=None,**lowercase_ : str,)-> Dict: '''simple docstring''' super().__init__( vocab_size=lowercase_,hidden_size=lowercase_,num_hidden_layers=lowercase_,num_attention_heads=lowercase_,intermediate_size=lowercase_,hidden_act=lowercase_,hidden_dropout_prob=lowercase_,attention_probs_dropout_prob=lowercase_,max_position_embeddings=lowercase_,type_vocab_size=lowercase_,initializer_range=lowercase_,layer_norm_eps=lowercase_,pad_token_id=lowercase_,bos_token_id=lowercase_,eos_token_id=lowercase_,**lowercase_,) A__ = max_ad_position_embeddings A__ = coordinate_size A__ = shape_size A__ = has_relative_attention_bias A__ = rel_pos_bins A__ = max_rel_pos A__ = has_spatial_attention_bias A__ = rel_ad_pos_bins A__ = max_rel_ad_pos A__ = text_embed A__ = visual_embed A__ = input_size A__ = num_channels A__ = patch_size A__ = classifier_dropout class A ( __A ): """simple docstring""" lowerCamelCase = version.parse('1.12' ) @property def snake_case__ ( self : str )-> Mapping[str, Mapping[int, str]]: '''simple docstring''' if self.task in ["question-answering", "sequence-classification"]: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels', 2: 'height', 3: 'width'}), ] ) else: return OrderedDict( [ ('input_ids', {0: 'batch', 1: 'sequence'}), ('bbox', {0: 'batch', 1: 'sequence'}), ('attention_mask', {0: 'batch', 1: 'sequence'}), ('pixel_values', {0: 'batch', 1: 'num_channels'}), ] ) @property def snake_case__ ( self : Tuple )-> float: '''simple docstring''' return 1E-5 @property def snake_case__ ( self : List[Any] )-> int: '''simple docstring''' return 1_2 def snake_case__ ( self : List[str],lowercase_ : "ProcessorMixin",lowercase_ : int = -1,lowercase_ : int = -1,lowercase_ : bool = False,lowercase_ : Optional["TensorType"] = None,lowercase_ : int = 3,lowercase_ : int = 4_0,lowercase_ : int = 4_0,)-> Mapping[str, Any]: '''simple docstring''' setattr(processor.image_processor,'apply_ocr',lowercase_ ) # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX A__ = compute_effective_axis_dimension( lowercase_,fixed_dimension=OnnxConfig.default_fixed_batch,num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX A__ = processor.tokenizer.num_special_tokens_to_add(lowercase_ ) A__ = compute_effective_axis_dimension( lowercase_,fixed_dimension=OnnxConfig.default_fixed_sequence,num_token_to_add=lowercase_ ) # Generate dummy inputs according to compute batch and sequence A__ = [[" ".join([processor.tokenizer.unk_token] ) * seq_length]] * batch_size # Generate dummy bounding boxes A__ = [[[4_8, 8_4, 7_3, 1_2_8]]] * batch_size # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX # batch_size = compute_effective_axis_dimension(batch_size, fixed_dimension=OnnxConfig.default_fixed_batch) A__ = self._generate_dummy_images(lowercase_,lowercase_,lowercase_,lowercase_ ) A__ = dict( processor( lowercase_,text=lowercase_,boxes=lowercase_,return_tensors=lowercase_,) ) return inputs
7
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. a_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_a) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE : Union[str, Any] = terminalreporter.config.getoption("--make-reports") if make_reports: pytest_terminal_summary_main(_a , id=_a)
76
0
from __future__ import annotations def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[str]: '''simple docstring''' print(F"""Vertex\tShortest Distance from vertex {src}""" ) for i, d in enumerate(_a ): print(F"""{i}\t\t{d}""" ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' for j in range(_a ): UpperCAmelCase = (graph[j][k] for k in ["src", "dst", "weight"]) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: return True return False def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' UpperCAmelCase = [float('''inf''' )] * vertex_count UpperCAmelCase = 0.0 for _ in range(vertex_count - 1 ): for j in range(_a ): UpperCAmelCase = (graph[j][k] for k in ["src", "dst", "weight"]) if distance[u] != float('''inf''' ) and distance[u] + w < distance[v]: UpperCAmelCase = distance[u] + w UpperCAmelCase = check_negative_cycle(_a , _a , _a ) if negative_cycle_exists: raise Exception('''Negative cycle found''' ) return distance if __name__ == "__main__": import doctest doctest.testmod() __A : Optional[int] = int(input("Enter number of vertices: ").strip()) __A : Any = int(input("Enter number of edges: ").strip()) __A : str = [{} for _ in range(E)] for i in range(E): print("Edge ", i + 1) __A , __A , __A : Any = ( int(x) for x in input("Enter source, destination, weight: ").strip().split(" ") ) __A : Any = {"src": src, "dst": dest, "weight": weight} __A : List[str] = int(input("\nEnter shortest path source:").strip()) __A : Tuple = bellman_ford(graph, V, E, source) print_distance(shortest_distance, 0)
273
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , a : int , a : Optional[int]=13 , a : Optional[int]=3 , a : int=224 , a : Optional[int]=30 , a : int=400 , a : Union[str, Any]=True , a : int=None , a : Tuple=True , a : Tuple=[0.5, 0.5, 0.5] , a : Optional[int]=[0.5, 0.5, 0.5] , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = size if size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE : Union[str, Any] = parent SCREAMING_SNAKE_CASE : int = batch_size SCREAMING_SNAKE_CASE : int = num_channels SCREAMING_SNAKE_CASE : Any = image_size SCREAMING_SNAKE_CASE : Tuple = min_resolution SCREAMING_SNAKE_CASE : str = max_resolution SCREAMING_SNAKE_CASE : int = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : int = do_normalize SCREAMING_SNAKE_CASE : Tuple = image_mean SCREAMING_SNAKE_CASE : Tuple = image_std def __UpperCamelCase ( self : Any ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCamelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =ViTImageProcessor if is_vision_available() else None def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = EfficientFormerImageProcessorTester(self ) @property def __UpperCamelCase ( self : Any ) -> List[str]: """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def __UpperCamelCase ( self : List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , "image_mean" ) ) self.assertTrue(hasattr(a , "image_std" ) ) self.assertTrue(hasattr(a , "do_normalize" ) ) self.assertTrue(hasattr(a , "do_resize" ) ) self.assertTrue(hasattr(a , "size" ) ) def __UpperCamelCase ( self : int ) -> str: """simple docstring""" pass def __UpperCamelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE : List[str] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE : int = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Any = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
76
0
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import re from ..models.auto import AutoProcessor from ..models.vision_encoder_decoder import VisionEncoderDecoderModel from ..utils import is_vision_available from .base import PipelineTool if is_vision_available(): from PIL import Image class snake_case__ ( __A ): lowercase__ : str = '''naver-clova-ix/donut-base-finetuned-docvqa''' lowercase__ : Dict = ( '''This is a tool that answers a question about an document (pdf). It takes an input named `document` which ''' '''should be the document containing the information, as well as a `question` that is the question about the ''' '''document. It returns a text that contains the answer to the question.''' ) lowercase__ : Dict = '''document_qa''' lowercase__ : Dict = AutoProcessor lowercase__ : Any = VisionEncoderDecoderModel lowercase__ : Any = ['''image''', '''text'''] lowercase__ : List[Any] = ['''text'''] def __init__( self , *lowerCAmelCase__ , **lowerCAmelCase__ ) -> Optional[int]: if not is_vision_available(): raise ValueError("""Pillow must be installed to use the DocumentQuestionAnsweringTool.""" ) super().__init__(*lowerCAmelCase__ , **lowerCAmelCase__ ) def __magic_name__ ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> Dict: __magic_name__ : List[str] = "<s_docvqa><s_question>{user_input}</s_question><s_answer>" __magic_name__ : Any = task_prompt.replace("""{user_input}""" , lowerCAmelCase__ ) __magic_name__ : List[Any] = self.pre_processor.tokenizer( lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ , return_tensors="""pt""" ).input_ids __magic_name__ : Optional[Any] = self.pre_processor(lowerCAmelCase__ , return_tensors="""pt""" ).pixel_values return {"decoder_input_ids": decoder_input_ids, "pixel_values": pixel_values} def __magic_name__ ( self , lowerCAmelCase__ ) -> int: return self.model.generate( inputs["""pixel_values"""].to(self.device ) , decoder_input_ids=inputs["""decoder_input_ids"""].to(self.device ) , max_length=self.model.decoder.config.max_position_embeddings , early_stopping=lowerCAmelCase__ , pad_token_id=self.pre_processor.tokenizer.pad_token_id , eos_token_id=self.pre_processor.tokenizer.eos_token_id , use_cache=lowerCAmelCase__ , num_beams=1 , bad_words_ids=[[self.pre_processor.tokenizer.unk_token_id]] , return_dict_in_generate=lowerCAmelCase__ , ).sequences def __magic_name__ ( self , lowerCAmelCase__ ) -> int: __magic_name__ : Any = self.pre_processor.batch_decode(lowerCAmelCase__ )[0] __magic_name__ : Dict = sequence.replace(self.pre_processor.tokenizer.eos_token , """""" ) __magic_name__ : int = sequence.replace(self.pre_processor.tokenizer.pad_token , """""" ) __magic_name__ : List[Any] = re.sub(R"""<.*?>""" , """""" , lowerCAmelCase__ , count=1 ).strip() # remove first task start token __magic_name__ : Any = self.pre_processor.tokenajson(lowerCAmelCase__ ) return sequence["answer"]
342
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Any = tokenizer(example["content"] , truncation=_a)["input_ids"] SCREAMING_SNAKE_CASE : Dict = len(example["content"]) / len(output["input_ids"]) return output a_ = HfArgumentParser(PretokenizationArguments) a_ = parser.parse_args() if args.num_workers is None: a_ = multiprocessing.cpu_count() a_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) a_ = time.time() a_ = load_dataset(args.dataset_name, split='train') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a_ = time.time() a_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ 'repo_name', 'path', 'copies', 'size', 'content', 'license', 'hash', 'line_mean', 'line_max', 'alpha_frac', 'autogenerated', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
76
0
'''simple docstring''' from __future__ import annotations import math def _UpperCamelCase ( __A , __A ) -> Optional[Any]: '''simple docstring''' if len(_a ) != 2 or len(a[0] ) != 2 or len(_a ) != 2 or len(b[0] ) != 2: raise Exception("Matrices are not 2x2" ) UpperCamelCase__ = [ [a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]], [a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]], ] return new_matrix def _UpperCamelCase ( __A , __A ) -> Union[str, Any]: '''simple docstring''' return [ [matrix_a[row][col] + matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(_a ) ) ] def _UpperCamelCase ( __A , __A ) -> Optional[Any]: '''simple docstring''' return [ [matrix_a[row][col] - matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(_a ) ) ] def _UpperCamelCase ( __A ) -> List[Any]: '''simple docstring''' if len(_a ) % 2 != 0 or len(a[0] ) % 2 != 0: raise Exception("Odd matrices are not supported!" ) UpperCamelCase__ = len(_a ) UpperCamelCase__ = matrix_length // 2 UpperCamelCase__ = [[a[i][j] for j in range(_a , _a )] for i in range(_a )] UpperCamelCase__ = [ [a[i][j] for j in range(_a , _a )] for i in range(_a , _a ) ] UpperCamelCase__ = [[a[i][j] for j in range(_a )] for i in range(_a )] UpperCamelCase__ = [[a[i][j] for j in range(_a )] for i in range(_a , _a )] return top_left, top_right, bot_left, bot_right def _UpperCamelCase ( __A ) -> int: '''simple docstring''' return len(_a ), len(matrix[0] ) def _UpperCamelCase ( __A ) -> str: '''simple docstring''' print("\n".join(str(_a ) for line in matrix ) ) def _UpperCamelCase ( __A , __A ) -> str: '''simple docstring''' if matrix_dimensions(_a ) == (2, 2): return default_matrix_multiplication(_a , _a ) UpperCamelCase__ = split_matrix(_a ) UpperCamelCase__ = split_matrix(_a ) UpperCamelCase__ = actual_strassen(_a , matrix_subtraction(_a , _a ) ) UpperCamelCase__ = actual_strassen(matrix_addition(_a , _a ) , _a ) UpperCamelCase__ = actual_strassen(matrix_addition(_a , _a ) , _a ) UpperCamelCase__ = actual_strassen(_a , matrix_subtraction(_a , _a ) ) UpperCamelCase__ = actual_strassen(matrix_addition(_a , _a ) , matrix_addition(_a , _a ) ) UpperCamelCase__ = actual_strassen(matrix_subtraction(_a , _a ) , matrix_addition(_a , _a ) ) UpperCamelCase__ = actual_strassen(matrix_subtraction(_a , _a ) , matrix_addition(_a , _a ) ) UpperCamelCase__ = matrix_addition(matrix_subtraction(matrix_addition(_a , _a ) , _a ) , _a ) UpperCamelCase__ = matrix_addition(_a , _a ) UpperCamelCase__ = matrix_addition(_a , _a ) UpperCamelCase__ = matrix_subtraction(matrix_subtraction(matrix_addition(_a , _a ) , _a ) , _a ) # construct the new matrix from our 4 quadrants UpperCamelCase__ = [] for i in range(len(_a ) ): new_matrix.append(top_left[i] + top_right[i] ) for i in range(len(_a ) ): new_matrix.append(bot_left[i] + bot_right[i] ) return new_matrix def _UpperCamelCase ( __A , __A ) -> Optional[Any]: '''simple docstring''' if matrix_dimensions(_a )[1] != matrix_dimensions(_a )[0]: UpperCamelCase__ = ( "Unable to multiply these matrices, please check the dimensions.\n" F'''Matrix A: {matrixa}\n''' F'''Matrix B: {matrixa}''' ) raise Exception(_a ) UpperCamelCase__ = matrix_dimensions(_a ) UpperCamelCase__ = matrix_dimensions(_a ) if dimensiona[0] == dimensiona[1] and dimensiona[0] == dimensiona[1]: return [matrixa, matrixa] UpperCamelCase__ = max(*_a , *_a ) UpperCamelCase__ = int(math.pow(2 , math.ceil(math.loga(_a ) ) ) ) UpperCamelCase__ = matrixa UpperCamelCase__ = matrixa # Adding zeros to the matrices so that the arrays dimensions are the same and also # power of 2 for i in range(0 , _a ): if i < dimensiona[0]: for _ in range(dimensiona[1] , _a ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) if i < dimensiona[0]: for _ in range(dimensiona[1] , _a ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) UpperCamelCase__ = actual_strassen(_a , _a ) # Removing the additional zeros for i in range(0 , _a ): if i < dimensiona[0]: for _ in range(dimensiona[1] , _a ): final_matrix[i].pop() else: final_matrix.pop() return final_matrix if __name__ == "__main__": a__ : List[str] = [ [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 2, 3, 1], ] a__ : Optional[int] = [[0, 2, 1, 1], [1_6, 2, 3, 3], [2, 2, 7, 7], [1_3, 1_1, 2_2, 4]] print(strassen(matrixa, matrixa))
80
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): # initialize config if "resnet-50" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-50") elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-101") else: raise ValueError("Model name should include either resnet50 or resnet101") SCREAMING_SNAKE_CASE : str = DetrConfig(use_timm_backbone=_a , backbone_config=_a) # set label attributes SCREAMING_SNAKE_CASE : List[str] = "panoptic" in model_name if is_panoptic: SCREAMING_SNAKE_CASE : Union[str, Any] = 250 else: SCREAMING_SNAKE_CASE : Union[str, Any] = 91 SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : Union[str, Any] = "coco-detection-id2label.json" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(_a , _a , repo_type="dataset") , "r")) SCREAMING_SNAKE_CASE : int = {int(_a): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCamelCase__ ( _a): # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE : Union[str, Any] = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight")) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight")) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias")) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean")) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var")) # stages for stage_idx in range(len(config.backbone_config.depths)): for layer_idx in range(config.backbone_config.depths[stage_idx]): # shortcut if layer_idx == 0: rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var", )) # 3 convs for i in range(3): rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var", )) # fmt: on for i in range(config.encoder_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", )) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", )) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ]) return rename_keys def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : str = state_dict.pop(_a) SCREAMING_SNAKE_CASE : int = val def lowerCamelCase__ ( _a , _a=False): SCREAMING_SNAKE_CASE : Optional[Any] = "" if is_panoptic: SCREAMING_SNAKE_CASE : Optional[int] = "detr." # first: transformer encoder for i in range(6): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : int = in_proj_bias[:256] SCREAMING_SNAKE_CASE : Tuple = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : str = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : str = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[:256] SCREAMING_SNAKE_CASE : List[Any] = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : Any = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight") SCREAMING_SNAKE_CASE : int = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias") # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE : Tuple = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[-256:] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Union[str, Any] = Image.open(requests.get(_a , stream=_a).raw) return im @torch.no_grad() def lowerCamelCase__ ( _a , _a=None , _a=False): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Optional[int] = get_detr_config(_a) # load original model from torch hub SCREAMING_SNAKE_CASE : Union[str, Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f"Converting model {model_name}...") SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=_a).eval() SCREAMING_SNAKE_CASE : Tuple = detr.state_dict() # rename keys for src, dest in create_rename_keys(_a): if is_panoptic: SCREAMING_SNAKE_CASE : List[str] = "detr." + src rename_key(_a , _a , _a) # query, key and value matrices need special treatment read_in_q_k_v(_a , is_panoptic=_a) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE : List[Any] = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr") and not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor") ): SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Optional[int] = val elif key.startswith("bbox_attention") or key.startswith("mask_head"): continue else: SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : List[Any] = val else: if not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor"): SCREAMING_SNAKE_CASE : Any = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Any = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE : int = DetrForSegmentation(_a) if is_panoptic else DetrForObjectDetection(_a) model.load_state_dict(_a) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE : int = "coco_panoptic" if is_panoptic else "coco_detection" SCREAMING_SNAKE_CASE : Optional[int] = DetrImageProcessor(format=_a) SCREAMING_SNAKE_CASE : List[str] = processor(images=prepare_img() , return_tensors="pt") SCREAMING_SNAKE_CASE : Any = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = detr(_a) SCREAMING_SNAKE_CASE : Any = model(_a) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4) print("Looks ok!") if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}...") Path(_a).mkdir(exist_ok=_a) model.save_pretrained(_a) processor.save_pretrained(_a) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub...") model.push_to_hub(f"nielsr/{model_name}") processor.push_to_hub(f"nielsr/{model_name}") if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') a_ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
76
0
from __future__ import annotations def _A ( SCREAMING_SNAKE_CASE__ : int , SCREAMING_SNAKE_CASE__ : Optional[Any] ): # Checks if the entire collection has been sorted if len(_a ) <= 1 or n <= 1: return insert_next(_a , n - 1 ) rec_insertion_sort(_a , n - 1 ) def _A ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : str ): # Checks order between adjacent elements if index >= len(_a ) or collection[index - 1] <= collection[index]: return # Swaps adjacent elements since they are not in ascending order UpperCamelCase :List[Any] = ( collection[index], collection[index - 1], ) insert_next(_a , index + 1 ) if __name__ == "__main__": __snake_case = input("""Enter integers separated by spaces: """) __snake_case = [int(num) for num in numbers.split()] rec_insertion_sort(number_list, len(number_list)) print(number_list)
259
import os def lowerCamelCase__ ( ): with open(os.path.dirname(_a) + "/p022_names.txt") as file: SCREAMING_SNAKE_CASE : List[str] = str(file.readlines()[0]) SCREAMING_SNAKE_CASE : List[Any] = names.replace("\"" , "").split(",") names.sort() SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Dict = 0 for i, name in enumerate(_a): for letter in name: name_score += ord(_a) - 64 total_score += (i + 1) * name_score SCREAMING_SNAKE_CASE : str = 0 return total_score if __name__ == "__main__": print(solution())
76
0
__A ={ '''Pillow''': '''Pillow''', '''accelerate''': '''accelerate>=0.11.0''', '''compel''': '''compel==0.1.8''', '''black''': '''black~=23.1''', '''datasets''': '''datasets''', '''filelock''': '''filelock''', '''flax''': '''flax>=0.4.1''', '''hf-doc-builder''': '''hf-doc-builder>=0.3.0''', '''huggingface-hub''': '''huggingface-hub>=0.13.2''', '''requests-mock''': '''requests-mock==1.10.0''', '''importlib_metadata''': '''importlib_metadata''', '''invisible-watermark''': '''invisible-watermark''', '''isort''': '''isort>=5.5.4''', '''jax''': '''jax>=0.2.8,!=0.3.2''', '''jaxlib''': '''jaxlib>=0.1.65''', '''Jinja2''': '''Jinja2''', '''k-diffusion''': '''k-diffusion>=0.0.12''', '''torchsde''': '''torchsde''', '''note_seq''': '''note_seq''', '''librosa''': '''librosa''', '''numpy''': '''numpy''', '''omegaconf''': '''omegaconf''', '''parameterized''': '''parameterized''', '''protobuf''': '''protobuf>=3.20.3,<4''', '''pytest''': '''pytest''', '''pytest-timeout''': '''pytest-timeout''', '''pytest-xdist''': '''pytest-xdist''', '''ruff''': '''ruff>=0.0.241''', '''safetensors''': '''safetensors''', '''sentencepiece''': '''sentencepiece>=0.1.91,!=0.1.92''', '''scipy''': '''scipy''', '''onnx''': '''onnx''', '''regex''': '''regex!=2019.12.17''', '''requests''': '''requests''', '''tensorboard''': '''tensorboard''', '''torch''': '''torch>=1.4''', '''torchvision''': '''torchvision''', '''transformers''': '''transformers>=4.25.1''', '''urllib3''': '''urllib3<=2.0.0''', }
19
from collections.abc import Callable import numpy as np def lowerCamelCase__ ( _a , _a , _a , _a , _a): SCREAMING_SNAKE_CASE : Dict = int(np.ceil((x_end - xa) / step_size)) SCREAMING_SNAKE_CASE : Tuple = np.zeros((n + 1,)) SCREAMING_SNAKE_CASE : int = ya SCREAMING_SNAKE_CASE : int = xa for k in range(_a): SCREAMING_SNAKE_CASE : Any = y[k] + step_size * ode_func(_a , y[k]) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
76
0
from datetime import datetime as dt import os from github import Github lowerCAmelCase__ : Any = [ '''good first issue''', '''good second issue''', '''good difficult issue''', '''feature request''', '''new model''', '''wip''', ] def UpperCamelCase__ ( ) -> Optional[int]: snake_case__ : int = Github(os.environ['GITHUB_TOKEN'] ) snake_case__ : List[str] = g.get_repo('huggingface/transformers' ) snake_case__ : Optional[int] = repo.get_issues(state='open' ) for issue in open_issues: snake_case__ : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda A__ : i.created_at , reverse=_a ) snake_case__ : str = comments[0] if len(_a ) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state='closed' ) elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels() ) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( 'This issue has been automatically marked as stale because it has not had ' 'recent activity. If you think this still needs to be addressed ' 'please comment on this thread.\n\nPlease note that issues that do not follow the ' '[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) ' 'are likely to be ignored.' ) if __name__ == "__main__": main()
143
def lowerCamelCase__ ( _a , _a): return int((input_a, input_a).count(1) != 0) def lowerCamelCase__ ( ): assert or_gate(0 , 0) == 0 assert or_gate(0 , 1) == 1 assert or_gate(1 , 0) == 1 assert or_gate(1 , 1) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
76
0
import os import tempfile import unittest from transformers import FlaubertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FlaubertForMultipleChoice, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertModel, FlaubertWithLMHeadModel, ) from transformers.models.flaubert.modeling_flaubert import FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST class __A ( __A ): """simple docstring""" def __init__( self , lowerCamelCase__ , lowerCamelCase__=13 , lowerCamelCase__=7 , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=False , lowerCamelCase__=2 , lowerCamelCase__=99 , lowerCamelCase__=0 , lowerCamelCase__=32 , lowerCamelCase__=5 , lowerCamelCase__=4 , lowerCamelCase__=0.1 , lowerCamelCase__=0.1 , lowerCamelCase__=512 , lowerCamelCase__=12 , lowerCamelCase__=2 , lowerCamelCase__=0.02 , lowerCamelCase__=3 , lowerCamelCase__=4 , lowerCamelCase__="last" , lowerCamelCase__=None , lowerCamelCase__=None , ): """simple docstring""" __UpperCamelCase : Any =parent __UpperCamelCase : Tuple =batch_size __UpperCamelCase : Any =seq_length __UpperCamelCase : int =is_training __UpperCamelCase : Any =use_input_lengths __UpperCamelCase : List[Any] =use_token_type_ids __UpperCamelCase : Union[str, Any] =use_labels __UpperCamelCase : List[str] =gelu_activation __UpperCamelCase : Dict =sinusoidal_embeddings __UpperCamelCase : List[Any] =causal __UpperCamelCase : Optional[int] =asm __UpperCamelCase : Dict =n_langs __UpperCamelCase : Tuple =vocab_size __UpperCamelCase : List[Any] =n_special __UpperCamelCase : Optional[int] =hidden_size __UpperCamelCase : Dict =num_hidden_layers __UpperCamelCase : Any =num_attention_heads __UpperCamelCase : Union[str, Any] =hidden_dropout_prob __UpperCamelCase : List[Any] =attention_probs_dropout_prob __UpperCamelCase : Optional[int] =max_position_embeddings __UpperCamelCase : List[str] =type_vocab_size __UpperCamelCase : Optional[Any] =type_sequence_label_size __UpperCamelCase : str =initializer_range __UpperCamelCase : Tuple =num_labels __UpperCamelCase : Tuple =num_choices __UpperCamelCase : str =summary_type __UpperCamelCase : Tuple =use_proj __UpperCamelCase : Tuple =scope def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __UpperCamelCase : Optional[int] =random_attention_mask([self.batch_size, self.seq_length] ) __UpperCamelCase : Dict =None if self.use_input_lengths: __UpperCamelCase : List[Any] =( ids_tensor([self.batch_size] , vocab_size=2 ) + self.seq_length - 2 ) # small variation of seq_length __UpperCamelCase : List[str] =None if self.use_token_type_ids: __UpperCamelCase : Dict =ids_tensor([self.batch_size, self.seq_length] , self.n_langs ) __UpperCamelCase : Optional[Any] =None __UpperCamelCase : int =None __UpperCamelCase : Tuple =None if self.use_labels: __UpperCamelCase : int =ids_tensor([self.batch_size] , self.type_sequence_label_size ) __UpperCamelCase : str =ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __UpperCamelCase : List[str] =ids_tensor([self.batch_size] , 2 ).float() __UpperCamelCase : Dict =ids_tensor([self.batch_size] , self.num_choices ) __UpperCamelCase : Dict =self.get_config() return ( config, input_ids, token_type_ids, input_lengths, sequence_labels, token_labels, is_impossible_labels, choice_labels, input_mask, ) def __lowercase ( self ): """simple docstring""" return FlaubertConfig( vocab_size=self.vocab_size , n_special=self.n_special , emb_dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , gelu_activation=self.gelu_activation , sinusoidal_embeddings=self.sinusoidal_embeddings , asm=self.asm , causal=self.causal , n_langs=self.n_langs , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , summary_type=self.summary_type , use_proj=self.use_proj , ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Tuple =FlaubertModel(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : Optional[int] =model(lowerCamelCase__ , lengths=lowerCamelCase__ , langs=lowerCamelCase__ ) __UpperCamelCase : str =model(lowerCamelCase__ , langs=lowerCamelCase__ ) __UpperCamelCase : Optional[Any] =model(lowerCamelCase__ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Optional[Any] =FlaubertWithLMHeadModel(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : List[Any] =model(lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : int =FlaubertForQuestionAnsweringSimple(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : Tuple =model(lowerCamelCase__ ) __UpperCamelCase : int =model(lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Tuple =FlaubertForQuestionAnswering(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : Tuple =model(lowerCamelCase__ ) __UpperCamelCase : List[str] =model( lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ , cls_index=lowerCamelCase__ , is_impossible=lowerCamelCase__ , p_mask=lowerCamelCase__ , ) __UpperCamelCase : Tuple =model( lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ , cls_index=lowerCamelCase__ , is_impossible=lowerCamelCase__ , ) (__UpperCamelCase) : int =result_with_labels.to_tuple() __UpperCamelCase : List[str] =model(lowerCamelCase__ , start_positions=lowerCamelCase__ , end_positions=lowerCamelCase__ ) (__UpperCamelCase) : Union[str, Any] =result_with_labels.to_tuple() self.parent.assertEqual(result_with_labels.loss.shape , () ) self.parent.assertEqual(result.start_top_log_probs.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual(result.start_top_index.shape , (self.batch_size, model.config.start_n_top) ) self.parent.assertEqual( result.end_top_log_probs.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual( result.end_top_index.shape , (self.batch_size, model.config.start_n_top * model.config.end_n_top) ) self.parent.assertEqual(result.cls_logits.shape , (self.batch_size,) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Any =FlaubertForSequenceClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : Tuple =model(lowerCamelCase__ ) __UpperCamelCase : Optional[int] =model(lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Dict =self.num_labels __UpperCamelCase : List[Any] =FlaubertForTokenClassification(lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : Any =model(lowerCamelCase__ , attention_mask=lowerCamelCase__ , labels=lowerCamelCase__ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , ): """simple docstring""" __UpperCamelCase : Tuple =self.num_choices __UpperCamelCase : Any =FlaubertForMultipleChoice(config=lowerCamelCase__ ) model.to(lowerCamelCase__ ) model.eval() __UpperCamelCase : str =input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase : str =token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase : Optional[Any] =input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __UpperCamelCase : List[Any] =model( lowerCamelCase__ , attention_mask=lowerCamelCase__ , token_type_ids=lowerCamelCase__ , labels=lowerCamelCase__ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =self.prepare_config_and_inputs() ( __UpperCamelCase ) : Any =config_and_inputs __UpperCamelCase : Union[str, Any] ={ "input_ids": input_ids, "token_type_ids": token_type_ids, "lengths": input_lengths, "attention_mask": input_mask, } return config, inputs_dict @require_torch class __A ( __A , __A , unittest.TestCase ): """simple docstring""" UpperCamelCase__ : Union[str, Any] =( ( FlaubertModel, FlaubertWithLMHeadModel, FlaubertForQuestionAnswering, FlaubertForQuestionAnsweringSimple, FlaubertForSequenceClassification, FlaubertForTokenClassification, FlaubertForMultipleChoice, ) if is_torch_available() else () ) UpperCamelCase__ : Optional[int] =( { """feature-extraction""": FlaubertModel, """fill-mask""": FlaubertWithLMHeadModel, """question-answering""": FlaubertForQuestionAnsweringSimple, """text-classification""": FlaubertForSequenceClassification, """token-classification""": FlaubertForTokenClassification, """zero-shot""": FlaubertForSequenceClassification, } if is_torch_available() else {} ) def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__ ): """simple docstring""" if ( pipeline_test_casse_name == "QAPipelineTests" and tokenizer_name is not None and not tokenizer_name.endswith('Fast' ) ): # `QAPipelineTests` fails for a few models when the slower tokenizer are used. # (The slower tokenizers were never used for pipeline tests before the pipeline testing rework) # TODO: check (and possibly fix) the `QAPipelineTests` with slower tokenizer return True return False def __lowercase ( self , lowerCamelCase__ , lowerCamelCase__ , lowerCamelCase__=False ): """simple docstring""" __UpperCamelCase : Dict =super()._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ , return_labels=lowerCamelCase__ ) if return_labels: if model_class.__name__ == "FlaubertForQuestionAnswering": __UpperCamelCase : Optional[Any] =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase__ ) __UpperCamelCase : Dict =torch.zeros( self.model_tester.batch_size , dtype=torch.long , device=lowerCamelCase__ ) return inputs_dict def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =FlaubertModelTester(self ) __UpperCamelCase : Dict =ConfigTester(self , config_class=lowerCamelCase__ , emb_dim=37 ) def __lowercase ( self ): """simple docstring""" self.config_tester.run_common_tests() def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_model(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_lm_head(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_simple_qa(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Tuple =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_qa(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_sequence_classif(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_token_classif(*lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_flaubert_multiple_choice(*lowerCamelCase__ ) @slow def __lowercase ( self ): """simple docstring""" for model_name in FLAUBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __UpperCamelCase : Dict =FlaubertModel.from_pretrained(lowerCamelCase__ ) self.assertIsNotNone(lowerCamelCase__ ) @slow @require_torch_gpu def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # FlauBertForMultipleChoice behaves incorrectly in JIT environments. if model_class == FlaubertForMultipleChoice: return __UpperCamelCase : Optional[int] =True __UpperCamelCase : List[str] =model_class(config=lowerCamelCase__ ) __UpperCamelCase : Optional[int] =self._prepare_for_class(lowerCamelCase__ , lowerCamelCase__ ) __UpperCamelCase : int =torch.jit.trace( lowerCamelCase__ , (inputs_dict['input_ids'].to('cpu' ), inputs_dict['attention_mask'].to('cpu' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(lowerCamelCase__ , os.path.join(lowerCamelCase__ , 'traced_model.pt' ) ) __UpperCamelCase : str =torch.jit.load(os.path.join(lowerCamelCase__ , 'traced_model.pt' ) , map_location=lowerCamelCase__ ) loaded(inputs_dict['input_ids'].to(lowerCamelCase__ ) , inputs_dict['attention_mask'].to(lowerCamelCase__ ) ) @require_torch class __A ( unittest.TestCase ): """simple docstring""" @slow def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =FlaubertModel.from_pretrained('flaubert/flaubert_base_cased' ) __UpperCamelCase : Tuple =torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) with torch.no_grad(): __UpperCamelCase : List[str] =model(lowerCamelCase__ )[0] __UpperCamelCase : int =torch.Size((1, 11, 768) ) self.assertEqual(output.shape , lowerCamelCase__ ) __UpperCamelCase : str =torch.tensor( [[[-2.6_251, -1.4_298, -0.0_227], [-2.8_510, -1.6_387, 0.2_258], [-2.8_114, -1.1_832, -0.3_066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , lowerCamelCase__ , atol=1E-4 ) )
71
a_ = 8.314_4598 def lowerCamelCase__ ( _a , _a): if temperature < 0: raise Exception("Temperature cannot be less than 0 K") if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol") else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a_ = 300 a_ = 28 a_ = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
76
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_blip_2": [ "BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Blip2Config", "Blip2QFormerConfig", "Blip2VisionConfig", ], "processing_blip_2": ["Blip2Processor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST", "Blip2Model", "Blip2QFormerModel", "Blip2PreTrainedModel", "Blip2ForConditionalGeneration", "Blip2VisionModel", ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
45
a_ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : int = set() # keep track of all the paths to be checked SCREAMING_SNAKE_CASE : int = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) # get the last node from the path SCREAMING_SNAKE_CASE : Union[str, Any] = path[-1] if node not in explored: SCREAMING_SNAKE_CASE : List[str] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: SCREAMING_SNAKE_CASE : List[Any] = list(_a) new_path.append(_a) queue.append(_a) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_a) # in case there's no path between the 2 nodes return [] def lowerCamelCase__ ( _a , _a , _a): if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 SCREAMING_SNAKE_CASE : str = [start] SCREAMING_SNAKE_CASE : Optional[Any] = set(_a) # Keep tab on distances from `start` node. SCREAMING_SNAKE_CASE : Union[str, Any] = {start: 0, target: -1} while queue: SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) if node == target: SCREAMING_SNAKE_CASE : Union[str, Any] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node]) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_a) queue.append(_a) SCREAMING_SNAKE_CASE : Optional[Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
76
0
import os import time from dataclasses import dataclass, field from enum import Enum from typing import Dict, List, Optional, Union import torch from filelock import FileLock from torch.utils.data import Dataset from ...models.auto.modeling_auto import MODEL_FOR_QUESTION_ANSWERING_MAPPING from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging from ..processors.squad import SquadFeatures, SquadVaProcessor, SquadVaProcessor, squad_convert_examples_to_features UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = list(MODEL_FOR_QUESTION_ANSWERING_MAPPING.keys()) UpperCamelCase = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class __UpperCAmelCase : __snake_case : int = field( default=__A ,metadata={"help": "Model type selected in the list: " + ", ".join(__A )} ) __snake_case : int = field( default=__A ,metadata={"help": "The input data dir. Should contain the .json files for the SQuAD task."} ) __snake_case : int = field( default=128 ,metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } ,) __snake_case : List[Any] = field( default=128 ,metadata={"help": "When splitting up a long document into chunks, how much stride to take between chunks."} ,) __snake_case : Any = field( default=64 ,metadata={ "help": ( "The maximum number of tokens for the question. Questions longer than this will " "be truncated to this length." ) } ,) __snake_case : List[str] = field( default=30 ,metadata={ "help": ( "The maximum length of an answer that can be generated. This is needed because the start " "and end predictions are not conditioned on one another." ) } ,) __snake_case : List[Any] = field( default=__A ,metadata={"help": "Overwrite the cached training and evaluation sets"} ) __snake_case : Optional[int] = field( default=__A ,metadata={"help": "If true, the SQuAD examples contain some that do not have an answer."} ) __snake_case : Tuple = field( default=0.0 ,metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) __snake_case : List[Any] = field( default=20 ,metadata={"help": "If null_score - best_non_null is greater than the threshold predict null."} ) __snake_case : Dict = field( default=0 ,metadata={ "help": ( "language id of input for language-specific xlm models (see" " tokenization_xlm.PRETRAINED_INIT_CONFIGURATION)" ) } ,) __snake_case : int = field(default=1 ,metadata={"help": "multiple threads for converting example to features"} ) class __UpperCAmelCase (__A ): __snake_case : List[Any] = "train" __snake_case : str = "dev" class __UpperCAmelCase (__A ): __snake_case : List[Any] = 42 __snake_case : List[Any] = 42 __snake_case : str = 42 __snake_case : Optional[Any] = 42 def __init__( self: List[str] , UpperCAmelCase_: SquadDataTrainingArguments , UpperCAmelCase_: PreTrainedTokenizer , UpperCAmelCase_: Optional[int] = None , UpperCAmelCase_: Union[str, Split] = Split.train , UpperCAmelCase_: Optional[bool] = False , UpperCAmelCase_: Optional[str] = None , UpperCAmelCase_: Optional[str] = "pt" , ): '''simple docstring''' _SCREAMING_SNAKE_CASE = args _SCREAMING_SNAKE_CASE = is_language_sensitive _SCREAMING_SNAKE_CASE = SquadVaProcessor() if args.version_2_with_negative else SquadVaProcessor() if isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): try: _SCREAMING_SNAKE_CASE = Split[mode] except KeyError: raise KeyError("""mode is not a valid split name""" ) _SCREAMING_SNAKE_CASE = mode # Load data features from cache or dataset file _SCREAMING_SNAKE_CASE = "v2" if args.version_2_with_negative else "v1" _SCREAMING_SNAKE_CASE = os.path.join( cache_dir if cache_dir is not None else args.data_dir , F'cached_{mode.value}_{tokenizer.__class__.__name__}_{args.max_seq_length}_{version_tag}' , ) # Make sure only the first process in distributed training processes the dataset, # and the others will use the cache. _SCREAMING_SNAKE_CASE = cached_features_file + ".lock" with FileLock(UpperCAmelCase_ ): if os.path.exists(UpperCAmelCase_ ) and not args.overwrite_cache: _SCREAMING_SNAKE_CASE = time.time() _SCREAMING_SNAKE_CASE = torch.load(UpperCAmelCase_ ) # Legacy cache files have only features, while new cache files # will have dataset and examples also. _SCREAMING_SNAKE_CASE = self.old_features["features"] _SCREAMING_SNAKE_CASE = self.old_features.get("""dataset""" , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = self.old_features.get("""examples""" , UpperCAmelCase_ ) logger.info( F'Loading features from cached file {cached_features_file} [took %.3f s]' , time.time() - start ) if self.dataset is None or self.examples is None: logger.warning( F'Deleting cached file {cached_features_file} will allow dataset and examples to be cached in' """ future run""" ) else: if mode == Split.dev: _SCREAMING_SNAKE_CASE = self.processor.get_dev_examples(args.data_dir ) else: _SCREAMING_SNAKE_CASE = self.processor.get_train_examples(args.data_dir ) _SCREAMING_SNAKE_CASE = squad_convert_examples_to_features( examples=self.examples , tokenizer=UpperCAmelCase_ , max_seq_length=args.max_seq_length , doc_stride=args.doc_stride , max_query_length=args.max_query_length , is_training=mode == Split.train , threads=args.threads , return_dataset=UpperCAmelCase_ , ) _SCREAMING_SNAKE_CASE = time.time() torch.save( {"""features""": self.features, """dataset""": self.dataset, """examples""": self.examples} , UpperCAmelCase_ , ) # ^ This seems to take a lot of time so I want to investigate why and how we can improve. logger.info( F'Saving features into cached file {cached_features_file} [took {time.time() - start:.3f} s]' ) def __len__( self: str ): '''simple docstring''' return len(self.features ) def __getitem__( self: Tuple , UpperCAmelCase_: Dict ): '''simple docstring''' _SCREAMING_SNAKE_CASE = self.features[i] _SCREAMING_SNAKE_CASE = torch.tensor(feature.input_ids , dtype=torch.long ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.attention_mask , dtype=torch.long ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.token_type_ids , dtype=torch.long ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.cls_index , dtype=torch.long ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.p_mask , dtype=torch.float ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.is_impossible , dtype=torch.float ) _SCREAMING_SNAKE_CASE = { "input_ids": input_ids, "attention_mask": attention_mask, "token_type_ids": token_type_ids, } if self.args.model_type in ["xlm", "roberta", "distilbert", "camembert"]: del inputs["token_type_ids"] if self.args.model_type in ["xlnet", "xlm"]: inputs.update({"""cls_index""": cls_index, """p_mask""": p_mask} ) if self.args.version_2_with_negative: inputs.update({"""is_impossible""": is_impossible} ) if self.is_language_sensitive: inputs.update({"""langs""": (torch.ones(input_ids.shape , dtype=torch.intaa ) * self.args.lang_id)} ) if self.mode == Split.train: _SCREAMING_SNAKE_CASE = torch.tensor(feature.start_position , dtype=torch.long ) _SCREAMING_SNAKE_CASE = torch.tensor(feature.end_position , dtype=torch.long ) inputs.update({"""start_positions""": start_positions, """end_positions""": end_positions} ) return inputs
306
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCamelCase ( self : str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : List[str] = AutoTokenizer.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : Tuple = tokenizer("Hello there" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer("Hi I am" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : str = model(a , labels=a ).loss SCREAMING_SNAKE_CASE : Any = -tf.math.reduce_mean(a ).numpy() SCREAMING_SNAKE_CASE : Union[str, Any] = -21.22_8168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
76
0
def UpperCAmelCase_ ( __lowerCAmelCase ) -> Dict: __lowercase : set[int] = set() # To detect a back edge, keep track of vertices currently in the recursion stack __lowercase : set[int] = set() return any( node not in visited and depth_first_search(_a , _a , _a , _a ) for node in graph ) def UpperCAmelCase_ ( __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase , __lowerCAmelCase ) -> str: visited.add(_a ) rec_stk.add(_a ) for node in graph[vertex]: if node not in visited: if depth_first_search(_a , _a , _a , _a ): return True elif node in rec_stk: return True # The node needs to be removed from recursion stack before function ends rec_stk.remove(_a ) return False if __name__ == "__main__": from doctest import testmod testmod()
156
from math import factorial def lowerCamelCase__ ( _a , _a , _a): if successes > trials: raise ValueError("successes must be lower or equal to trials") if trials < 0 or successes < 0: raise ValueError("the function is defined for non-negative integers") if not isinstance(_a , _a) or not isinstance(_a , _a): raise ValueError("the function is defined for non-negative integers") if not 0 < prob < 1: raise ValueError("prob has to be in range of 1 - 0") SCREAMING_SNAKE_CASE : int = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! SCREAMING_SNAKE_CASE : List[Any] = float(factorial(_a)) coefficient /= factorial(_a) * factorial(trials - successes) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('Probability of 2 successes out of 4 trails') print('with probability of 0.75 is:', end=' ') print(binomial_distribution(2, 4, 0.75))
76
0
# Usage: # ./gen-card-allenai-wmt16.py import os from pathlib import Path def _snake_case( SCREAMING_SNAKE_CASE__ : Union[str, Any] , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : Tuple ) -> Dict: '''simple docstring''' A__ = { "en": "Machine learning is great, isn't it?", "ru": "Машинное обучение - это здорово, не так ли?", "de": "Maschinelles Lernen ist großartig, nicht wahr?", } # BLUE scores as follows: # "pair": [fairseq, transformers] A__ = { "wmt16-en-de-dist-12-1": [28.3, 27.52], "wmt16-en-de-dist-6-1": [27.4, 27.11], "wmt16-en-de-12-1": [26.9, 25.75], } A__ = f'{src_lang}-{tgt_lang}' A__ = f'\n---\nlanguage:\n- {src_lang}\n- {tgt_lang}\nthumbnail:\ntags:\n- translation\n- wmt16\n- allenai\nlicense: apache-2.0\ndatasets:\n- wmt16\nmetrics:\n- bleu\n---\n\n# FSMT\n\n## Model description\n\nThis is a ported version of fairseq-based [wmt16 transformer](https://github.com/jungokasai/deep-shallow/) for {src_lang}-{tgt_lang}.\n\nFor more details, please, see [Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation](https://arxiv.org/abs/2006.10369).\n\nAll 3 models are available:\n\n* [wmt16-en-de-dist-12-1](https://huggingface.co/allenai/wmt16-en-de-dist-12-1)\n* [wmt16-en-de-dist-6-1](https://huggingface.co/allenai/wmt16-en-de-dist-6-1)\n* [wmt16-en-de-12-1](https://huggingface.co/allenai/wmt16-en-de-12-1)\n\n\n## Intended uses & limitations\n\n#### How to use\n\n```python\nfrom transformers import FSMTForConditionalGeneration, FSMTTokenizer\nmname = \"allenai/{model_name}\"\ntokenizer = FSMTTokenizer.from_pretrained(mname)\nmodel = FSMTForConditionalGeneration.from_pretrained(mname)\n\ninput = \"{texts[src_lang]}\"\ninput_ids = tokenizer.encode(input, return_tensors=\"pt\")\noutputs = model.generate(input_ids)\ndecoded = tokenizer.decode(outputs[0], skip_special_tokens=True)\nprint(decoded) # {texts[tgt_lang]}\n\n```\n\n#### Limitations and bias\n\n\n## Training data\n\nPretrained weights were left identical to the original model released by allenai. For more details, please, see the [paper](https://arxiv.org/abs/2006.10369).\n\n## Eval results\n\nHere are the BLEU scores:\n\nmodel | fairseq | transformers\n-------|---------|----------\n{model_name} | {scores[model_name][0]} | {scores[model_name][1]}\n\nThe score is slightly below the score reported in the paper, as the researchers don\'t use `sacrebleu` and measure the score on tokenized outputs. `transformers` score was measured using `sacrebleu` on detokenized outputs.\n\nThe score was calculated using this code:\n\n```bash\ngit clone https://github.com/huggingface/transformers\ncd transformers\nexport PAIR={pair}\nexport DATA_DIR=data/$PAIR\nexport SAVE_DIR=data/$PAIR\nexport BS=8\nexport NUM_BEAMS=5\nmkdir -p $DATA_DIR\nsacrebleu -t wmt16 -l $PAIR --echo src > $DATA_DIR/val.source\nsacrebleu -t wmt16 -l $PAIR --echo ref > $DATA_DIR/val.target\necho $PAIR\nPYTHONPATH=\"src:examples/seq2seq\" python examples/seq2seq/run_eval.py allenai/{model_name} $DATA_DIR/val.source $SAVE_DIR/test_translations.txt --reference_path $DATA_DIR/val.target --score_path $SAVE_DIR/test_bleu.json --bs $BS --task translation --num_beams $NUM_BEAMS\n```\n\n## Data Sources\n\n- [training, etc.](http://www.statmt.org/wmt16/)\n- [test set](http://matrix.statmt.org/test_sets/newstest2016.tgz?1504722372)\n\n\n### BibTeX entry and citation info\n\n```\n@misc{{kasai2020deep,\n title={{Deep Encoder, Shallow Decoder: Reevaluating the Speed-Quality Tradeoff in Machine Translation}},\n author={{Jungo Kasai and Nikolaos Pappas and Hao Peng and James Cross and Noah A. Smith}},\n year={{2020}},\n eprint={{2006.10369}},\n archivePrefix={{arXiv}},\n primaryClass={{cs.CL}}\n}}\n```\n\n' model_card_dir.mkdir(parents=_a , exist_ok=_a ) A__ = os.path.join(_a , 'README.md' ) print(f'Generating {path}' ) with open(_a , 'w' , encoding='utf-8' ) as f: f.write(_a ) # make sure we are under the root of the project lowercase_ = Path(__file__).resolve().parent.parent.parent lowercase_ = repo_dir / "model_cards" for model_name in ["wmt16-en-de-dist-12-1", "wmt16-en-de-dist-6-1", "wmt16-en-de-12-1"]: lowercase_ = model_cards_dir / "allenai" / model_name write_model_card(model_card_dir, src_lang="en", tgt_lang="de", model_name=model_name)
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
from __future__ import annotations class A_ : def __init__( self , _A , _A ): '''simple docstring''' UpperCAmelCase = text, pattern UpperCAmelCase = len(_A ), len(_A ) def _lowercase ( self , _A ): '''simple docstring''' for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def _lowercase ( self , _A ): '''simple docstring''' for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = [] for i in range(self.textLen - self.patLen + 1 ): UpperCAmelCase = self.mismatch_in_text(_A ) if mismatch_index == -1: positions.append(_A ) else: UpperCAmelCase = self.match_in_pattern(self.text[mismatch_index] ) UpperCAmelCase = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions __A : List[str] = "ABAABA" __A : Tuple = "AB" __A : List[str] = BoyerMooreSearch(text, pattern) __A : List[Any] = bms.bad_character_heuristic() if len(positions) == 0: print("No match found") else: print("Pattern found in following positions: ") print(positions)
273
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a_ = logging.getLogger(__name__) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = False def __UpperCamelCase ( self : str , a : str , a : Optional[int] , a : Any , a : str ) -> List[Any]: """simple docstring""" if not self.initialized: SCREAMING_SNAKE_CASE : List[str] = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[int] = True def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def __UpperCamelCase ( self : Optional[Any] , a : List[Any] , a : Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Dict = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : Tuple , a : Any , a : Tuple , a : Tuple , a : Tuple , a : List[Any]=None ) -> Optional[int]: """simple docstring""" if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def __UpperCamelCase ( self : Any ) -> Dict: """simple docstring""" logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __UpperCamelCase ( self : Tuple , a : Optional[int] , a : Any ) -> int: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. SCREAMING_SNAKE_CASE : Optional[Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = ray.get(random_worker.retrieve.remote(a , a ) ) else: SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Any = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def __UpperCamelCase ( cls : str , a : Optional[Any] , a : Any=None , **a : List[Any] ) -> str: """simple docstring""" return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def __UpperCamelCase ( cls : Union[str, Any] , a : int , a : Any , a : List[Any]=None , **a : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : str = kwargs.pop("config" , a ) or RagConfig.from_pretrained(a , **a ) SCREAMING_SNAKE_CASE : List[Any] = RagTokenizer.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.question_encoder SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.generator if indexed_dataset is not None: SCREAMING_SNAKE_CASE : str = "custom" SCREAMING_SNAKE_CASE : List[Any] = CustomHFIndex(config.retrieval_vector_size , a ) else: SCREAMING_SNAKE_CASE : List[str] = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
76
0
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class snake_case__ ( unittest.TestCase ): @slow def __magic_name__ ( self ) -> List[str]: __magic_name__ : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained("""google/mt5-small""" ) __magic_name__ : List[str] = AutoTokenizer.from_pretrained("""google/mt5-small""" ) __magic_name__ : Tuple = tokenizer("""Hello there""" , return_tensors="""tf""" ).input_ids __magic_name__ : Optional[Any] = tokenizer("""Hi I am""" , return_tensors="""tf""" ).input_ids __magic_name__ : str = model(lowerCAmelCase__ , labels=lowerCAmelCase__ ).loss __magic_name__ : Any = -tf.math.reduce_mean(lowerCAmelCase__ ).numpy() __magic_name__ : Union[str, Any] = -2_1.2_2_8_1_6_8 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
342
from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : Dict , a : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : int = data SCREAMING_SNAKE_CASE : int = None def __repr__( self : str ) -> str: """simple docstring""" return F"Node({self.data})" class _UpperCamelCase : '''simple docstring''' def __init__( self : List[str] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = None def __iter__( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.head while node: yield node.data SCREAMING_SNAKE_CASE : List[str] = node.next def __len__( self : str ) -> int: """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return "->".join([str(a ) for item in self] ) def __getitem__( self : List[Any] , a : int ) -> Any: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Tuple , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) SCREAMING_SNAKE_CASE : str = self.head for _ in range(a ): SCREAMING_SNAKE_CASE : str = current.next SCREAMING_SNAKE_CASE : Any = data def __UpperCamelCase ( self : List[str] , a : Any ) -> None: """simple docstring""" self.insert_nth(len(self ) , a ) def __UpperCamelCase ( self : Union[str, Any] , a : Any ) -> None: """simple docstring""" self.insert_nth(0 , a ) def __UpperCamelCase ( self : Optional[Any] , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) SCREAMING_SNAKE_CASE : Any = Node(a ) if self.head is None: SCREAMING_SNAKE_CASE : Optional[int] = new_node elif index == 0: SCREAMING_SNAKE_CASE : Optional[int] = self.head # link new_node to head SCREAMING_SNAKE_CASE : List[Any] = new_node else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : int = new_node def __UpperCamelCase ( self : Optional[int] ) -> None: # print every node data """simple docstring""" print(self ) def __UpperCamelCase ( self : int ) -> Any: """simple docstring""" return self.delete_nth(0 ) def __UpperCamelCase ( self : Any ) -> Any: # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def __UpperCamelCase ( self : List[str] , a : int = 0 ) -> Any: """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) SCREAMING_SNAKE_CASE : Tuple = self.head # default first node if index == 0: SCREAMING_SNAKE_CASE : List[str] = self.head.next else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Any = temp.next SCREAMING_SNAKE_CASE : List[Any] = temp.next SCREAMING_SNAKE_CASE : List[str] = temp.next.next return delete_node.data def __UpperCamelCase ( self : List[Any] ) -> bool: """simple docstring""" return self.head is None def __UpperCamelCase ( self : Optional[int] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : str = self.head while current: # Store the current node's next node. SCREAMING_SNAKE_CASE : Any = current.next # Make the current node's next point backwards SCREAMING_SNAKE_CASE : List[Any] = prev # Make the previous node be the current node SCREAMING_SNAKE_CASE : Any = current # Make the current node the next node (to progress iteration) SCREAMING_SNAKE_CASE : str = next_node # Return prev in order to put the head at the end SCREAMING_SNAKE_CASE : Optional[Any] = prev def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_a) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10): assert len(_a) == i linked_list.insert_nth(_a , i + 1) assert str(_a) == "->".join(str(_a) for i in range(1 , 11)) linked_list.insert_head(0) linked_list.insert_tail(11) assert str(_a) == "->".join(str(_a) for i in range(0 , 12)) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9) == 10 assert linked_list.delete_tail() == 11 assert len(_a) == 9 assert str(_a) == "->".join(str(_a) for i in range(1 , 10)) assert all(linked_list[i] == i + 1 for i in range(0 , 9)) is True for i in range(0 , 9): SCREAMING_SNAKE_CASE : str = -i assert all(linked_list[i] == -i for i in range(0 , 9)) is True linked_list.reverse() assert str(_a) == "->".join(str(_a) for i in range(-8 , 1)) def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Optional[Any] = [ -9, 100, Node(77345112), "dlrow olleH", 7, 5555, 0, -192.5_5555, "Hello, world!", 77.9, Node(10), None, None, 12.20, ] SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in test_input: linked_list.insert_tail(_a) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_a) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head SCREAMING_SNAKE_CASE : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail SCREAMING_SNAKE_CASE : Any = linked_list.delete_tail() assert result == 12.2 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list SCREAMING_SNAKE_CASE : Any = linked_list.delete_nth(10) assert result is None assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!")) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_a) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_a) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowerCamelCase__ ( ): from doctest import testmod testmod() SCREAMING_SNAKE_CASE : Optional[int] = LinkedList() linked_list.insert_head(input("Inserting 1st at head ").strip()) linked_list.insert_head(input("Inserting 2nd at head ").strip()) print("\nPrint list:") linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail ").strip()) linked_list.insert_tail(input("Inserting 2nd at tail ").strip()) print("\nPrint list:") linked_list.print_list() print("\nDelete head") linked_list.delete_head() print("Delete tail") linked_list.delete_tail() print("\nPrint list:") linked_list.print_list() print("\nReverse linked list") linked_list.reverse() print("\nPrint list:") linked_list.print_list() print("\nString representation of linked list:") print(_a) print("\nReading/changing Node data using indexing:") print(f"Element at Position 1: {linked_list[1]}") SCREAMING_SNAKE_CASE : Dict = input("Enter New Value: ").strip() print("New list:") print(_a) print(f"length of linked_list is : {len(_a)}") if __name__ == "__main__": main()
76
0
'''simple docstring''' import copy from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING a__ : Dict = logging.get_logger(__name__) a__ : Union[str, Any] = { 'microsoft/conditional-detr-resnet-50': ( 'https://huggingface.co/microsoft/conditional-detr-resnet-50/resolve/main/config.json' ), } class lowercase_ ( __A ): __UpperCAmelCase = 'conditional_detr' __UpperCAmelCase = ['past_key_values'] __UpperCAmelCase = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self , a=True , a=None , a=3 , a=3_00 , a=6 , a=20_48 , a=8 , a=6 , a=20_48 , a=8 , a=0.0 , a=0.0 , a=True , a="relu" , a=2_56 , a=0.1 , a=0.0 , a=0.0 , a=0.02 , a=1.0 , a=False , a="sine" , a="resnet50" , a=True , a=False , a=2 , a=5 , a=2 , a=1 , a=1 , a=2 , a=5 , a=2 , a=0.25 , **a , ): if backbone_config is not None and use_timm_backbone: raise ValueError("You can't specify both `backbone_config` and `use_timm_backbone`." ) if not use_timm_backbone: if backbone_config is None: logger.info("`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone." ) UpperCamelCase__ = CONFIG_MAPPING["resnet"](out_features=["stage4"] ) elif isinstance(a , a ): UpperCamelCase__ = backbone_config.get("model_type" ) UpperCamelCase__ = CONFIG_MAPPING[backbone_model_type] UpperCamelCase__ = config_class.from_dict(a ) UpperCamelCase__ = use_timm_backbone UpperCamelCase__ = backbone_config UpperCamelCase__ = num_channels UpperCamelCase__ = num_queries UpperCamelCase__ = d_model UpperCamelCase__ = encoder_ffn_dim UpperCamelCase__ = encoder_layers UpperCamelCase__ = encoder_attention_heads UpperCamelCase__ = decoder_ffn_dim UpperCamelCase__ = decoder_layers UpperCamelCase__ = decoder_attention_heads UpperCamelCase__ = dropout UpperCamelCase__ = attention_dropout UpperCamelCase__ = activation_dropout UpperCamelCase__ = activation_function UpperCamelCase__ = init_std UpperCamelCase__ = init_xavier_std UpperCamelCase__ = encoder_layerdrop UpperCamelCase__ = decoder_layerdrop UpperCamelCase__ = encoder_layers UpperCamelCase__ = auxiliary_loss UpperCamelCase__ = position_embedding_type UpperCamelCase__ = backbone UpperCamelCase__ = use_pretrained_backbone UpperCamelCase__ = dilation # Hungarian matcher UpperCamelCase__ = class_cost UpperCamelCase__ = bbox_cost UpperCamelCase__ = giou_cost # Loss coefficients UpperCamelCase__ = mask_loss_coefficient UpperCamelCase__ = dice_loss_coefficient UpperCamelCase__ = cls_loss_coefficient UpperCamelCase__ = bbox_loss_coefficient UpperCamelCase__ = giou_loss_coefficient UpperCamelCase__ = focal_alpha super().__init__(is_encoder_decoder=a , **a ) @property def __a ( self ): return self.encoder_attention_heads @property def __a ( self ): return self.d_model def __a ( self ): UpperCamelCase__ = copy.deepcopy(self.__dict__ ) if self.backbone_config is not None: UpperCamelCase__ = self.backbone_config.to_dict() UpperCamelCase__ = self.__class__.model_type return output class lowercase_ ( __A ): __UpperCAmelCase = version.parse('1.11' ) @property def __a ( self ): return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ("pixel_mask", {0: "batch"}), ] ) @property def __a ( self ): return 1e-5 @property def __a ( self ): return 12
80
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ = get_logger(__name__) class _UpperCamelCase ( enum.Enum ): '''simple docstring''' lowerCamelCase__ ='all_checks' lowerCamelCase__ ='basic_checks' lowerCamelCase__ ='no_checks' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a , _a=None): if expected_checksums is None: logger.info("Unable to verify checksums.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedDownloadedFile(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE : Tuple = " for " + verification_name if verification_name is not None else "" if len(_a) > 0: raise NonMatchingChecksumError( f"Checksums didn't match{for_verification_name}:\n" f"{bad_urls}\n" "Set `verification_mode='no_checks'` to skip checksums verification and ignore this error") logger.info("All the checksums matched successfully" + for_verification_name) class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a): if expected_splits is None: logger.info("Unable to verify splits sizes.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreSplits(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedSplits(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : List[str] = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_a) > 0: raise NonMatchingSplitsSizesError(str(_a)) logger.info("All the splits matched successfully.") def lowerCamelCase__ ( _a , _a = True): if record_checksum: SCREAMING_SNAKE_CASE : List[str] = shaaaa() with open(_a , "rb") as f: for chunk in iter(lambda: f.read(1 << 20) , b""): m.update(_a) SCREAMING_SNAKE_CASE : Optional[int] = m.hexdigest() else: SCREAMING_SNAKE_CASE : List[str] = None return {"num_bytes": os.path.getsize(_a), "checksum": checksum} def lowerCamelCase__ ( _a): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
76
0
def _A ( SCREAMING_SNAKE_CASE__ : Dict = 3 , SCREAMING_SNAKE_CASE__ : Union[str, Any] = 7 , SCREAMING_SNAKE_CASE__ : Tuple = 1000000 ): UpperCamelCase :str = 0 UpperCamelCase :int = 1 for current_denominator in range(1 , limit + 1 ): UpperCamelCase :int = current_denominator * numerator // denominator if current_denominator % denominator == 0: current_numerator -= 1 if current_numerator * max_denominator > current_denominator * max_numerator: UpperCamelCase :List[str] = current_numerator UpperCamelCase :List[Any] = current_denominator return max_numerator if __name__ == "__main__": print(solution(numerator=3, denominator=7, limit=1_00_00_00))
259
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( _a , _a): # Load checkpoint SCREAMING_SNAKE_CASE : int = torch.load(_a , map_location="cpu") SCREAMING_SNAKE_CASE : Dict = chkpt["model"] # We have the base model one level deeper than the original XLM repository SCREAMING_SNAKE_CASE : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: SCREAMING_SNAKE_CASE : List[str] = v else: SCREAMING_SNAKE_CASE : int = v SCREAMING_SNAKE_CASE : int = chkpt["params"] SCREAMING_SNAKE_CASE : Union[str, Any] = {n: v for n, v in config.items() if not isinstance(_a , (torch.FloatTensor, numpy.ndarray))} SCREAMING_SNAKE_CASE : List[Any] = chkpt["dico_word2id"] SCREAMING_SNAKE_CASE : List[Any] = {s + "</w>" if s.find("@@") == -1 and i > 13 else s.replace("@@" , ""): i for s, i in vocab.items()} # Save pytorch-model SCREAMING_SNAKE_CASE : Tuple = pytorch_dump_folder_path + "/" + WEIGHTS_NAME SCREAMING_SNAKE_CASE : Any = pytorch_dump_folder_path + "/" + CONFIG_NAME SCREAMING_SNAKE_CASE : Optional[int] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(f"Save PyTorch model to {pytorch_weights_dump_path}") torch.save(_a , _a) print(f"Save configuration file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") print(f"Save vocab file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) a_ = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
76
0
import os def lowerCamelCase_ ( lowerCamelCase__ = "input.txt" ): with open(os.path.join(os.path.dirname(_a ) , _a ) ) as input_file: lowerCamelCase_ = [ [int(_a ) for element in line.split("," )] for line in input_file.readlines() ] lowerCamelCase_ = len(_a ) lowerCamelCase_ = len(matrix[0] ) lowerCamelCase_ = [[-1 for _ in range(_a )] for _ in range(_a )] for i in range(_a ): lowerCamelCase_ = matrix[i][0] for j in range(1 , _a ): for i in range(_a ): lowerCamelCase_ = minimal_path_sums[i][j - 1] + matrix[i][j] for i in range(1 , _a ): lowerCamelCase_ = min( minimal_path_sums[i][j] , minimal_path_sums[i - 1][j] + matrix[i][j] ) for i in range(rows - 2 , -1 , -1 ): lowerCamelCase_ = min( minimal_path_sums[i][j] , minimal_path_sums[i + 1][j] + matrix[i][j] ) return min(minimal_path_sums_row[-1] for minimal_path_sums_row in minimal_path_sums ) if __name__ == "__main__": print(F"""{solution() = }""")
19
def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(a - b) for a, b in zip(_a , _a))) def lowerCamelCase__ ( _a): if point: if isinstance(_a , _a): for item in point: if not isinstance(_a , (int, float)): SCREAMING_SNAKE_CASE : List[Any] = ( "Expected a list of numbers as input, found " f"{type(_a).__name__}" ) raise TypeError(_a) else: SCREAMING_SNAKE_CASE : List[Any] = f"Expected a list of numbers as input, found {type(_a).__name__}" raise TypeError(_a) else: raise ValueError("Missing an input") def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(x - y) for x, y in zip(_a , _a))) if __name__ == "__main__": import doctest doctest.testmod()
76
0
import tempfile import unittest from pathlib import Path from shutil import copyfile from transformers import MaMaaaTokenizer, is_torch_available from transformers.testing_utils import ( get_tests_dir, nested_simplify, require_sentencepiece, require_tokenizers, require_torch, slow, ) from transformers.utils import is_sentencepiece_available if is_sentencepiece_available(): from transformers.models.mam_aaa.tokenization_mam_aaa import VOCAB_FILES_NAMES, save_json from ...test_tokenization_common import TokenizerTesterMixin if is_sentencepiece_available(): lowerCAmelCase__ : List[str] = get_tests_dir('''fixtures/test_sentencepiece.model''') if is_torch_available(): from transformers.models.mam_aaa.modeling_mam_aaa import shift_tokens_right lowerCAmelCase__ : Optional[Any] = 12_80_22 lowerCAmelCase__ : int = 12_80_28 @require_sentencepiece class __snake_case ( __A ,unittest.TestCase ): __lowerCamelCase = MaMaaaTokenizer __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = True def __a ( self ) -> str: '''simple docstring''' super().setUp() snake_case__ : int = ["</s>", "<unk>", "▁This", "▁is", "▁a", "▁t", "est", "\u0120", "<pad>"] snake_case__ : str = dict(zip(__UpperCamelCase , range(len(__UpperCamelCase ) ) ) ) snake_case__ : Union[str, Any] = Path(self.tmpdirname ) save_json(__UpperCamelCase , save_dir / VOCAB_FILES_NAMES['vocab_file'] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(__UpperCamelCase , save_dir / VOCAB_FILES_NAMES['spm_file'] ) snake_case__ : List[str] = MaMaaaTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def __a ( self , **__UpperCamelCase ) -> List[Any]: '''simple docstring''' return MaMaaaTokenizer.from_pretrained(self.tmpdirname , **__UpperCamelCase ) def __a ( self , __UpperCamelCase ) -> Any: '''simple docstring''' return ( "This is a test", "This is a test", ) def __a ( self ) -> Dict: '''simple docstring''' snake_case__ : List[Any] = "</s>" snake_case__ : List[str] = 0 self.assertEqual(self.get_tokenizer()._convert_token_to_id(__UpperCamelCase ) , __UpperCamelCase ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(__UpperCamelCase ) , __UpperCamelCase ) def __a ( self ) -> Union[str, Any]: '''simple docstring''' snake_case__ : List[str] = self.get_tokenizer() snake_case__ : List[Any] = list(tokenizer.get_vocab().keys() ) self.assertEqual(vocab_keys[0] , '</s>' ) self.assertEqual(vocab_keys[1] , '<unk>' ) self.assertEqual(vocab_keys[-1] , '<s>' ) self.assertEqual(len(__UpperCamelCase ) , tokenizer.vocab_size + len(tokenizer.get_added_vocab() ) ) @unittest.skip('Skip this test while all models are still to be uploaded.' ) def __a ( self ) -> Dict: '''simple docstring''' pass def __a ( self ) -> List[str]: '''simple docstring''' snake_case__ : Union[str, Any] = self.get_tokenizer() snake_case__ : Tuple = tokenizer.tokenize('This is a test' ) self.assertListEqual(__UpperCamelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(__UpperCamelCase ) , [2, 3, 4, 5, 6] , ) snake_case__ : Optional[Any] = tokenizer.convert_ids_to_tokens([2, 3, 4, 5, 6] ) self.assertListEqual(__UpperCamelCase , ['▁This', '▁is', '▁a', '▁t', 'est'] ) snake_case__ : List[Any] = tokenizer.convert_tokens_to_string(__UpperCamelCase ) self.assertEqual(__UpperCamelCase , 'This is a test' ) @slow def __a ( self ) -> Any: '''simple docstring''' snake_case__ : Optional[Any] = {"input_ids": [[128022, 110108, 397, 11, 38272, 2247, 124811, 285, 18105, 1586, 207, 7, 39534, 4428, 397, 1019, 18105, 1586, 207, 7, 41337, 16786, 241, 7, 20214, 17, 125690, 10398, 7, 44378, 58069, 68342, 7798, 7343, 11, 299, 33310, 4, 158, 37350, 94077, 4569, 299, 33310, 90, 4, 52840, 290, 4, 31270, 112, 299, 682, 4, 52840, 39953, 14079, 193, 52519, 90894, 17894, 120697, 11, 40445, 551, 17, 1019, 52519, 90894, 17756, 963, 11, 40445, 480, 17, 9792, 1120, 5173, 1393, 6240, 16786, 241, 120996, 28, 1245, 1393, 118240, 11123, 1019, 93612, 2691, 10618, 98058, 120409, 1928, 279, 4, 40683, 367, 178, 207, 1019, 103, 103121, 506, 65296, 5, 2], [128022, 21217, 367, 117, 125450, 128, 719, 7, 7308, 40, 93612, 12669, 1116, 16704, 71, 17785, 3699, 15592, 35, 144, 9584, 241, 11943, 713, 950, 799, 2247, 88427, 150, 149, 118813, 120706, 1019, 106906, 81518, 28, 1224, 22799, 397, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [128022, 1658, 123311, 5155, 5578, 4722, 279, 14947, 2366, 1120, 1197, 14, 1348, 9232, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=__UpperCamelCase , model_name='facebook/m2m100_418M' , revision='c168bae485c864188cf9aa0e4108b0b6934dc91e' , ) @require_torch @require_sentencepiece @require_tokenizers class __snake_case ( unittest.TestCase ): __lowerCamelCase = """facebook/m2m100_418M""" __lowerCamelCase = [ """In my opinion, there are two levels of response from the French government.""", """NSA Affair Emphasizes Complete Lack of Debate on Intelligence""", ] __lowerCamelCase = [ """Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.""", """L\'affaire NSA souligne l\'absence totale de débat sur le renseignement""", ] # fmt: off __lowerCamelCase = [EN_CODE, 593, 1949, 11_5781, 4, 7_1586, 4234, 6_0633, 12_6233, 432, 12_3808, 1_5592, 1197, 11_7132, 12_0618, 5, 2] @classmethod def __a ( cls ) -> List[str]: '''simple docstring''' snake_case__ : MaMaaaTokenizer = MaMaaaTokenizer.from_pretrained( cls.checkpoint_name , src_lang='en' , tgt_lang='fr' ) snake_case__ : Union[str, Any] = 1 return cls def __a ( self ) -> Dict: '''simple docstring''' self.assertEqual(self.tokenizer.get_lang_id('ar' ) , 128006 ) self.assertEqual(self.tokenizer.get_lang_id('en' ) , 128022 ) self.assertEqual(self.tokenizer.get_lang_id('ro' ) , 128076 ) self.assertEqual(self.tokenizer.get_lang_id('mr' ) , 128063 ) def __a ( self ) -> int: '''simple docstring''' snake_case__ : Any = self.tokenizer.get_vocab() self.assertEqual(len(__UpperCamelCase ) , self.tokenizer.vocab_size ) self.assertEqual(vocab['<unk>'] , 3 ) self.assertIn(self.tokenizer.get_lang_token('en' ) , __UpperCamelCase ) def __a ( self ) -> Tuple: '''simple docstring''' snake_case__ : Any = "en" snake_case__ : List[Any] = self.tokenizer.batch_encode_plus(self.src_text ).input_ids[0] self.assertListEqual(self.expected_src_tokens , __UpperCamelCase ) def __a ( self ) -> List[Any]: '''simple docstring''' self.assertIn(__UpperCamelCase , self.tokenizer.all_special_ids ) # fmt: off snake_case__ : Union[str, Any] = [FR_CODE, 5364, 82, 8642, 4, 294, 47, 8, 14028, 136, 3286, 9706, 6, 90797, 6, 144012, 162, 88128, 30061, 5, 2] # fmt: on snake_case__ : str = self.tokenizer.decode(__UpperCamelCase , skip_special_tokens=__UpperCamelCase ) snake_case__ : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=__UpperCamelCase ) self.assertEqual(__UpperCamelCase , __UpperCamelCase ) self.assertNotIn(self.tokenizer.eos_token , __UpperCamelCase ) def __a ( self ) -> Union[str, Any]: '''simple docstring''' snake_case__ : Tuple = tempfile.mkdtemp() snake_case__ : Dict = self.tokenizer.lang_token_to_id self.tokenizer.save_pretrained(__UpperCamelCase ) snake_case__ : int = MaMaaaTokenizer.from_pretrained(__UpperCamelCase ) self.assertDictEqual(new_tok.lang_token_to_id , __UpperCamelCase ) @require_torch def __a ( self ) -> Optional[Any]: '''simple docstring''' snake_case__ : Tuple = "en" snake_case__ : List[Any] = "fr" snake_case__ : str = self.tokenizer(self.src_text , text_target=self.tgt_text , padding=__UpperCamelCase , return_tensors='pt' ) snake_case__ : int = shift_tokens_right( batch['labels'] , self.tokenizer.pad_token_id , self.tokenizer.eos_token_id ) for k in batch: snake_case__ : List[str] = batch[k].tolist() # batch = {k: v.tolist() for k,v in batch.items()} # fairseq batch: https://gist.github.com/sshleifer/cba08bc2109361a74ac3760a7e30e4f4 # batch.decoder_inputs_ids[0][0] == assert batch.input_ids[1][0] == EN_CODE assert batch.input_ids[1][-1] == 2 assert batch.labels[1][0] == FR_CODE assert batch.labels[1][-1] == 2 assert batch.decoder_input_ids[1][:2] == [2, FR_CODE] @require_torch def __a ( self ) -> Optional[int]: '''simple docstring''' snake_case__ : Any = "mr" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('mr' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) snake_case__ : Any = "zh" self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('zh' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) @require_torch def __a ( self ) -> int: '''simple docstring''' snake_case__ : List[Any] = "mr" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('mr' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) snake_case__ : int = "zh" self.tokenizer._switch_to_target_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id('zh' )] ) self.assertListEqual(self.tokenizer.suffix_tokens , [self.tokenizer.eos_token_id] ) self.tokenizer._switch_to_input_mode() self.assertListEqual(self.tokenizer.prefix_tokens , [self.tokenizer.get_lang_id(self.tokenizer.src_lang )] ) @require_torch def __a ( self ) -> Tuple: '''simple docstring''' snake_case__ : List[str] = self.tokenizer._build_translation_inputs('A test' , return_tensors='pt' , src_lang='en' , tgt_lang='ar' ) self.assertEqual( nested_simplify(__UpperCamelCase ) , { # en_XX, A, test, EOS 'input_ids': [[128022, 58, 4183, 2]], 'attention_mask': [[1, 1, 1, 1]], # ar_AR 'forced_bos_token_id': 128006, } , )
143
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ ='vit_msn' def __init__( self : str , a : Tuple=768 , a : Tuple=12 , a : Any=12 , a : int=3072 , a : List[Any]="gelu" , a : Dict=0.0 , a : int=0.0 , a : str=0.02 , a : List[str]=1e-06 , a : List[Any]=224 , a : Union[str, Any]=16 , a : Union[str, Any]=3 , a : Tuple=True , **a : Dict , ) -> List[Any]: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Dict = hidden_size SCREAMING_SNAKE_CASE : Optional[Any] = num_hidden_layers SCREAMING_SNAKE_CASE : Optional[Any] = num_attention_heads SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : Any = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = initializer_range SCREAMING_SNAKE_CASE : int = layer_norm_eps SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : Tuple = patch_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : List[str] = qkv_bias
76
0
from ...configuration_utils import PretrainedConfig from ...utils import logging A_ :Tuple = logging.get_logger(__name__) A_ :int = { '''sayakpaul/vit-msn-base''': '''https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json''', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class __A ( __A ): """simple docstring""" UpperCamelCase__ : int ="""vit_msn""" def __init__( self , lowerCamelCase__=768 , lowerCamelCase__=12 , lowerCamelCase__=12 , lowerCamelCase__=3072 , lowerCamelCase__="gelu" , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.02 , lowerCamelCase__=1E-06 , lowerCamelCase__=224 , lowerCamelCase__=16 , lowerCamelCase__=3 , lowerCamelCase__=True , **lowerCamelCase__ , ): """simple docstring""" super().__init__(**lowerCamelCase__ ) __UpperCamelCase : Dict =hidden_size __UpperCamelCase : Optional[Any] =num_hidden_layers __UpperCamelCase : Optional[Any] =num_attention_heads __UpperCamelCase : Optional[int] =intermediate_size __UpperCamelCase : int =hidden_act __UpperCamelCase : Union[str, Any] =hidden_dropout_prob __UpperCamelCase : Any =attention_probs_dropout_prob __UpperCamelCase : List[Any] =initializer_range __UpperCamelCase : int =layer_norm_eps __UpperCamelCase : Dict =image_size __UpperCamelCase : Tuple =patch_size __UpperCamelCase : Optional[int] =num_channels __UpperCamelCase : List[str] =qkv_bias
71
import baseaa def lowerCamelCase__ ( _a): return baseaa.aaaencode(string.encode("utf-8")) def lowerCamelCase__ ( _a): return baseaa.aaadecode(_a).decode("utf-8") if __name__ == "__main__": import doctest doctest.testmod()
76
0
"""simple docstring""" def lowercase ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Dict ) -> Dict: # Check if the input is valid if not len(_a ) == len(_a ) == 3: raise ValueError('''Please enter a valid equation.''' ) if equationa[0] == equationa[1] == equationa[0] == equationa[1] == 0: raise ValueError('''Both a & b of two equations can\'t be zero.''' ) # Extract the coefficients __a = equationa __a = equationa # Calculate the determinants of the matrices __a = aa * ba - aa * ba __a = ca * ba - ca * ba __a = aa * ca - aa * ca # Check if the system of linear equations has a solution (using Cramer's rule) if determinant == 0: if determinant_x == determinant_y == 0: raise ValueError('''Infinite solutions. (Consistent system)''' ) else: raise ValueError('''No solution. (Inconsistent system)''' ) else: if determinant_x == determinant_y == 0: # Trivial solution (Inconsistent system) return (0.0, 0.0) else: __a = determinant_x / determinant __a = determinant_y / determinant # Non-Trivial Solution (Consistent system) return (x, y)
45
from datetime import datetime as dt import os from github import Github a_ = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : int = Github(os.environ["GITHUB_TOKEN"]) SCREAMING_SNAKE_CASE : List[str] = g.get_repo("huggingface/transformers") SCREAMING_SNAKE_CASE : Optional[int] = repo.get_issues(state="open") for issue in open_issues: SCREAMING_SNAKE_CASE : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda _a: i.created_at , reverse=_a) SCREAMING_SNAKE_CASE : str = comments[0] if len(_a) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="closed") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") if __name__ == "__main__": main()
76
0
import random from typing import Any def __lowerCamelCase ( snake_case__ ) -> Union[str, Any]: """simple docstring""" for _ in range(len(_a ) ): _SCREAMING_SNAKE_CASE = random.randint(0 ,len(_a ) - 1 ) _SCREAMING_SNAKE_CASE = random.randint(0 ,len(_a ) - 1 ) _SCREAMING_SNAKE_CASE = data[b], data[a] return data if __name__ == "__main__": UpperCamelCase = [0, 1, 2, 3, 4, 5, 6, 7] UpperCamelCase = ['''python''', '''says''', '''hello''', '''!'''] print('''Fisher-Yates Shuffle:''') print('''List''', integers, strings) print('''FY Shuffle''', fisher_yates_shuffle(integers), fisher_yates_shuffle(strings))
306
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): if isinstance(_a , (list, tuple)) and isinstance(videos[0] , (list, tuple)) and is_valid_image(videos[0][0]): return videos elif isinstance(_a , (list, tuple)) and is_valid_image(videos[0]): return [videos] elif is_valid_image(_a): return [[videos]] raise ValueError(f"Could not make batched video from {videos}") class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =['pixel_values'] def __init__( self : Optional[Any] , a : bool = True , a : Dict[str, int] = None , a : PILImageResampling = PILImageResampling.BILINEAR , a : bool = True , a : Dict[str, int] = None , a : bool = True , a : Union[int, float] = 1 / 255 , a : bool = True , a : bool = True , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , **a : Tuple , ) -> None: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Tuple = size if size is not None else {"shortest_edge": 256} SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE : str = get_size_dict(a , param_name="crop_size" ) SCREAMING_SNAKE_CASE : Dict = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop SCREAMING_SNAKE_CASE : int = crop_size SCREAMING_SNAKE_CASE : int = resample SCREAMING_SNAKE_CASE : Any = do_rescale SCREAMING_SNAKE_CASE : int = rescale_factor SCREAMING_SNAKE_CASE : Tuple = offset SCREAMING_SNAKE_CASE : str = do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __UpperCamelCase ( self : Optional[Any] , a : np.ndarray , a : Dict[str, int] , a : PILImageResampling = PILImageResampling.BILINEAR , a : Optional[Union[str, ChannelDimension]] = None , **a : Union[str, Any] , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) if "shortest_edge" in size: SCREAMING_SNAKE_CASE : str = get_resize_output_image_size(a , size["shortest_edge"] , default_to_square=a ) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE : Dict = (size["height"], size["width"]) else: raise ValueError(F"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(a , size=a , resample=a , data_format=a , **a ) def __UpperCamelCase ( self : List[str] , a : np.ndarray , a : Dict[str, int] , a : Optional[Union[str, ChannelDimension]] = None , **a : str , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : str = get_size_dict(a ) if "height" not in size or "width" not in size: raise ValueError(F"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(a , size=(size["height"], size["width"]) , data_format=a , **a ) def __UpperCamelCase ( self : List[Any] , a : np.ndarray , a : Union[int, float] , a : bool = True , a : Optional[Union[str, ChannelDimension]] = None , **a : Tuple , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : int = image.astype(np.floataa ) if offset: SCREAMING_SNAKE_CASE : Union[str, Any] = image - (scale / 2) return rescale(a , scale=a , data_format=a , **a ) def __UpperCamelCase ( self : int , a : np.ndarray , a : Union[float, List[float]] , a : Union[float, List[float]] , a : Optional[Union[str, ChannelDimension]] = None , **a : List[str] , ) -> np.ndarray: """simple docstring""" return normalize(a , mean=a , std=a , data_format=a , **a ) def __UpperCamelCase ( self : Tuple , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: """simple docstring""" if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) if offset and not do_rescale: raise ValueError("For offset, do_rescale must also be set to True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : List[str] = to_numpy_array(a ) if do_resize: SCREAMING_SNAKE_CASE : Optional[Any] = self.resize(image=a , size=a , resample=a ) if do_center_crop: SCREAMING_SNAKE_CASE : Union[str, Any] = self.center_crop(a , size=a ) if do_rescale: SCREAMING_SNAKE_CASE : Any = self.rescale(image=a , scale=a , offset=a ) if do_normalize: SCREAMING_SNAKE_CASE : Tuple = self.normalize(image=a , mean=a , std=a ) SCREAMING_SNAKE_CASE : Optional[int] = to_channel_dimension_format(a , a ) return image def __UpperCamelCase ( self : Dict , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[str, TensorType]] = None , a : ChannelDimension = ChannelDimension.FIRST , **a : Tuple , ) -> PIL.Image.Image: """simple docstring""" SCREAMING_SNAKE_CASE : str = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Union[str, Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : int = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : str = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : Optional[Any] = offset if offset is not None else self.offset SCREAMING_SNAKE_CASE : str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : int = size if size is not None else self.size SCREAMING_SNAKE_CASE : List[Any] = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = get_size_dict(a , param_name="crop_size" ) if not valid_images(a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) SCREAMING_SNAKE_CASE : Optional[int] = make_batched(a ) SCREAMING_SNAKE_CASE : List[Any] = [ [ self._preprocess_image( image=a , do_resize=a , size=a , resample=a , do_center_crop=a , crop_size=a , do_rescale=a , rescale_factor=a , offset=a , do_normalize=a , image_mean=a , image_std=a , data_format=a , ) for img in video ] for video in videos ] SCREAMING_SNAKE_CASE : Optional[int] = {"pixel_values": videos} return BatchFeature(data=a , tensor_type=a )
76
0
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...file_utils import add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward from ...modeling_outputs import ( BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import logging from .configuration_regnet import RegNetConfig __lowerCAmelCase : Tuple = logging.get_logger(__name__) # General docstring __lowerCAmelCase : int = "RegNetConfig" # Base docstring __lowerCAmelCase : Tuple = "facebook/regnet-y-040" __lowerCAmelCase : Tuple = [1, 1_088, 7, 7] # Image classification docstring __lowerCAmelCase : int = "facebook/regnet-y-040" __lowerCAmelCase : Optional[int] = "tabby, tabby cat" __lowerCAmelCase : Union[str, Any] = [ "facebook/regnet-y-040", # See all regnet models at https://huggingface.co/models?filter=regnet ] class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , _snake_case : int , _snake_case : int , _snake_case : int = 3 , _snake_case : int = 1 , _snake_case : int = 1 , _snake_case : Optional[str] = "relu" , ): super().__init__() __lowercase : Optional[Any] = nn.Convad( _snake_case , _snake_case , kernel_size=_snake_case , stride=_snake_case , padding=kernel_size // 2 , groups=_snake_case , bias=_snake_case , ) __lowercase : List[str] = nn.BatchNormad(_snake_case ) __lowercase : int = ACTaFN[activation] if activation is not None else nn.Identity() def snake_case_ ( self : Any , _snake_case : Dict ): __lowercase : Any = self.convolution(_snake_case ) __lowercase : Optional[int] = self.normalization(_snake_case ) __lowercase : List[Any] = self.activation(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Any , _snake_case : RegNetConfig ): super().__init__() __lowercase : Optional[int] = RegNetConvLayer( config.num_channels , config.embedding_size , kernel_size=3 , stride=2 , activation=config.hidden_act ) __lowercase : str = config.num_channels def snake_case_ ( self : List[str] , _snake_case : Tuple ): __lowercase : int = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( '''Make sure that the channel dimension of the pixel values match with the one set in the configuration.''' ) __lowercase : Optional[Any] = self.embedder(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Union[str, Any] , _snake_case : int , _snake_case : int , _snake_case : int = 2 ): super().__init__() __lowercase : Dict = nn.Convad(_snake_case , _snake_case , kernel_size=1 , stride=_snake_case , bias=_snake_case ) __lowercase : Optional[Any] = nn.BatchNormad(_snake_case ) def snake_case_ ( self : str , _snake_case : Tensor ): __lowercase : int = self.convolution(_snake_case ) __lowercase : Optional[Any] = self.normalization(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : str , _snake_case : int , _snake_case : int ): super().__init__() __lowercase : str = nn.AdaptiveAvgPoolad((1, 1) ) __lowercase : Any = nn.Sequential( nn.Convad(_snake_case , _snake_case , kernel_size=1 ) , nn.ReLU() , nn.Convad(_snake_case , _snake_case , kernel_size=1 ) , nn.Sigmoid() , ) def snake_case_ ( self : int , _snake_case : Dict ): __lowercase : Union[str, Any] = self.pooler(_snake_case ) __lowercase : List[Any] = self.attention(_snake_case ) __lowercase : Optional[Any] = hidden_state * attention return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Dict , _snake_case : RegNetConfig , _snake_case : int , _snake_case : int , _snake_case : int = 1 ): super().__init__() __lowercase : Tuple = in_channels != out_channels or stride != 1 __lowercase : Union[str, Any] = max(1 , out_channels // config.groups_width ) __lowercase : Union[str, Any] = ( RegNetShortCut(_snake_case , _snake_case , stride=_snake_case ) if should_apply_shortcut else nn.Identity() ) __lowercase : int = nn.Sequential( RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , stride=_snake_case , groups=_snake_case , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=_snake_case ) , ) __lowercase : Any = ACTaFN[config.hidden_act] def snake_case_ ( self : List[Any] , _snake_case : str ): __lowercase : Union[str, Any] = hidden_state __lowercase : Optional[Any] = self.layer(_snake_case ) __lowercase : List[Any] = self.shortcut(_snake_case ) hidden_state += residual __lowercase : List[str] = self.activation(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[Any] , _snake_case : RegNetConfig , _snake_case : int , _snake_case : int , _snake_case : int = 1 ): super().__init__() __lowercase : List[str] = in_channels != out_channels or stride != 1 __lowercase : Union[str, Any] = max(1 , out_channels // config.groups_width ) __lowercase : Union[str, Any] = ( RegNetShortCut(_snake_case , _snake_case , stride=_snake_case ) if should_apply_shortcut else nn.Identity() ) __lowercase : str = nn.Sequential( RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=config.hidden_act ) , RegNetConvLayer(_snake_case , _snake_case , stride=_snake_case , groups=_snake_case , activation=config.hidden_act ) , RegNetSELayer(_snake_case , reduced_channels=int(round(in_channels / 4 ) ) ) , RegNetConvLayer(_snake_case , _snake_case , kernel_size=1 , activation=_snake_case ) , ) __lowercase : Optional[int] = ACTaFN[config.hidden_act] def snake_case_ ( self : List[str] , _snake_case : List[str] ): __lowercase : str = hidden_state __lowercase : Optional[Any] = self.layer(_snake_case ) __lowercase : Dict = self.shortcut(_snake_case ) hidden_state += residual __lowercase : Dict = self.activation(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[Any] , _snake_case : RegNetConfig , _snake_case : int , _snake_case : int , _snake_case : int = 2 , _snake_case : int = 2 , ): super().__init__() __lowercase : Tuple = RegNetXLayer if config.layer_type == "x" else RegNetYLayer __lowercase : Dict = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer( _snake_case , _snake_case , _snake_case , stride=_snake_case , ) , *[layer(_snake_case , _snake_case , _snake_case ) for _ in range(depth - 1 )] , ) def snake_case_ ( self : Dict , _snake_case : Any ): __lowercase : Tuple = self.layers(_snake_case ) return hidden_state class __lowerCAmelCase ( nn.Module ): """simple docstring""" def __init__( self : Optional[int] , _snake_case : RegNetConfig ): super().__init__() __lowercase : Dict = nn.ModuleList([] ) # based on `downsample_in_first_stage`, the first layer of the first stage may or may not downsample the input self.stages.append( RegNetStage( _snake_case , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) __lowercase : Optional[Any] = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(_snake_case , config.depths[1:] ): self.stages.append(RegNetStage(_snake_case , _snake_case , _snake_case , depth=_snake_case ) ) def snake_case_ ( self : List[str] , _snake_case : Tensor , _snake_case : bool = False , _snake_case : bool = True ): __lowercase : Optional[Any] = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: __lowercase : Any = hidden_states + (hidden_state,) __lowercase : Optional[int] = stage_module(_snake_case ) if output_hidden_states: __lowercase : Any = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention(last_hidden_state=_snake_case , hidden_states=_snake_case ) class __lowerCAmelCase ( __A ): """simple docstring""" A__ : Union[str, Any] = RegNetConfig A__ : Optional[Any] = '''regnet''' A__ : List[str] = '''pixel_values''' A__ : str = True def snake_case_ ( self : Tuple , _snake_case : List[str] ): if isinstance(_snake_case , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode='''fan_out''' , nonlinearity='''relu''' ) elif isinstance(_snake_case , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def snake_case_ ( self : List[Any] , _snake_case : List[Any] , _snake_case : Optional[Any]=False ): if isinstance(_snake_case , _snake_case ): __lowercase : int = value __lowerCAmelCase : Optional[int] = R"\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`RegNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __lowerCAmelCase : Dict = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~file_utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( '''The bare RegNet model outputting raw features without any specific head on top.''' , __A , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetModel with RESNET->REGNET,ResNet->RegNet class __lowerCAmelCase ( __A ): """simple docstring""" def __init__( self : List[str] , _snake_case : Optional[Any] ): super().__init__(_snake_case ) __lowercase : List[str] = config __lowercase : str = RegNetEmbeddings(_snake_case ) __lowercase : List[Any] = RegNetEncoder(_snake_case ) __lowercase : Optional[Any] = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_snake_case ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=_snake_case , config_class=_CONFIG_FOR_DOC , modality='''vision''' , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def snake_case_ ( self : List[Any] , _snake_case : Tensor , _snake_case : Optional[bool] = None , _snake_case : Optional[bool] = None ): __lowercase : List[Any] = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) __lowercase : Dict = return_dict if return_dict is not None else self.config.use_return_dict __lowercase : Any = self.embedder(_snake_case ) __lowercase : Any = self.encoder( _snake_case , output_hidden_states=_snake_case , return_dict=_snake_case ) __lowercase : str = encoder_outputs[0] __lowercase : str = self.pooler(_snake_case ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=_snake_case , pooler_output=_snake_case , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '''\n RegNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ''' , __A , ) # Copied from transformers.models.resnet.modeling_resnet.ResNetForImageClassification with RESNET->REGNET,ResNet->RegNet,resnet->regnet class __lowerCAmelCase ( __A ): """simple docstring""" def __init__( self : str , _snake_case : Optional[int] ): super().__init__(_snake_case ) __lowercase : Optional[Any] = config.num_labels __lowercase : List[Any] = RegNetModel(_snake_case ) # classification head __lowercase : int = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(_snake_case ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=_snake_case , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def snake_case_ ( self : Union[str, Any] , _snake_case : Optional[torch.FloatTensor] = None , _snake_case : Optional[torch.LongTensor] = None , _snake_case : Optional[bool] = None , _snake_case : Optional[bool] = None , ): __lowercase : Union[str, Any] = return_dict if return_dict is not None else self.config.use_return_dict __lowercase : Dict = self.regnet(_snake_case , output_hidden_states=_snake_case , return_dict=_snake_case ) __lowercase : List[Any] = outputs.pooler_output if return_dict else outputs[1] __lowercase : Optional[Any] = self.classifier(_snake_case ) __lowercase : Optional[Any] = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: __lowercase : Union[str, Any] = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): __lowercase : Optional[int] = "single_label_classification" else: __lowercase : Any = "multi_label_classification" if self.config.problem_type == "regression": __lowercase : Tuple = MSELoss() if self.num_labels == 1: __lowercase : List[Any] = loss_fct(logits.squeeze() , labels.squeeze() ) else: __lowercase : str = loss_fct(_snake_case , _snake_case ) elif self.config.problem_type == "single_label_classification": __lowercase : List[str] = CrossEntropyLoss() __lowercase : Any = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": __lowercase : List[str] = BCEWithLogitsLoss() __lowercase : Dict = loss_fct(_snake_case , _snake_case ) if not return_dict: __lowercase : List[Any] = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=_snake_case , logits=_snake_case , hidden_states=outputs.hidden_states )
156
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a_ = logging.get_logger(__name__) a_ = {'vocab_file': 'vocab.txt'} a_ = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a_ = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } a_ = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =ConvBertTokenizer def __init__( self : List[str] , a : Union[str, Any]=None , a : Optional[int]=None , a : int=True , a : Tuple="[UNK]" , a : Dict="[SEP]" , a : Dict="[PAD]" , a : List[Any]="[CLS]" , a : Tuple="[MASK]" , a : Dict=True , a : Optional[Any]=None , **a : str , ) -> Dict: """simple docstring""" super().__init__( a , tokenizer_file=a , do_lower_case=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , tokenize_chinese_chars=a , strip_accents=a , **a , ) SCREAMING_SNAKE_CASE : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , a ) != do_lower_case or normalizer_state.get("strip_accents" , a ) != strip_accents or normalizer_state.get("handle_chinese_chars" , a ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE : List[str] = getattr(a , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE : Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE : Any = strip_accents SCREAMING_SNAKE_CASE : Optional[int] = tokenize_chinese_chars SCREAMING_SNAKE_CASE : List[str] = normalizer_class(**a ) SCREAMING_SNAKE_CASE : str = do_lower_case def __UpperCamelCase ( self : Union[str, Any] , a : List[Any] , a : int=None ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : Tuple , a : str , a : Optional[str] = None ) -> Tuple[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self._tokenizer.model.save(a , name=a ) return tuple(a )
76
0
import unittest import torch from torch import nn from accelerate.test_utils import require_cuda from accelerate.utils.memory import find_executable_batch_size, release_memory def _snake_case( ) -> Optional[Any]: '''simple docstring''' raise RuntimeError('CUDA out of memory.' ) class A ( nn.Module ): """simple docstring""" def __init__( self : List[Any] )-> List[str]: '''simple docstring''' super().__init__() A__ = nn.Linear(3,4 ) A__ = nn.BatchNormad(4 ) A__ = nn.Linear(4,5 ) def snake_case__ ( self : Optional[int],lowercase_ : Optional[Any] )-> Optional[int]: '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(lowercase_ ) ) ) class A ( unittest.TestCase ): """simple docstring""" def snake_case__ ( self : Any )-> str: '''simple docstring''' A__ = [] @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowercase_ : str ): nonlocal batch_sizes batch_sizes.append(lowercase_ ) if batch_size != 8: raise_fake_out_of_memory() mock_training_loop_function() self.assertListEqual(lowercase_,[1_2_8, 6_4, 3_2, 1_6, 8] ) def snake_case__ ( self : str )-> Optional[int]: '''simple docstring''' A__ = [] @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowercase_ : List[Any],lowercase_ : Optional[int] ): nonlocal batch_sizes batch_sizes.append(lowercase_ ) if batch_size != 8: raise_fake_out_of_memory() return batch_size, arga A__ = mock_training_loop_function('hello' ) self.assertListEqual(lowercase_,[1_2_8, 6_4, 3_2, 1_6, 8] ) self.assertListEqual([bs, arga],[8, 'hello'] ) def snake_case__ ( self : Optional[int] )-> int: '''simple docstring''' @find_executable_batch_size(starting_batch_size=0 ) def mock_training_loop_function(lowercase_ : Any ): pass with self.assertRaises(lowercase_ ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.',cm.exception.args[0] ) def snake_case__ ( self : List[str] )-> Tuple: '''simple docstring''' @find_executable_batch_size(starting_batch_size=1_6 ) def mock_training_loop_function(lowercase_ : Any ): if batch_size > 0: raise_fake_out_of_memory() pass with self.assertRaises(lowercase_ ) as cm: mock_training_loop_function() self.assertIn('No executable batch size found, reached zero.',cm.exception.args[0] ) def snake_case__ ( self : List[Any] )-> Tuple: '''simple docstring''' @find_executable_batch_size(starting_batch_size=1_2_8 ) def mock_training_loop_function(lowercase_ : List[str],lowercase_ : List[str],lowercase_ : Any ): if batch_size != 8: raise raise_fake_out_of_memory() with self.assertRaises(lowercase_ ) as cm: mock_training_loop_function(1_2_8,'hello','world' ) self.assertIn('Batch size was passed into `f`',cm.exception.args[0] ) self.assertIn('`f(arg1=\'hello\', arg2=\'world\')',cm.exception.args[0] ) def snake_case__ ( self : Union[str, Any] )-> Union[str, Any]: '''simple docstring''' @find_executable_batch_size(starting_batch_size=1_6 ) def mock_training_loop_function(lowercase_ : List[Any] ): raise ValueError('Oops, we had an error!' ) with self.assertRaises(lowercase_ ) as cm: mock_training_loop_function() self.assertIn('Oops, we had an error!',cm.exception.args[0] ) @require_cuda def snake_case__ ( self : Any )-> int: '''simple docstring''' A__ = torch.cuda.memory_allocated() A__ = ModelForTest() model.cuda() self.assertGreater(torch.cuda.memory_allocated(),lowercase_ ) A__ = release_memory(lowercase_ ) self.assertEqual(torch.cuda.memory_allocated(),lowercase_ )
7
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. a_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_a) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE : Union[str, Any] = terminalreporter.config.getoption("--make-reports") if make_reports: pytest_terminal_summary_main(_a , id=_a)
76
0
from __future__ import annotations import math import numpy as np from numpy.linalg import norm def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' return math.sqrt(sum(pow(a - b , 2 ) for a, b in zip(_a , _a ) ) ) def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' if dataset.ndim != value_array.ndim: UpperCAmelCase = ( "Wrong input data's dimensions... " F"""dataset : {dataset.ndim}, value_array : {value_array.ndim}""" ) raise ValueError(_a ) try: if dataset.shape[1] != value_array.shape[1]: UpperCAmelCase = ( "Wrong input data's shape... " F"""dataset : {dataset.shape[1]}, value_array : {value_array.shape[1]}""" ) raise ValueError(_a ) except IndexError: if dataset.ndim != value_array.ndim: raise TypeError('''Wrong shape''' ) if dataset.dtype != value_array.dtype: UpperCAmelCase = ( "Input data have different datatype... " F"""dataset : {dataset.dtype}, value_array : {value_array.dtype}""" ) raise TypeError(_a ) UpperCAmelCase = [] for value in value_array: UpperCAmelCase = euclidean(_a , dataset[0] ) UpperCAmelCase = dataset[0].tolist() for dataset_value in dataset[1:]: UpperCAmelCase = euclidean(_a , _a ) if dist > temp_dist: UpperCAmelCase = temp_dist UpperCAmelCase = dataset_value.tolist() answer.append([vector, dist] ) return answer def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ , UpperCamelCase__ ) -> Optional[Any]: '''simple docstring''' return np.dot(_a , _a ) / (norm(_a ) * norm(_a )) if __name__ == "__main__": import doctest doctest.testmod()
273
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , a : int , a : Optional[int]=13 , a : Optional[int]=3 , a : int=224 , a : Optional[int]=30 , a : int=400 , a : Union[str, Any]=True , a : int=None , a : Tuple=True , a : Tuple=[0.5, 0.5, 0.5] , a : Optional[int]=[0.5, 0.5, 0.5] , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = size if size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE : Union[str, Any] = parent SCREAMING_SNAKE_CASE : int = batch_size SCREAMING_SNAKE_CASE : int = num_channels SCREAMING_SNAKE_CASE : Any = image_size SCREAMING_SNAKE_CASE : Tuple = min_resolution SCREAMING_SNAKE_CASE : str = max_resolution SCREAMING_SNAKE_CASE : int = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : int = do_normalize SCREAMING_SNAKE_CASE : Tuple = image_mean SCREAMING_SNAKE_CASE : Tuple = image_std def __UpperCamelCase ( self : Any ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCamelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =ViTImageProcessor if is_vision_available() else None def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = EfficientFormerImageProcessorTester(self ) @property def __UpperCamelCase ( self : Any ) -> List[str]: """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def __UpperCamelCase ( self : List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , "image_mean" ) ) self.assertTrue(hasattr(a , "image_std" ) ) self.assertTrue(hasattr(a , "do_normalize" ) ) self.assertTrue(hasattr(a , "do_resize" ) ) self.assertTrue(hasattr(a , "size" ) ) def __UpperCamelCase ( self : int ) -> str: """simple docstring""" pass def __UpperCamelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE : List[str] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE : int = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Any = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
76
0
import math import sys def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : Tuple = "" try: with open(_a, """rb""" ) as binary_file: __magic_name__ : str = binary_file.read() for dat in data: __magic_name__ : Tuple = f'{dat:08b}' result += curr_byte return result except OSError: print("""File not accessible""" ) sys.exit() def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : Dict = {"0": "0", "1": "1"} __magic_name__ : List[Any] = "", "" __magic_name__ : Any = len(_a ) for i in range(len(_a ) ): curr_string += data_bits[i] if curr_string not in lexicon: continue __magic_name__ : List[str] = lexicon[curr_string] result += last_match_id __magic_name__ : Dict = last_match_id + "0" if math.loga(_a ).is_integer(): __magic_name__ : Dict = {} for curr_key in list(_a ): __magic_name__ : Optional[Any] = lexicon.pop(_a ) __magic_name__ : Union[str, Any] = new_lex __magic_name__ : Dict = last_match_id + "1" index += 1 __magic_name__ : List[Any] = "" return result def UpperCamelCase ( _A, _A ): """simple docstring""" __magic_name__ : List[str] = 8 try: with open(_a, """wb""" ) as opened_file: __magic_name__ : str = [ to_write[i : i + byte_length] for i in range(0, len(_a ), _a ) ] if len(result_byte_array[-1] ) % byte_length == 0: result_byte_array.append("""10000000""" ) else: result_byte_array[-1] += "1" + "0" * ( byte_length - len(result_byte_array[-1] ) - 1 ) for elem in result_byte_array[:-1]: opened_file.write(int(_a, 2 ).to_bytes(1, byteorder="""big""" ) ) except OSError: print("""File not accessible""" ) sys.exit() def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : int = 0 for letter in data_bits: if letter == "1": break counter += 1 __magic_name__ : Union[str, Any] = data_bits[counter:] __magic_name__ : Dict = data_bits[counter + 1 :] return data_bits def UpperCamelCase ( _A, _A ): """simple docstring""" __magic_name__ : Optional[Any] = read_file_binary(_a ) __magic_name__ : int = remove_prefix(_a ) __magic_name__ : Union[str, Any] = decompress_data(_a ) write_file_binary(_a, _a ) if __name__ == "__main__": compress(sys.argv[1], sys.argv[2])
342
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Any = tokenizer(example["content"] , truncation=_a)["input_ids"] SCREAMING_SNAKE_CASE : Dict = len(example["content"]) / len(output["input_ids"]) return output a_ = HfArgumentParser(PretokenizationArguments) a_ = parser.parse_args() if args.num_workers is None: a_ = multiprocessing.cpu_count() a_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) a_ = time.time() a_ = load_dataset(args.dataset_name, split='train') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a_ = time.time() a_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ 'repo_name', 'path', 'copies', 'size', 'content', 'license', 'hash', 'line_mean', 'line_max', 'alpha_frac', 'autogenerated', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
76
0
'''simple docstring''' import unittest import numpy as np from transformers import RoFormerConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.roformer.modeling_flax_roformer import ( FlaxRoFormerForMaskedLM, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerModel, ) class lowercase_ ( unittest.TestCase ): def __init__( self , a , a=13 , a=7 , a=True , a=True , a=True , a=True , a=99 , a=32 , a=5 , a=4 , a=37 , a="gelu" , a=0.1 , a=0.1 , a=5_12 , a=16 , a=2 , a=0.02 , a=4 , ): UpperCamelCase__ = parent UpperCamelCase__ = batch_size UpperCamelCase__ = seq_length UpperCamelCase__ = is_training UpperCamelCase__ = use_attention_mask UpperCamelCase__ = use_token_type_ids UpperCamelCase__ = use_labels UpperCamelCase__ = vocab_size UpperCamelCase__ = hidden_size UpperCamelCase__ = num_hidden_layers UpperCamelCase__ = num_attention_heads UpperCamelCase__ = intermediate_size UpperCamelCase__ = hidden_act UpperCamelCase__ = hidden_dropout_prob UpperCamelCase__ = attention_probs_dropout_prob UpperCamelCase__ = max_position_embeddings UpperCamelCase__ = type_vocab_size UpperCamelCase__ = type_sequence_label_size UpperCamelCase__ = initializer_range UpperCamelCase__ = num_choices def __a ( self ): UpperCamelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCamelCase__ = None if self.use_attention_mask: UpperCamelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCamelCase__ = None if self.use_token_type_ids: UpperCamelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCamelCase__ = RoFormerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=a , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def __a ( self ): UpperCamelCase__ = self.prepare_config_and_inputs() UpperCamelCase__ = config_and_inputs UpperCamelCase__ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class lowercase_ ( __A , unittest.TestCase ): __UpperCAmelCase = True __UpperCAmelCase = ( ( FlaxRoFormerModel, FlaxRoFormerForMaskedLM, FlaxRoFormerForSequenceClassification, FlaxRoFormerForTokenClassification, FlaxRoFormerForMultipleChoice, FlaxRoFormerForQuestionAnswering, ) if is_flax_available() else () ) def __a ( self ): UpperCamelCase__ = FlaxRoFormerModelTester(self ) @slow def __a ( self ): for model_class_name in self.all_model_classes: UpperCamelCase__ = model_class_name.from_pretrained("junnyu/roformer_chinese_small" , from_pt=a ) UpperCamelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(a ) @require_flax class lowercase_ ( unittest.TestCase ): @slow def __a ( self ): UpperCamelCase__ = FlaxRoFormerForMaskedLM.from_pretrained("junnyu/roformer_chinese_base" ) UpperCamelCase__ = jnp.array([[0, 1, 2, 3, 4, 5]] ) UpperCamelCase__ = model(a )[0] UpperCamelCase__ = 5_00_00 UpperCamelCase__ = (1, 6, vocab_size) self.assertEqual(output.shape , a ) UpperCamelCase__ = jnp.array( [[[-0.1205, -1.0265, 0.2922], [-1.5134, 0.1974, 0.1519], [-5.0135, -3.9003, -0.8404]]] ) self.assertTrue(jnp.allclose(output[:, :3, :3] , a , atol=1e-4 ) )
80
import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DetrConfig, DetrForObjectDetection, DetrForSegmentation, DetrImageProcessor, ResNetConfig from transformers.utils import logging logging.set_verbosity_info() a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): # initialize config if "resnet-50" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-50") elif "resnet-101" in model_name: SCREAMING_SNAKE_CASE : int = ResNetConfig.from_pretrained("microsoft/resnet-101") else: raise ValueError("Model name should include either resnet50 or resnet101") SCREAMING_SNAKE_CASE : str = DetrConfig(use_timm_backbone=_a , backbone_config=_a) # set label attributes SCREAMING_SNAKE_CASE : List[str] = "panoptic" in model_name if is_panoptic: SCREAMING_SNAKE_CASE : Union[str, Any] = 250 else: SCREAMING_SNAKE_CASE : Union[str, Any] = 91 SCREAMING_SNAKE_CASE : str = "huggingface/label-files" SCREAMING_SNAKE_CASE : Union[str, Any] = "coco-detection-id2label.json" SCREAMING_SNAKE_CASE : Optional[Any] = json.load(open(hf_hub_download(_a , _a , repo_type="dataset") , "r")) SCREAMING_SNAKE_CASE : int = {int(_a): v for k, v in idalabel.items()} SCREAMING_SNAKE_CASE : List[Any] = idalabel SCREAMING_SNAKE_CASE : List[Any] = {v: k for k, v in idalabel.items()} return config, is_panoptic def lowerCamelCase__ ( _a): # here we list all keys to be renamed (original name on the left, our name on the right) SCREAMING_SNAKE_CASE : Union[str, Any] = [] # stem # fmt: off rename_keys.append(("backbone.0.body.conv1.weight", "backbone.conv_encoder.model.embedder.embedder.convolution.weight")) rename_keys.append(("backbone.0.body.bn1.weight", "backbone.conv_encoder.model.embedder.embedder.normalization.weight")) rename_keys.append(("backbone.0.body.bn1.bias", "backbone.conv_encoder.model.embedder.embedder.normalization.bias")) rename_keys.append(("backbone.0.body.bn1.running_mean", "backbone.conv_encoder.model.embedder.embedder.normalization.running_mean")) rename_keys.append(("backbone.0.body.bn1.running_var", "backbone.conv_encoder.model.embedder.embedder.normalization.running_var")) # stages for stage_idx in range(len(config.backbone_config.depths)): for layer_idx in range(config.backbone_config.depths[stage_idx]): # shortcut if layer_idx == 0: rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.0.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.downsample.1.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.shortcut.normalization.running_var", )) # 3 convs for i in range(3): rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.conv{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.convolution.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.weight", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.weight", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.bias", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.bias", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_mean", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_mean", )) rename_keys.append( ( f"backbone.0.body.layer{stage_idx + 1}.{layer_idx}.bn{i+1}.running_var", f"backbone.conv_encoder.model.encoder.stages.{stage_idx}.layers.{layer_idx}.layer.{i}.normalization.running_var", )) # fmt: on for i in range(config.encoder_layers): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( ( f"transformer.encoder.layers.{i}.self_attn.out_proj.weight", f"encoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.encoder.layers.{i}.self_attn.out_proj.bias", f"encoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.weight", f"encoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear1.bias", f"encoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.weight", f"encoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.linear2.bias", f"encoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.weight", f"encoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm1.bias", f"encoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.encoder.layers.{i}.norm2.weight", f"encoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.encoder.layers.{i}.norm2.bias", f"encoder.layers.{i}.final_layer_norm.bias")) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( ( f"transformer.decoder.layers.{i}.self_attn.out_proj.weight", f"decoder.layers.{i}.self_attn.out_proj.weight", )) rename_keys.append( (f"transformer.decoder.layers.{i}.self_attn.out_proj.bias", f"decoder.layers.{i}.self_attn.out_proj.bias")) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.weight", f"decoder.layers.{i}.encoder_attn.out_proj.weight", )) rename_keys.append( ( f"transformer.decoder.layers.{i}.multihead_attn.out_proj.bias", f"decoder.layers.{i}.encoder_attn.out_proj.bias", )) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.weight", f"decoder.layers.{i}.fc1.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear1.bias", f"decoder.layers.{i}.fc1.bias")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.weight", f"decoder.layers.{i}.fc2.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.linear2.bias", f"decoder.layers.{i}.fc2.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.weight", f"decoder.layers.{i}.self_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm1.bias", f"decoder.layers.{i}.self_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.weight", f"decoder.layers.{i}.encoder_attn_layer_norm.weight")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm2.bias", f"decoder.layers.{i}.encoder_attn_layer_norm.bias")) rename_keys.append( (f"transformer.decoder.layers.{i}.norm3.weight", f"decoder.layers.{i}.final_layer_norm.weight")) rename_keys.append((f"transformer.decoder.layers.{i}.norm3.bias", f"decoder.layers.{i}.final_layer_norm.bias")) # convolutional projection + query embeddings + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("input_proj.weight", "input_projection.weight"), ("input_proj.bias", "input_projection.bias"), ("query_embed.weight", "query_position_embeddings.weight"), ("transformer.decoder.norm.weight", "decoder.layernorm.weight"), ("transformer.decoder.norm.bias", "decoder.layernorm.bias"), ("class_embed.weight", "class_labels_classifier.weight"), ("class_embed.bias", "class_labels_classifier.bias"), ("bbox_embed.layers.0.weight", "bbox_predictor.layers.0.weight"), ("bbox_embed.layers.0.bias", "bbox_predictor.layers.0.bias"), ("bbox_embed.layers.1.weight", "bbox_predictor.layers.1.weight"), ("bbox_embed.layers.1.bias", "bbox_predictor.layers.1.bias"), ("bbox_embed.layers.2.weight", "bbox_predictor.layers.2.weight"), ("bbox_embed.layers.2.bias", "bbox_predictor.layers.2.bias"), ]) return rename_keys def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : str = state_dict.pop(_a) SCREAMING_SNAKE_CASE : int = val def lowerCamelCase__ ( _a , _a=False): SCREAMING_SNAKE_CASE : Optional[Any] = "" if is_panoptic: SCREAMING_SNAKE_CASE : Optional[int] = "detr." # first: transformer encoder for i in range(6): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(f"{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : int = in_proj_bias[:256] SCREAMING_SNAKE_CASE : Tuple = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : List[Any] = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : str = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6): # read in weights + bias of input projection layer of self-attention SCREAMING_SNAKE_CASE : List[str] = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight") SCREAMING_SNAKE_CASE : str = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias") # next, add query, keys and values (in that order) to the state dict SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_weight[:256, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias[:256] SCREAMING_SNAKE_CASE : List[Any] = in_proj_weight[256:512, :] SCREAMING_SNAKE_CASE : Any = in_proj_bias[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop( f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight") SCREAMING_SNAKE_CASE : int = state_dict.pop(f"{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias") # next, add query, keys and values (in that order) of cross-attention to the state dict SCREAMING_SNAKE_CASE : Tuple = in_proj_weight_cross_attn[:256, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[:256] SCREAMING_SNAKE_CASE : Optional[Any] = in_proj_weight_cross_attn[256:512, :] SCREAMING_SNAKE_CASE : Dict = in_proj_bias_cross_attn[256:512] SCREAMING_SNAKE_CASE : Optional[int] = in_proj_weight_cross_attn[-256:, :] SCREAMING_SNAKE_CASE : Union[str, Any] = in_proj_bias_cross_attn[-256:] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" SCREAMING_SNAKE_CASE : Union[str, Any] = Image.open(requests.get(_a , stream=_a).raw) return im @torch.no_grad() def lowerCamelCase__ ( _a , _a=None , _a=False): SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Optional[int] = get_detr_config(_a) # load original model from torch hub SCREAMING_SNAKE_CASE : Union[str, Any] = { "detr-resnet-50": "detr_resnet50", "detr-resnet-101": "detr_resnet101", } logger.info(f"Converting model {model_name}...") SCREAMING_SNAKE_CASE : Optional[int] = torch.hub.load("facebookresearch/detr" , model_name_to_original_name[model_name] , pretrained=_a).eval() SCREAMING_SNAKE_CASE : Tuple = detr.state_dict() # rename keys for src, dest in create_rename_keys(_a): if is_panoptic: SCREAMING_SNAKE_CASE : List[str] = "detr." + src rename_key(_a , _a , _a) # query, key and value matrices need special treatment read_in_q_k_v(_a , is_panoptic=_a) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them SCREAMING_SNAKE_CASE : List[Any] = "detr.model." if is_panoptic else "model." for key in state_dict.copy().keys(): if is_panoptic: if ( key.startswith("detr") and not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor") ): SCREAMING_SNAKE_CASE : Optional[int] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Union[str, Any] = val elif "class_labels_classifier" in key or "bbox_predictor" in key: SCREAMING_SNAKE_CASE : Union[str, Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Optional[int] = val elif key.startswith("bbox_attention") or key.startswith("mask_head"): continue else: SCREAMING_SNAKE_CASE : Optional[Any] = state_dict.pop(_a) SCREAMING_SNAKE_CASE : List[Any] = val else: if not key.startswith("class_labels_classifier") and not key.startswith("bbox_predictor"): SCREAMING_SNAKE_CASE : Any = state_dict.pop(_a) SCREAMING_SNAKE_CASE : Any = val # finally, create HuggingFace model and load state dict SCREAMING_SNAKE_CASE : int = DetrForSegmentation(_a) if is_panoptic else DetrForObjectDetection(_a) model.load_state_dict(_a) model.eval() # verify our conversion on an image SCREAMING_SNAKE_CASE : int = "coco_panoptic" if is_panoptic else "coco_detection" SCREAMING_SNAKE_CASE : Optional[int] = DetrImageProcessor(format=_a) SCREAMING_SNAKE_CASE : List[str] = processor(images=prepare_img() , return_tensors="pt") SCREAMING_SNAKE_CASE : Any = encoding["pixel_values"] SCREAMING_SNAKE_CASE : Optional[Any] = detr(_a) SCREAMING_SNAKE_CASE : Any = model(_a) assert torch.allclose(outputs.logits , original_outputs["pred_logits"] , atol=1E-3) assert torch.allclose(outputs.pred_boxes , original_outputs["pred_boxes"] , atol=1E-3) if is_panoptic: assert torch.allclose(outputs.pred_masks , original_outputs["pred_masks"] , atol=1E-4) print("Looks ok!") if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f"Saving PyTorch model and image processor to {pytorch_dump_folder_path}...") Path(_a).mkdir(exist_ok=_a) model.save_pretrained(_a) processor.save_pretrained(_a) if push_to_hub: # Upload model and image processor to the hub logger.info("Uploading PyTorch model and image processor to the hub...") model.push_to_hub(f"nielsr/{model_name}") processor.push_to_hub(f"nielsr/{model_name}") if __name__ == "__main__": a_ = argparse.ArgumentParser() parser.add_argument( '--model_name', default='detr-resnet-50', type=str, choices=['detr-resnet-50', 'detr-resnet-101'], help='Name of the DETR model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the folder to output PyTorch model.' ) parser.add_argument('--push_to_hub', action='store_true', help='Whether to push the model to the hub or not.') a_ = parser.parse_args() convert_detr_checkpoint(args.model_name, args.pytorch_dump_folder_path, args.push_to_hub)
76
0
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def _A ( SCREAMING_SNAKE_CASE__ : Tuple , SCREAMING_SNAKE_CASE__ : Optional[Any] ): # Load checkpoint UpperCamelCase :int = torch.load(_a , map_location='''cpu''' ) UpperCamelCase :Dict = chkpt["model"] # We have the base model one level deeper than the original XLM repository UpperCamelCase :Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: UpperCamelCase :List[str] = v else: UpperCamelCase :int = v UpperCamelCase :int = chkpt["params"] UpperCamelCase :Union[str, Any] = {n: v for n, v in config.items() if not isinstance(_a , (torch.FloatTensor, numpy.ndarray) )} UpperCamelCase :List[Any] = chkpt["dico_word2id"] UpperCamelCase :List[Any] = {s + "</w>" if s.find('''@@''' ) == -1 and i > 13 else s.replace('''@@''' , '''''' ): i for s, i in vocab.items()} # Save pytorch-model UpperCamelCase :Tuple = pytorch_dump_folder_path + "/" + WEIGHTS_NAME UpperCamelCase :Any = pytorch_dump_folder_path + "/" + CONFIG_NAME UpperCamelCase :Optional[int] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(F'''Save PyTorch model to {pytorch_weights_dump_path}''' ) torch.save(_a , _a ) print(F'''Save configuration file to {pytorch_config_dump_path}''' ) with open(_a , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(_a , indent=2 ) + '''\n''' ) print(F'''Save vocab file to {pytorch_config_dump_path}''' ) with open(_a , '''w''' , encoding='''utf-8''' ) as f: f.write(json.dumps(_a , indent=2 ) + '''\n''' ) if __name__ == "__main__": __snake_case = argparse.ArgumentParser() # Required parameters parser.add_argument( """--xlm_checkpoint_path""", default=None, type=str, required=True, help="""Path the official PyTorch dump.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) __snake_case = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
259
import os def lowerCamelCase__ ( ): with open(os.path.dirname(_a) + "/p022_names.txt") as file: SCREAMING_SNAKE_CASE : List[str] = str(file.readlines()[0]) SCREAMING_SNAKE_CASE : List[Any] = names.replace("\"" , "").split(",") names.sort() SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Dict = 0 for i, name in enumerate(_a): for letter in name: name_score += ord(_a) - 64 total_score += (i + 1) * name_score SCREAMING_SNAKE_CASE : str = 0 return total_score if __name__ == "__main__": print(solution())
76
0
import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _SCREAMING_SNAKE_CASE ( __A ): def __init__( self , lowercase , lowercase , lowercase ) -> List[str]: lowerCamelCase_ = dataset lowerCamelCase_ = process lowerCamelCase_ = params def __len__( self ) -> int: return len(self.dataset ) def __getitem__( self , lowercase ) -> Union[str, Any]: lowerCamelCase_ = self.dataset[i] lowerCamelCase_ = self.process(lowercase , **self.params ) return processed class _SCREAMING_SNAKE_CASE ( __A ): def __init__( self , lowercase , lowercase , lowercase , lowercase=None ) -> Optional[Any]: lowerCamelCase_ = loader lowerCamelCase_ = infer lowerCamelCase_ = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether lowerCamelCase_ = None lowerCamelCase_ = loader_batch_size # Internal bookkeeping lowerCamelCase_ = None lowerCamelCase_ = None def __len__( self ) -> List[str]: return len(self.loader ) def __iter__( self ) -> Union[str, Any]: lowerCamelCase_ = iter(self.loader ) return self def SCREAMING_SNAKE_CASE_( self ) -> List[str]: if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice lowerCamelCase_ = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) lowerCamelCase_ = {} for k, element in self._loader_batch_data.items(): if isinstance(lowercase , lowercase ): # Convert ModelOutput to tuple first lowerCamelCase_ = element.to_tuple() if isinstance(element[0] , torch.Tensor ): lowerCamelCase_ = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): lowerCamelCase_ = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(lowercase , lowercase ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): lowerCamelCase_ = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): lowerCamelCase_ = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around lowerCamelCase_ = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers lowerCamelCase_ = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers lowerCamelCase_ = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. lowerCamelCase_ = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 lowerCamelCase_ = self._loader_batch_data.__class__(lowercase ) self._loader_batch_index += 1 return result def SCREAMING_SNAKE_CASE_( self ) -> Dict: if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch lowerCamelCase_ = next(self.iterator ) lowerCamelCase_ = self.infer(lowercase , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(lowercase , torch.Tensor ): lowerCamelCase_ = processed else: lowerCamelCase_ = list(processed.keys() )[0] lowerCamelCase_ = processed[key] if isinstance(lowercase , lowercase ): lowerCamelCase_ = len(lowercase ) else: lowerCamelCase_ = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. lowerCamelCase_ = observed_batch_size # Setting internal index to unwrap the batch lowerCamelCase_ = processed lowerCamelCase_ = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _SCREAMING_SNAKE_CASE ( __A ): def __init__( self , lowercase , lowercase , lowercase , lowercase=None ) -> Dict: super().__init__(lowercase , lowercase , lowercase ) def __iter__( self ) -> Optional[Any]: lowerCamelCase_ = iter(self.loader ) lowerCamelCase_ = None return self def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: if self.subiterator is None: lowerCamelCase_ = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item lowerCamelCase_ = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators lowerCamelCase_ = self.infer(next(self.iterator ) , **self.params ) lowerCamelCase_ = next(self.subiterator ) return processed class _SCREAMING_SNAKE_CASE ( __A ): def __iter__( self ) -> Optional[int]: lowerCamelCase_ = iter(self.loader ) return self def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = False lowerCamelCase_ = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: lowerCamelCase_ = self.loader_batch_item() lowerCamelCase_ = item.pop("is_last" ) accumulator.append(lowercase ) if is_last: return accumulator while not is_last: lowerCamelCase_ = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(lowercase , torch.Tensor ): lowerCamelCase_ = processed else: lowerCamelCase_ = list(processed.keys() )[0] lowerCamelCase_ = processed[key] if isinstance(lowercase , lowercase ): lowerCamelCase_ = len(lowercase ) else: lowerCamelCase_ = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. lowerCamelCase_ = observed_batch_size lowerCamelCase_ = processed lowerCamelCase_ = 0 while self._loader_batch_index < self.loader_batch_size: lowerCamelCase_ = self.loader_batch_item() lowerCamelCase_ = item.pop("is_last" ) accumulator.append(lowercase ) if is_last: return accumulator else: lowerCamelCase_ = processed lowerCamelCase_ = item.pop("is_last" ) accumulator.append(lowercase ) return accumulator class _SCREAMING_SNAKE_CASE ( __A ): def __init__( self , lowercase , lowercase ) -> Tuple: lowerCamelCase_ = dataset lowerCamelCase_ = key def __len__( self ) -> Dict: return len(self.dataset ) def __getitem__( self , lowercase ) -> List[str]: return self.dataset[i][self.key] class _SCREAMING_SNAKE_CASE ( __A ): def __init__( self , lowercase , lowercase , lowercase ) -> Optional[Any]: lowerCamelCase_ = dataset lowerCamelCase_ = keya lowerCamelCase_ = keya def __len__( self ) -> Any: return len(self.dataset ) def __getitem__( self , lowercase ) -> List[str]: return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
19
from collections.abc import Callable import numpy as np def lowerCamelCase__ ( _a , _a , _a , _a , _a): SCREAMING_SNAKE_CASE : Dict = int(np.ceil((x_end - xa) / step_size)) SCREAMING_SNAKE_CASE : Tuple = np.zeros((n + 1,)) SCREAMING_SNAKE_CASE : int = ya SCREAMING_SNAKE_CASE : int = xa for k in range(_a): SCREAMING_SNAKE_CASE : Any = y[k] + step_size * ode_func(_a , y[k]) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
76
0
lowerCAmelCase__ : Dict = { '''A''': ['''B''', '''C''', '''E'''], '''B''': ['''A''', '''D''', '''E'''], '''C''': ['''A''', '''F''', '''G'''], '''D''': ['''B'''], '''E''': ['''A''', '''B''', '''D'''], '''F''': ['''C'''], '''G''': ['''C'''], } def UpperCamelCase__ ( A__ , A__ , A__ ) -> str: snake_case__ : int = set() # keep track of all the paths to be checked snake_case__ : int = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue snake_case__ : Optional[int] = queue.pop(0 ) # get the last node from the path snake_case__ : Union[str, Any] = path[-1] if node not in explored: snake_case__ : List[str] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: snake_case__ : List[Any] = list(_a ) new_path.append(_a ) queue.append(_a ) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_a ) # in case there's no path between the 2 nodes return [] def UpperCamelCase__ ( A__ , A__ , A__ ) -> Optional[int]: if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 snake_case__ : str = [start] snake_case__ : Optional[Any] = set(_a ) # Keep tab on distances from `start` node. snake_case__ : Union[str, Any] = {start: 0, target: -1} while queue: snake_case__ : Optional[int] = queue.pop(0 ) if node == target: snake_case__ : Union[str, Any] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node] ) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_a ) queue.append(_a ) snake_case__ : Optional[Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, '''G''', '''D''')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, '''G''', '''D''')) # returns 4
143
def lowerCamelCase__ ( _a , _a): return int((input_a, input_a).count(1) != 0) def lowerCamelCase__ ( ): assert or_gate(0 , 0) == 0 assert or_gate(0 , 1) == 1 assert or_gate(1 , 0) == 1 assert or_gate(1 , 1) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
76
0
import unittest import numpy as np from transformers.testing_utils import require_flax, require_tf, require_torch from transformers.utils import ( expand_dims, flatten_dict, is_flax_available, is_tf_available, is_torch_available, reshape, squeeze, transpose, ) if is_flax_available(): import jax.numpy as jnp if is_tf_available(): import tensorflow as tf if is_torch_available(): import torch class __A ( unittest.TestCase ): """simple docstring""" def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] ={ "task_specific_params": { "summarization": {"length_penalty": 1.0, "max_length": 128, "min_length": 12, "num_beams": 4}, "summarization_cnn": {"length_penalty": 2.0, "max_length": 142, "min_length": 56, "num_beams": 4}, "summarization_xsum": {"length_penalty": 1.0, "max_length": 62, "min_length": 11, "num_beams": 6}, } } __UpperCamelCase : int ={ "task_specific_params.summarization.length_penalty": 1.0, "task_specific_params.summarization.max_length": 128, "task_specific_params.summarization.min_length": 12, "task_specific_params.summarization.num_beams": 4, "task_specific_params.summarization_cnn.length_penalty": 2.0, "task_specific_params.summarization_cnn.max_length": 142, "task_specific_params.summarization_cnn.min_length": 56, "task_specific_params.summarization_cnn.num_beams": 4, "task_specific_params.summarization_xsum.length_penalty": 1.0, "task_specific_params.summarization_xsum.max_length": 62, "task_specific_params.summarization_xsum.min_length": 11, "task_specific_params.summarization_xsum.num_beams": 6, } self.assertEqual(flatten_dict(lowerCamelCase__ ) , lowerCamelCase__ ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =np.random.randn(3 , 4 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , x.transpose() ) ) __UpperCamelCase : str =np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , x.transpose((1, 2, 0) ) ) ) @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =np.random.randn(3 , 4 ) __UpperCamelCase : Optional[Any] =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : Optional[int] =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Optional[Any] =np.random.randn(3 , 4 ) __UpperCamelCase : Tuple =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , transpose(lowerCamelCase__ ).numpy() ) ) __UpperCamelCase : int =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : Dict =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , transpose(lowerCamelCase__ , axes=(1, 2, 0) ).numpy() ) ) @require_flax def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 ) __UpperCamelCase : List[Any] =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ ) , np.asarray(transpose(lowerCamelCase__ ) ) ) ) __UpperCamelCase : List[Any] =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : Any =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) , np.asarray(transpose(lowerCamelCase__ , axes=(1, 2, 0) ) ) ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =np.random.randn(3 , 4 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.reshape(lowerCamelCase__ , (4, 3) ) ) ) __UpperCamelCase : Tuple =np.random.randn(3 , 4 , 5 ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , np.reshape(lowerCamelCase__ , (12, 5) ) ) ) @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =np.random.randn(3 , 4 ) __UpperCamelCase : List[str] =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) __UpperCamelCase : Dict =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : Any =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , reshape(lowerCamelCase__ , (12, 5) ).numpy() ) ) @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 ) __UpperCamelCase : Optional[int] =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , reshape(lowerCamelCase__ , (4, 3) ).numpy() ) ) __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : int =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , reshape(lowerCamelCase__ , (12, 5) ).numpy() ) ) @require_flax def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[Any] =np.random.randn(3 , 4 ) __UpperCamelCase : Optional[Any] =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (4, 3) ) , np.asarray(reshape(lowerCamelCase__ , (4, 3) ) ) ) ) __UpperCamelCase : str =np.random.randn(3 , 4 , 5 ) __UpperCamelCase : Union[str, Any] =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(reshape(lowerCamelCase__ , (12, 5) ) , np.asarray(reshape(lowerCamelCase__ , (12, 5) ) ) ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =np.random.randn(1 , 3 , 4 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.squeeze(lowerCamelCase__ ) ) ) __UpperCamelCase : str =np.random.randn(1 , 4 , 1 , 5 ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.squeeze(lowerCamelCase__ , axis=2 ) ) ) @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =np.random.randn(1 , 3 , 4 ) __UpperCamelCase : Union[str, Any] =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) __UpperCamelCase : int =np.random.randn(1 , 4 , 1 , 5 ) __UpperCamelCase : Dict =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =np.random.randn(1 , 3 , 4 ) __UpperCamelCase : str =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , squeeze(lowerCamelCase__ ).numpy() ) ) __UpperCamelCase : List[str] =np.random.randn(1 , 4 , 1 , 5 ) __UpperCamelCase : Any =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , squeeze(lowerCamelCase__ , axis=2 ).numpy() ) ) @require_flax def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Dict =np.random.randn(1 , 3 , 4 ) __UpperCamelCase : List[str] =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ ) , np.asarray(squeeze(lowerCamelCase__ ) ) ) ) __UpperCamelCase : int =np.random.randn(1 , 4 , 1 , 5 ) __UpperCamelCase : str =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(squeeze(lowerCamelCase__ , axis=2 ) , np.asarray(squeeze(lowerCamelCase__ , axis=2 ) ) ) ) def __lowercase ( self ): """simple docstring""" __UpperCamelCase : List[str] =np.random.randn(3 , 4 ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.expand_dims(lowerCamelCase__ , axis=1 ) ) ) @require_torch def __lowercase ( self ): """simple docstring""" __UpperCamelCase : str =np.random.randn(3 , 4 ) __UpperCamelCase : Tuple =torch.tensor(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_tf def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 ) __UpperCamelCase : Dict =tf.constant(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , expand_dims(lowerCamelCase__ , axis=1 ).numpy() ) ) @require_flax def __lowercase ( self ): """simple docstring""" __UpperCamelCase : Union[str, Any] =np.random.randn(3 , 4 ) __UpperCamelCase : Union[str, Any] =jnp.array(lowerCamelCase__ ) self.assertTrue(np.allclose(expand_dims(lowerCamelCase__ , axis=1 ) , np.asarray(expand_dims(lowerCamelCase__ , axis=1 ) ) ) )
71
a_ = 8.314_4598 def lowerCamelCase__ ( _a , _a): if temperature < 0: raise Exception("Temperature cannot be less than 0 K") if molar_mass <= 0: raise Exception("Molar mass cannot be less than or equal to 0 kg/mol") else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example a_ = 300 a_ = 28 a_ = rms_speed_of_molecule(temperature, molar_mass) print(F'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
76
0
"""simple docstring""" import math def lowercase ( ) -> Optional[int]: __a = input('''Enter message: ''' ) __a = int(input(f'''Enter key [2-{len(_a ) - 1}]: ''' ) ) __a = input('''Encryption/Decryption [e/d]: ''' ) if mode.lower().startswith('''e''' ): __a = encrypt_message(_a , _a ) elif mode.lower().startswith('''d''' ): __a = decrypt_message(_a , _a ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f'''Output:\n{text + '|'}''' ) def lowercase ( lowerCAmelCase__ : List[str] , lowerCAmelCase__ : List[str] ) -> Optional[Any]: __a = [""] * key for col in range(_a ): __a = col while pointer < len(_a ): cipher_text[col] += message[pointer] pointer += key return "".join(_a ) def lowercase ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[Any] ) -> List[Any]: __a = math.ceil(len(_a ) / key ) __a = key __a = (num_cols * num_rows) - len(_a ) __a = [""] * num_cols __a = 0 __a = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): __a = 0 row += 1 return "".join(_a ) if __name__ == "__main__": import doctest doctest.testmod() main()
45
a_ = { 'A': ['B', 'C', 'E'], 'B': ['A', 'D', 'E'], 'C': ['A', 'F', 'G'], 'D': ['B'], 'E': ['A', 'B', 'D'], 'F': ['C'], 'G': ['C'], } def lowerCamelCase__ ( _a , _a , _a): SCREAMING_SNAKE_CASE : int = set() # keep track of all the paths to be checked SCREAMING_SNAKE_CASE : int = [[start]] # return path if start is goal if start == goal: return [start] # keeps looping until all possible paths have been checked while queue: # pop the first path from the queue SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) # get the last node from the path SCREAMING_SNAKE_CASE : Union[str, Any] = path[-1] if node not in explored: SCREAMING_SNAKE_CASE : List[str] = graph[node] # go through all neighbour nodes, construct a new path and # push it into the queue for neighbour in neighbours: SCREAMING_SNAKE_CASE : List[Any] = list(_a) new_path.append(_a) queue.append(_a) # return path if neighbour is goal if neighbour == goal: return new_path # mark node as explored explored.add(_a) # in case there's no path between the 2 nodes return [] def lowerCamelCase__ ( _a , _a , _a): if not graph or start not in graph or target not in graph: return -1 if start == target: return 0 SCREAMING_SNAKE_CASE : str = [start] SCREAMING_SNAKE_CASE : Optional[Any] = set(_a) # Keep tab on distances from `start` node. SCREAMING_SNAKE_CASE : Union[str, Any] = {start: 0, target: -1} while queue: SCREAMING_SNAKE_CASE : Optional[int] = queue.pop(0) if node == target: SCREAMING_SNAKE_CASE : Union[str, Any] = ( dist[node] if dist[target] == -1 else min(dist[target] , dist[node]) ) for adjacent in graph[node]: if adjacent not in visited: visited.add(_a) queue.append(_a) SCREAMING_SNAKE_CASE : Optional[Any] = dist[node] + 1 return dist[target] if __name__ == "__main__": print(bfs_shortest_path(demo_graph, 'G', 'D')) # returns ['G', 'C', 'A', 'B', 'D'] print(bfs_shortest_path_distance(demo_graph, 'G', 'D')) # returns 4
76
0
import importlib import json import os from collections import OrderedDict from typing import Dict, Optional, Union # Build the list of all image processors from ...configuration_utils import PretrainedConfig from ...dynamic_module_utils import get_class_from_dynamic_module, resolve_trust_remote_code from ...image_processing_utils import ImageProcessingMixin from ...utils import CONFIG_NAME, IMAGE_PROCESSOR_NAME, get_file_from_repo, logging from .auto_factory import _LazyAutoMapping from .configuration_auto import ( CONFIG_MAPPING_NAMES, AutoConfig, model_type_to_module_name, replace_list_option_in_docstrings, ) UpperCamelCase = logging.get_logger(__name__) UpperCamelCase = OrderedDict( [ ('''align''', '''EfficientNetImageProcessor'''), ('''beit''', '''BeitImageProcessor'''), ('''bit''', '''BitImageProcessor'''), ('''blip''', '''BlipImageProcessor'''), ('''blip-2''', '''BlipImageProcessor'''), ('''bridgetower''', '''BridgeTowerImageProcessor'''), ('''chinese_clip''', '''ChineseCLIPImageProcessor'''), ('''clip''', '''CLIPImageProcessor'''), ('''clipseg''', '''ViTImageProcessor'''), ('''conditional_detr''', '''ConditionalDetrImageProcessor'''), ('''convnext''', '''ConvNextImageProcessor'''), ('''convnextv2''', '''ConvNextImageProcessor'''), ('''cvt''', '''ConvNextImageProcessor'''), ('''data2vec-vision''', '''BeitImageProcessor'''), ('''deformable_detr''', '''DeformableDetrImageProcessor'''), ('''deit''', '''DeiTImageProcessor'''), ('''deta''', '''DetaImageProcessor'''), ('''detr''', '''DetrImageProcessor'''), ('''dinat''', '''ViTImageProcessor'''), ('''donut-swin''', '''DonutImageProcessor'''), ('''dpt''', '''DPTImageProcessor'''), ('''efficientformer''', '''EfficientFormerImageProcessor'''), ('''efficientnet''', '''EfficientNetImageProcessor'''), ('''flava''', '''FlavaImageProcessor'''), ('''focalnet''', '''BitImageProcessor'''), ('''git''', '''CLIPImageProcessor'''), ('''glpn''', '''GLPNImageProcessor'''), ('''groupvit''', '''CLIPImageProcessor'''), ('''imagegpt''', '''ImageGPTImageProcessor'''), ('''instructblip''', '''BlipImageProcessor'''), ('''layoutlmv2''', '''LayoutLMv2ImageProcessor'''), ('''layoutlmv3''', '''LayoutLMv3ImageProcessor'''), ('''levit''', '''LevitImageProcessor'''), ('''mask2former''', '''Mask2FormerImageProcessor'''), ('''maskformer''', '''MaskFormerImageProcessor'''), ('''mgp-str''', '''ViTImageProcessor'''), ('''mobilenet_v1''', '''MobileNetV1ImageProcessor'''), ('''mobilenet_v2''', '''MobileNetV2ImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevit''', '''MobileViTImageProcessor'''), ('''mobilevitv2''', '''MobileViTImageProcessor'''), ('''nat''', '''ViTImageProcessor'''), ('''oneformer''', '''OneFormerImageProcessor'''), ('''owlvit''', '''OwlViTImageProcessor'''), ('''perceiver''', '''PerceiverImageProcessor'''), ('''pix2struct''', '''Pix2StructImageProcessor'''), ('''poolformer''', '''PoolFormerImageProcessor'''), ('''regnet''', '''ConvNextImageProcessor'''), ('''resnet''', '''ConvNextImageProcessor'''), ('''sam''', '''SamImageProcessor'''), ('''segformer''', '''SegformerImageProcessor'''), ('''swiftformer''', '''ViTImageProcessor'''), ('''swin''', '''ViTImageProcessor'''), ('''swin2sr''', '''Swin2SRImageProcessor'''), ('''swinv2''', '''ViTImageProcessor'''), ('''table-transformer''', '''DetrImageProcessor'''), ('''timesformer''', '''VideoMAEImageProcessor'''), ('''tvlt''', '''TvltImageProcessor'''), ('''upernet''', '''SegformerImageProcessor'''), ('''van''', '''ConvNextImageProcessor'''), ('''videomae''', '''VideoMAEImageProcessor'''), ('''vilt''', '''ViltImageProcessor'''), ('''vit''', '''ViTImageProcessor'''), ('''vit_hybrid''', '''ViTHybridImageProcessor'''), ('''vit_mae''', '''ViTImageProcessor'''), ('''vit_msn''', '''ViTImageProcessor'''), ('''xclip''', '''CLIPImageProcessor'''), ('''yolos''', '''YolosImageProcessor'''), ] ) UpperCamelCase = _LazyAutoMapping(CONFIG_MAPPING_NAMES, IMAGE_PROCESSOR_MAPPING_NAMES) def __lowerCamelCase ( snake_case__ ) -> int: """simple docstring""" for module_name, extractors in IMAGE_PROCESSOR_MAPPING_NAMES.items(): if class_name in extractors: _SCREAMING_SNAKE_CASE = model_type_to_module_name(_a ) _SCREAMING_SNAKE_CASE = importlib.import_module(F'.{module_name}' ,"""transformers.models""" ) try: return getattr(_a ,_a ) except AttributeError: continue for _, extractor in IMAGE_PROCESSOR_MAPPING._extra_content.items(): if getattr(_a ,"""__name__""" ,_a ) == class_name: return extractor # We did not fine the class, but maybe it's because a dep is missing. In that case, the class will be in the main # init and we return the proper dummy to get an appropriate error message. _SCREAMING_SNAKE_CASE = importlib.import_module("""transformers""" ) if hasattr(_a ,_a ): return getattr(_a ,_a ) return None def __lowerCamelCase ( snake_case__ ,snake_case__ = None ,snake_case__ = False ,snake_case__ = False ,snake_case__ = None ,snake_case__ = None ,snake_case__ = None ,snake_case__ = False ,**snake_case__ ,) -> Union[str, Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = get_file_from_repo( _a ,_a ,cache_dir=_a ,force_download=_a ,resume_download=_a ,proxies=_a ,use_auth_token=_a ,revision=_a ,local_files_only=_a ,) if resolved_config_file is None: logger.info( """Could not locate the image processor configuration file, will try to use the model config instead.""" ) return {} with open(_a ,encoding="""utf-8""" ) as reader: return json.load(_a ) class __UpperCAmelCase : def __init__( self: Optional[int] ): '''simple docstring''' raise EnvironmentError( """AutoImageProcessor is designed to be instantiated """ """using the `AutoImageProcessor.from_pretrained(pretrained_model_name_or_path)` method.""" ) @classmethod @replace_list_option_in_docstrings(UpperCAmelCase_ ) def UpperCamelCase ( cls: Optional[int] , UpperCAmelCase_: List[str] , **UpperCAmelCase_: List[Any] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = kwargs.pop("""config""" , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = kwargs.pop("""trust_remote_code""" , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = True _SCREAMING_SNAKE_CASE = ImageProcessingMixin.get_image_processor_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = config_dict.get("""image_processor_type""" , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = None if "AutoImageProcessor" in config_dict.get("""auto_map""" , {} ): _SCREAMING_SNAKE_CASE = config_dict["auto_map"]["AutoImageProcessor"] # If we still don't have the image processor class, check if we're loading from a previous feature extractor config # and if so, infer the image processor class from there. if image_processor_class is None and image_processor_auto_map is None: _SCREAMING_SNAKE_CASE = config_dict.pop("""feature_extractor_type""" , UpperCAmelCase_ ) if feature_extractor_class is not None: logger.warning( """Could not find image processor class in the image processor config or the model config. Loading""" """ based on pattern matching with the model's feature extractor configuration.""" ) _SCREAMING_SNAKE_CASE = feature_extractor_class.replace("""FeatureExtractor""" , """ImageProcessor""" ) if "AutoFeatureExtractor" in config_dict.get("""auto_map""" , {} ): _SCREAMING_SNAKE_CASE = config_dict["auto_map"]["AutoFeatureExtractor"] _SCREAMING_SNAKE_CASE = feature_extractor_auto_map.replace("""FeatureExtractor""" , """ImageProcessor""" ) logger.warning( """Could not find image processor auto map in the image processor config or the model config.""" """ Loading based on pattern matching with the model's feature extractor configuration.""" ) # If we don't find the image processor class in the image processor config, let's try the model config. if image_processor_class is None and image_processor_auto_map is None: if not isinstance(UpperCAmelCase_ , UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = AutoConfig.from_pretrained(UpperCAmelCase_ , **UpperCAmelCase_ ) # It could be in `config.image_processor_type`` _SCREAMING_SNAKE_CASE = getattr(UpperCAmelCase_ , """image_processor_type""" , UpperCAmelCase_ ) if hasattr(UpperCAmelCase_ , """auto_map""" ) and "AutoImageProcessor" in config.auto_map: _SCREAMING_SNAKE_CASE = config.auto_map["AutoImageProcessor"] if image_processor_class is not None: _SCREAMING_SNAKE_CASE = image_processor_class_from_name(UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = image_processor_auto_map is not None _SCREAMING_SNAKE_CASE = image_processor_class is not None or type(UpperCAmelCase_ ) in IMAGE_PROCESSOR_MAPPING _SCREAMING_SNAKE_CASE = resolve_trust_remote_code( UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) if has_remote_code and trust_remote_code: _SCREAMING_SNAKE_CASE = get_class_from_dynamic_module( UpperCAmelCase_ , UpperCAmelCase_ , **UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = kwargs.pop("""code_revision""" , UpperCAmelCase_ ) if os.path.isdir(UpperCAmelCase_ ): image_processor_class.register_for_auto_class() return image_processor_class.from_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) elif image_processor_class is not None: return image_processor_class.from_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) # Last try: we use the IMAGE_PROCESSOR_MAPPING. elif type(UpperCAmelCase_ ) in IMAGE_PROCESSOR_MAPPING: _SCREAMING_SNAKE_CASE = IMAGE_PROCESSOR_MAPPING[type(UpperCAmelCase_ )] return image_processor_class.from_dict(UpperCAmelCase_ , **UpperCAmelCase_ ) raise ValueError( F'Unrecognized image processor in {pretrained_model_name_or_path}. Should have a ' F'`image_processor_type` key in its {IMAGE_PROCESSOR_NAME} of {CONFIG_NAME}, or one of the following ' F'`model_type` keys in its {CONFIG_NAME}: {", ".join(c for c in IMAGE_PROCESSOR_MAPPING_NAMES.keys() )}' ) @staticmethod def UpperCamelCase ( UpperCAmelCase_: Dict , UpperCAmelCase_: List[str] ): '''simple docstring''' IMAGE_PROCESSOR_MAPPING.register(UpperCAmelCase_ , UpperCAmelCase_ )
306
from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import tensorflow as tf from transformers import AutoTokenizer, TFAutoModelForSeqaSeqLM @require_tf @require_sentencepiece @require_tokenizers class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' @slow def __UpperCamelCase ( self : str ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Optional[Any] = TFAutoModelForSeqaSeqLM.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : List[str] = AutoTokenizer.from_pretrained("google/mt5-small" ) SCREAMING_SNAKE_CASE : Tuple = tokenizer("Hello there" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : Optional[Any] = tokenizer("Hi I am" , return_tensors="tf" ).input_ids SCREAMING_SNAKE_CASE : str = model(a , labels=a ).loss SCREAMING_SNAKE_CASE : Any = -tf.math.reduce_mean(a ).numpy() SCREAMING_SNAKE_CASE : Union[str, Any] = -21.22_8168 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 2e-4 )
76
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available __lowerCAmelCase : str = { "configuration_megatron_bert": ["MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP", "MegatronBertConfig"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: __lowerCAmelCase : Optional[Any] = [ "MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST", "MegatronBertForCausalLM", "MegatronBertForMaskedLM", "MegatronBertForMultipleChoice", "MegatronBertForNextSentencePrediction", "MegatronBertForPreTraining", "MegatronBertForQuestionAnswering", "MegatronBertForSequenceClassification", "MegatronBertForTokenClassification", "MegatronBertModel", "MegatronBertPreTrainedModel", ] if TYPE_CHECKING: from .configuration_megatron_bert import MEGATRON_BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, MegatronBertConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_megatron_bert import ( MEGATRON_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, MegatronBertForCausalLM, MegatronBertForMaskedLM, MegatronBertForMultipleChoice, MegatronBertForNextSentencePrediction, MegatronBertForPreTraining, MegatronBertForQuestionAnswering, MegatronBertForSequenceClassification, MegatronBertForTokenClassification, MegatronBertModel, MegatronBertPreTrainedModel, ) else: import sys __lowerCAmelCase : Optional[Any] = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
156
from math import factorial def lowerCamelCase__ ( _a , _a , _a): if successes > trials: raise ValueError("successes must be lower or equal to trials") if trials < 0 or successes < 0: raise ValueError("the function is defined for non-negative integers") if not isinstance(_a , _a) or not isinstance(_a , _a): raise ValueError("the function is defined for non-negative integers") if not 0 < prob < 1: raise ValueError("prob has to be in range of 1 - 0") SCREAMING_SNAKE_CASE : int = (prob**successes) * ((1 - prob) ** (trials - successes)) # Calculate the binomial coefficient: n! / k!(n-k)! SCREAMING_SNAKE_CASE : List[Any] = float(factorial(_a)) coefficient /= factorial(_a) * factorial(trials - successes) return probability * coefficient if __name__ == "__main__": from doctest import testmod testmod() print('Probability of 2 successes out of 4 trails') print('with probability of 0.75 is:', end=' ') print(binomial_distribution(2, 4, 0.75))
76
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_time_series_transformer": [ "TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP", "TimeSeriesTransformerConfig", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST", "TimeSeriesTransformerForPrediction", "TimeSeriesTransformerModel", "TimeSeriesTransformerPreTrainedModel", ] if TYPE_CHECKING: from .configuration_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, TimeSeriesTransformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_time_series_transformer import ( TIME_SERIES_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, TimeSeriesTransformerForPrediction, TimeSeriesTransformerModel, TimeSeriesTransformerPreTrainedModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
7
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =CustomTokenizer pass
76
0
__A : Optional[int] = "Tobias Carryer" from time import time class A_ : def __init__( self , _A , _A , _A , _A=int(time() ) ): # noqa: B008 '''simple docstring''' UpperCAmelCase = multiplier UpperCAmelCase = increment UpperCAmelCase = modulo UpperCAmelCase = seed def _lowercase ( self ): '''simple docstring''' UpperCAmelCase = (self.multiplier * self.seed + self.increment) % self.modulo return self.seed if __name__ == "__main__": # Show the LCG in action. __A : List[str] = LinearCongruentialGenerator(1_664_525, 1_013_904_223, 2 << 31) while True: print(lcg.next_number())
273
import logging import random import ray from transformers import RagConfig, RagRetriever, RagTokenizer from transformers.models.rag.retrieval_rag import CustomHFIndex a_ = logging.getLogger(__name__) class _UpperCamelCase : '''simple docstring''' def __init__( self : Any ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = False def __UpperCamelCase ( self : str , a : str , a : Optional[int] , a : Any , a : str ) -> List[Any]: """simple docstring""" if not self.initialized: SCREAMING_SNAKE_CASE : List[str] = RagRetriever( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[int] = True def __UpperCamelCase ( self : List[str] ) -> Optional[int]: """simple docstring""" self.retriever.index.init_index() def __UpperCamelCase ( self : Optional[Any] , a : List[Any] , a : Any ) -> int: """simple docstring""" SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Dict = self.retriever._main_retrieve(a , a ) return doc_ids, retrieved_doc_embeds class _UpperCamelCase ( __A ): '''simple docstring''' def __init__( self : Tuple , a : Any , a : Tuple , a : Tuple , a : Tuple , a : List[Any]=None ) -> Optional[int]: """simple docstring""" if index is not None and index.is_initialized() and len(a ) > 0: raise ValueError( "When using Ray for distributed fine-tuning, " "you'll need to provide the paths instead, " "as the dataset and the index are loaded " "separately. More info in examples/rag/use_own_knowledge_dataset.py " ) super().__init__( a , question_encoder_tokenizer=a , generator_tokenizer=a , index=a , init_retrieval=a , ) SCREAMING_SNAKE_CASE : Optional[Any] = retrieval_workers if len(self.retrieval_workers ) > 0: ray.get( [ worker.create_rag_retriever.remote(a , a , a , a ) for worker in self.retrieval_workers ] ) def __UpperCamelCase ( self : Any ) -> Dict: """simple docstring""" logger.info("initializing retrieval" ) if len(self.retrieval_workers ) > 0: ray.get([worker.init_retrieval.remote() for worker in self.retrieval_workers] ) else: # Non-distributed training. Load index into this same process. self.index.init_index() def __UpperCamelCase ( self : Tuple , a : Optional[int] , a : Any ) -> int: """simple docstring""" if len(self.retrieval_workers ) > 0: # Select a random retrieval actor. SCREAMING_SNAKE_CASE : Optional[Any] = self.retrieval_workers[random.randint(0 , len(self.retrieval_workers ) - 1 )] SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : str = ray.get(random_worker.retrieve.remote(a , a ) ) else: SCREAMING_SNAKE_CASE ,SCREAMING_SNAKE_CASE : Any = self._main_retrieve(a , a ) return retrieved_doc_embeds, doc_ids, self.index.get_doc_dicts(a ) @classmethod def __UpperCamelCase ( cls : str , a : Optional[Any] , a : Any=None , **a : List[Any] ) -> str: """simple docstring""" return super(a , cls ).get_tokenizers(a , a , **a ) @classmethod def __UpperCamelCase ( cls : Union[str, Any] , a : int , a : Any , a : List[Any]=None , **a : Optional[Any] ) -> str: """simple docstring""" SCREAMING_SNAKE_CASE : str = kwargs.pop("config" , a ) or RagConfig.from_pretrained(a , **a ) SCREAMING_SNAKE_CASE : List[Any] = RagTokenizer.from_pretrained(a , config=a ) SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.question_encoder SCREAMING_SNAKE_CASE : List[Any] = rag_tokenizer.generator if indexed_dataset is not None: SCREAMING_SNAKE_CASE : str = "custom" SCREAMING_SNAKE_CASE : List[Any] = CustomHFIndex(config.retrieval_vector_size , a ) else: SCREAMING_SNAKE_CASE : List[str] = cls._build_index(a ) return cls( a , question_encoder_tokenizer=a , generator_tokenizer=a , retrieval_workers=a , index=a , )
76
0
from __future__ import annotations def UpperCamelCase ( _A, _A ): """simple docstring""" if b == 0: return (1, 0) (__magic_name__) : Dict = extended_euclid(_a, a % b ) __magic_name__ : Tuple = a // b return (y, x - k * y) def UpperCamelCase ( _A, _A, _A, _A ): """simple docstring""" (__magic_name__) : Dict = extended_euclid(_a, _a ) __magic_name__ : str = na * na __magic_name__ : int = ra * x * na + ra * y * na return (n % m + m) % m def UpperCamelCase ( _A, _A ): """simple docstring""" (__magic_name__) : List[str] = extended_euclid(_a, _a ) if b < 0: __magic_name__ : int = (b % n + n) % n return b def UpperCamelCase ( _A, _A, _A, _A ): """simple docstring""" __magic_name__ : Dict = invert_modulo(_a, _a ), invert_modulo(_a, _a ) __magic_name__ : List[Any] = na * na __magic_name__ : str = ra * x * na + ra * y * na return (n % m + m) % m if __name__ == "__main__": from doctest import testmod testmod(name="chinese_remainder_theorem", verbose=True) testmod(name="chinese_remainder_theorem2", verbose=True) testmod(name="invert_modulo", verbose=True) testmod(name="extended_euclid", verbose=True)
342
from typing import Any class _UpperCamelCase : '''simple docstring''' def __init__( self : Dict , a : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : int = data SCREAMING_SNAKE_CASE : int = None def __repr__( self : str ) -> str: """simple docstring""" return F"Node({self.data})" class _UpperCamelCase : '''simple docstring''' def __init__( self : List[str] ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Any = None def __iter__( self : Any ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Union[str, Any] = self.head while node: yield node.data SCREAMING_SNAKE_CASE : List[str] = node.next def __len__( self : str ) -> int: """simple docstring""" return sum(1 for _ in self ) def __repr__( self : Optional[Any] ) -> str: """simple docstring""" return "->".join([str(a ) for item in self] ) def __getitem__( self : List[Any] , a : int ) -> Any: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) for i, node in enumerate(self ): if i == index: return node return None def __setitem__( self : Tuple , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index < len(self ): raise ValueError("list index out of range." ) SCREAMING_SNAKE_CASE : str = self.head for _ in range(a ): SCREAMING_SNAKE_CASE : str = current.next SCREAMING_SNAKE_CASE : Any = data def __UpperCamelCase ( self : List[str] , a : Any ) -> None: """simple docstring""" self.insert_nth(len(self ) , a ) def __UpperCamelCase ( self : Union[str, Any] , a : Any ) -> None: """simple docstring""" self.insert_nth(0 , a ) def __UpperCamelCase ( self : Optional[Any] , a : int , a : Any ) -> None: """simple docstring""" if not 0 <= index <= len(self ): raise IndexError("list index out of range" ) SCREAMING_SNAKE_CASE : Any = Node(a ) if self.head is None: SCREAMING_SNAKE_CASE : Optional[int] = new_node elif index == 0: SCREAMING_SNAKE_CASE : Optional[int] = self.head # link new_node to head SCREAMING_SNAKE_CASE : List[Any] = new_node else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : Optional[int] = temp.next SCREAMING_SNAKE_CASE : int = new_node def __UpperCamelCase ( self : Optional[int] ) -> None: # print every node data """simple docstring""" print(self ) def __UpperCamelCase ( self : int ) -> Any: """simple docstring""" return self.delete_nth(0 ) def __UpperCamelCase ( self : Any ) -> Any: # delete from tail """simple docstring""" return self.delete_nth(len(self ) - 1 ) def __UpperCamelCase ( self : List[str] , a : int = 0 ) -> Any: """simple docstring""" if not 0 <= index <= len(self ) - 1: # test if index is valid raise IndexError("List index out of range." ) SCREAMING_SNAKE_CASE : Tuple = self.head # default first node if index == 0: SCREAMING_SNAKE_CASE : List[str] = self.head.next else: SCREAMING_SNAKE_CASE : Optional[Any] = self.head for _ in range(index - 1 ): SCREAMING_SNAKE_CASE : Any = temp.next SCREAMING_SNAKE_CASE : List[Any] = temp.next SCREAMING_SNAKE_CASE : List[str] = temp.next.next return delete_node.data def __UpperCamelCase ( self : List[Any] ) -> bool: """simple docstring""" return self.head is None def __UpperCamelCase ( self : Optional[int] ) -> None: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = None SCREAMING_SNAKE_CASE : str = self.head while current: # Store the current node's next node. SCREAMING_SNAKE_CASE : Any = current.next # Make the current node's next point backwards SCREAMING_SNAKE_CASE : List[Any] = prev # Make the previous node be the current node SCREAMING_SNAKE_CASE : Any = current # Make the current node the next node (to progress iteration) SCREAMING_SNAKE_CASE : str = next_node # Return prev in order to put the head at the end SCREAMING_SNAKE_CASE : Optional[Any] = prev def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Union[str, Any] = LinkedList() assert linked_list.is_empty() is True assert str(_a) == "" try: linked_list.delete_head() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. try: linked_list.delete_tail() raise AssertionError # This should not happen. except IndexError: assert True # This should happen. for i in range(10): assert len(_a) == i linked_list.insert_nth(_a , i + 1) assert str(_a) == "->".join(str(_a) for i in range(1 , 11)) linked_list.insert_head(0) linked_list.insert_tail(11) assert str(_a) == "->".join(str(_a) for i in range(0 , 12)) assert linked_list.delete_head() == 0 assert linked_list.delete_nth(9) == 10 assert linked_list.delete_tail() == 11 assert len(_a) == 9 assert str(_a) == "->".join(str(_a) for i in range(1 , 10)) assert all(linked_list[i] == i + 1 for i in range(0 , 9)) is True for i in range(0 , 9): SCREAMING_SNAKE_CASE : str = -i assert all(linked_list[i] == -i for i in range(0 , 9)) is True linked_list.reverse() assert str(_a) == "->".join(str(_a) for i in range(-8 , 1)) def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : Optional[Any] = [ -9, 100, Node(77345112), "dlrow olleH", 7, 5555, 0, -192.5_5555, "Hello, world!", 77.9, Node(10), None, None, 12.20, ] SCREAMING_SNAKE_CASE : List[Any] = LinkedList() for i in test_input: linked_list.insert_tail(_a) # Check if it's empty or not assert linked_list.is_empty() is False assert ( str(_a) == "-9->100->Node(77345112)->dlrow olleH->7->5555->0->" "-192.55555->Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the head SCREAMING_SNAKE_CASE : List[Any] = linked_list.delete_head() assert result == -9 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None->12.2" ) # Delete the tail SCREAMING_SNAKE_CASE : Any = linked_list.delete_tail() assert result == 12.2 assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None->None" ) # Delete a node in specific location in linked list SCREAMING_SNAKE_CASE : Any = linked_list.delete_nth(10) assert result is None assert ( str(_a) == "100->Node(77345112)->dlrow olleH->7->5555->0->-192.55555->" "Hello, world!->77.9->Node(10)->None" ) # Add a Node instance to its head linked_list.insert_head(Node("Hello again, world!")) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None" ) # Add None to its tail linked_list.insert_tail(_a) assert ( str(_a) == "Node(Hello again, world!)->100->Node(77345112)->dlrow olleH->" "7->5555->0->-192.55555->Hello, world!->77.9->Node(10)->None->None" ) # Reverse the linked list linked_list.reverse() assert ( str(_a) == "None->None->Node(10)->77.9->Hello, world!->-192.55555->0->5555->" "7->dlrow olleH->Node(77345112)->100->Node(Hello again, world!)" ) def lowerCamelCase__ ( ): from doctest import testmod testmod() SCREAMING_SNAKE_CASE : Optional[int] = LinkedList() linked_list.insert_head(input("Inserting 1st at head ").strip()) linked_list.insert_head(input("Inserting 2nd at head ").strip()) print("\nPrint list:") linked_list.print_list() linked_list.insert_tail(input("\nInserting 1st at tail ").strip()) linked_list.insert_tail(input("Inserting 2nd at tail ").strip()) print("\nPrint list:") linked_list.print_list() print("\nDelete head") linked_list.delete_head() print("Delete tail") linked_list.delete_tail() print("\nPrint list:") linked_list.print_list() print("\nReverse linked list") linked_list.reverse() print("\nPrint list:") linked_list.print_list() print("\nString representation of linked list:") print(_a) print("\nReading/changing Node data using indexing:") print(f"Element at Position 1: {linked_list[1]}") SCREAMING_SNAKE_CASE : Dict = input("Enter New Value: ").strip() print("New list:") print(_a) print(f"length of linked_list is : {len(_a)}") if __name__ == "__main__": main()
76
0
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ....audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ....feature_extraction_sequence_utils import SequenceFeatureExtractor from ....feature_extraction_utils import BatchFeature from ....file_utils import PaddingStrategy, TensorType from ....utils import logging a__ : Union[str, Any] = logging.get_logger(__name__) class lowercase_ ( __A ): __UpperCAmelCase = ['input_features', 'attention_mask'] def __init__( self , a=80 , a=1_60_00 , a=0.0 , a=10 , a=25 , a="hamming_window" , a=3_27_68.0 , a=0.97 , a=1.0 , a=True , a=True , a=False , **a , ): super().__init__(feature_size=a , sampling_rate=a , padding_value=a , **a ) UpperCamelCase__ = feature_size UpperCamelCase__ = sampling_rate UpperCamelCase__ = padding_value UpperCamelCase__ = hop_length UpperCamelCase__ = win_length UpperCamelCase__ = frame_signal_scale UpperCamelCase__ = preemphasis_coeff UpperCamelCase__ = mel_floor UpperCamelCase__ = normalize_means UpperCamelCase__ = normalize_vars UpperCamelCase__ = win_function UpperCamelCase__ = return_attention_mask UpperCamelCase__ = win_length * sampling_rate // 10_00 UpperCamelCase__ = hop_length * sampling_rate // 10_00 UpperCamelCase__ = optimal_fft_length(self.sample_size ) UpperCamelCase__ = (self.n_fft // 2) + 1 def __a ( self , a ): if self.win_function == "hamming_window": UpperCamelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=a ) else: UpperCamelCase__ = window_function(window_length=self.sample_size , name=self.win_function ) UpperCamelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.feature_size , min_frequency=0.0 , max_frequency=self.sampling_rate / 2.0 , sampling_rate=self.sampling_rate , ) UpperCamelCase__ = spectrogram( one_waveform * self.frame_signal_scale , window=a , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , center=a , preemphasis=self.preemphasis_coeff , mel_filters=a , mel_floor=self.mel_floor , log_mel="log" , ) return msfc_features.T def __a ( self , a , a , a ): if self.normalize_means: UpperCamelCase__ = x[:input_length].mean(axis=0 ) UpperCamelCase__ = np.subtract(a , a ) if self.normalize_vars: UpperCamelCase__ = x[:input_length].std(axis=0 ) UpperCamelCase__ = np.divide(a , a ) if input_length < x.shape[0]: UpperCamelCase__ = padding_value # make sure array is in float32 UpperCamelCase__ = x.astype(np.floataa ) return x def __a ( self , a , a = None ): UpperCamelCase__ = attention_mask.sum(-1 ) if attention_mask is not None else [x.shape[0] for x in input_features] return [self._normalize_one(a , a , self.padding_value ) for x, n in zip(a , a )] def __call__( self , a , a = False , a = None , a = False , a = None , a = None , a = None , a = None , **a , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f'''The model corresponding to this feature extractor: {self} was trained using a sampling rate of''' f''' {self.sampling_rate}. Please make sure that the provided `raw_speech` input was sampled with''' f''' {self.sampling_rate} and not {sampling_rate}.''' ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) UpperCamelCase__ = isinstance(a , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(f'''Only mono-channel audio is supported for input to {self}''' ) UpperCamelCase__ = is_batched_numpy or ( isinstance(a , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCamelCase__ = [np.asarray(a , dtype=np.floataa ) for speech in raw_speech] elif not is_batched and not isinstance(a , np.ndarray ): UpperCamelCase__ = np.asarray(a , dtype=np.floataa ) elif isinstance(a , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): UpperCamelCase__ = raw_speech.astype(np.floataa ) # always return batch if not is_batched: UpperCamelCase__ = [raw_speech] # extract fbank features UpperCamelCase__ = [self._extract_mfsc_features(a ) for one_waveform in raw_speech] # convert into correct format for padding UpperCamelCase__ = BatchFeature({"input_features": features} ) UpperCamelCase__ = self.pad( a , padding=a , max_length=a , truncation=a , pad_to_multiple_of=a , return_attention_mask=a , **a , ) # make sure list is in array format UpperCamelCase__ = padded_inputs.get("input_features" ) if isinstance(input_features[0] , a ): UpperCamelCase__ = [np.asarray(a , dtype=np.floataa ) for feature in input_features] UpperCamelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCamelCase__ = [np.asarray(a , dtype=np.intaa ) for array in attention_mask] if self.normalize_means or self.normalize_vars: UpperCamelCase__ = ( np.array(a , dtype=np.intaa ) if self._get_padding_strategies(a , max_length=a ) is not PaddingStrategy.DO_NOT_PAD and padding else None ) UpperCamelCase__ = self.normalize( padded_inputs["input_features"] , attention_mask=a ) if return_tensors is not None: UpperCamelCase__ = padded_inputs.convert_to_tensors(a ) return padded_inputs
80
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger a_ = get_logger(__name__) class _UpperCamelCase ( enum.Enum ): '''simple docstring''' lowerCamelCase__ ='all_checks' lowerCamelCase__ ='basic_checks' lowerCamelCase__ ='no_checks' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a , _a=None): if expected_checksums is None: logger.info("Unable to verify checksums.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreDownloadedFiles(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedDownloadedFile(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] SCREAMING_SNAKE_CASE : Tuple = " for " + verification_name if verification_name is not None else "" if len(_a) > 0: raise NonMatchingChecksumError( f"Checksums didn't match{for_verification_name}:\n" f"{bad_urls}\n" "Set `verification_mode='no_checks'` to skip checksums verification and ignore this error") logger.info("All the checksums matched successfully" + for_verification_name) class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' class _UpperCamelCase ( __A ): '''simple docstring''' def lowerCamelCase__ ( _a , _a): if expected_splits is None: logger.info("Unable to verify splits sizes.") return if len(set(_a) - set(_a)) > 0: raise ExpectedMoreSplits(str(set(_a) - set(_a))) if len(set(_a) - set(_a)) > 0: raise UnexpectedSplits(str(set(_a) - set(_a))) SCREAMING_SNAKE_CASE : List[str] = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_a) > 0: raise NonMatchingSplitsSizesError(str(_a)) logger.info("All the splits matched successfully.") def lowerCamelCase__ ( _a , _a = True): if record_checksum: SCREAMING_SNAKE_CASE : List[str] = shaaaa() with open(_a , "rb") as f: for chunk in iter(lambda: f.read(1 << 20) , b""): m.update(_a) SCREAMING_SNAKE_CASE : Optional[int] = m.hexdigest() else: SCREAMING_SNAKE_CASE : List[str] = None return {"num_bytes": os.path.getsize(_a), "checksum": checksum} def lowerCamelCase__ ( _a): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
76
0
import enum import os from hashlib import shaaaa from typing import Optional from .. import config from .logging import get_logger __snake_case = get_logger(__name__) class UpperCAmelCase_ ( enum.Enum ): """simple docstring""" UpperCamelCase_ : Union[str, Any] ='all_checks' UpperCamelCase_ : Optional[Any] ='basic_checks' UpperCamelCase_ : Dict ='no_checks' class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" def _A ( SCREAMING_SNAKE_CASE__ : Any , SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str]=None ): if expected_checksums is None: logger.info('''Unable to verify checksums.''' ) return if len(set(_a ) - set(_a ) ) > 0: raise ExpectedMoreDownloadedFiles(str(set(_a ) - set(_a ) ) ) if len(set(_a ) - set(_a ) ) > 0: raise UnexpectedDownloadedFile(str(set(_a ) - set(_a ) ) ) UpperCamelCase :str = [url for url in expected_checksums if expected_checksums[url] != recorded_checksums[url]] UpperCamelCase :Tuple = " for " + verification_name if verification_name is not None else "" if len(_a ) > 0: raise NonMatchingChecksumError( F'''Checksums didn\'t match{for_verification_name}:\n''' F'''{bad_urls}\n''' '''Set `verification_mode=\'no_checks\'` to skip checksums verification and ignore this error''' ) logger.info('''All the checksums matched successfully''' + for_verification_name ) class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" class UpperCAmelCase_ ( __A ): """simple docstring""" def _A ( SCREAMING_SNAKE_CASE__ : str , SCREAMING_SNAKE_CASE__ : Dict ): if expected_splits is None: logger.info('''Unable to verify splits sizes.''' ) return if len(set(_a ) - set(_a ) ) > 0: raise ExpectedMoreSplits(str(set(_a ) - set(_a ) ) ) if len(set(_a ) - set(_a ) ) > 0: raise UnexpectedSplits(str(set(_a ) - set(_a ) ) ) UpperCamelCase :List[str] = [ {"expected": expected_splits[name], "recorded": recorded_splits[name]} for name in expected_splits if expected_splits[name].num_examples != recorded_splits[name].num_examples ] if len(_a ) > 0: raise NonMatchingSplitsSizesError(str(_a ) ) logger.info('''All the splits matched successfully.''' ) def _A ( SCREAMING_SNAKE_CASE__ : Optional[Any] , SCREAMING_SNAKE_CASE__ : str = True ): if record_checksum: UpperCamelCase :List[str] = shaaaa() with open(_a , '''rb''' ) as f: for chunk in iter(lambda: f.read(1 << 20 ) , B'''''' ): m.update(_a ) UpperCamelCase :Optional[int] = m.hexdigest() else: UpperCamelCase :List[str] = None return {"num_bytes": os.path.getsize(_a ), "checksum": checksum} def _A ( SCREAMING_SNAKE_CASE__ : Any ): if dataset_size and config.IN_MEMORY_MAX_SIZE: return dataset_size < config.IN_MEMORY_MAX_SIZE else: return False
259
import argparse import json import numpy import torch from transformers.models.xlm.tokenization_xlm import VOCAB_FILES_NAMES from transformers.utils import CONFIG_NAME, WEIGHTS_NAME, logging logging.set_verbosity_info() def lowerCamelCase__ ( _a , _a): # Load checkpoint SCREAMING_SNAKE_CASE : int = torch.load(_a , map_location="cpu") SCREAMING_SNAKE_CASE : Dict = chkpt["model"] # We have the base model one level deeper than the original XLM repository SCREAMING_SNAKE_CASE : Optional[int] = {} for k, v in state_dict.items(): if "pred_layer" in k: SCREAMING_SNAKE_CASE : List[str] = v else: SCREAMING_SNAKE_CASE : int = v SCREAMING_SNAKE_CASE : int = chkpt["params"] SCREAMING_SNAKE_CASE : Union[str, Any] = {n: v for n, v in config.items() if not isinstance(_a , (torch.FloatTensor, numpy.ndarray))} SCREAMING_SNAKE_CASE : List[Any] = chkpt["dico_word2id"] SCREAMING_SNAKE_CASE : List[Any] = {s + "</w>" if s.find("@@") == -1 and i > 13 else s.replace("@@" , ""): i for s, i in vocab.items()} # Save pytorch-model SCREAMING_SNAKE_CASE : Tuple = pytorch_dump_folder_path + "/" + WEIGHTS_NAME SCREAMING_SNAKE_CASE : Any = pytorch_dump_folder_path + "/" + CONFIG_NAME SCREAMING_SNAKE_CASE : Optional[int] = pytorch_dump_folder_path + "/" + VOCAB_FILES_NAMES["vocab_file"] print(f"Save PyTorch model to {pytorch_weights_dump_path}") torch.save(_a , _a) print(f"Save configuration file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") print(f"Save vocab file to {pytorch_config_dump_path}") with open(_a , "w" , encoding="utf-8") as f: f.write(json.dumps(_a , indent=2) + "\n") if __name__ == "__main__": a_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--xlm_checkpoint_path', default=None, type=str, required=True, help='Path the official PyTorch dump.' ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) a_ = parser.parse_args() convert_xlm_checkpoint_to_pytorch(args.xlm_checkpoint_path, args.pytorch_dump_folder_path)
76
0
from __future__ import annotations import collections import tempfile import unittest import numpy as np from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import is_tf_available, is_vision_available from ...test_modeling_tf_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_tf_bert import TFBertModelTester from ..clip.test_modeling_tf_clip import TFCLIPVisionModelTester from ..deit.test_modeling_tf_deit import TFDeiTModelTester from ..roberta.test_modeling_tf_roberta import TFRobertaModelTester from ..vit.test_modeling_tf_vit import TFViTModelTester if is_tf_available(): from transformers import ( TFBertModel, TFCLIPVisionModel, TFDeiTModel, TFRobertaModel, TFVisionTextDualEncoderModel, TFViTModel, VisionTextDualEncoderConfig, ) if is_vision_available(): from PIL import Image from transformers import VisionTextDualEncoderProcessor def lowerCamelCase_ ( lowerCamelCase__ ): if isinstance(_a , collections.abc.Iterable ): return x return (x, x) @require_tf class _SCREAMING_SNAKE_CASE : def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> Union[str, Any]: pass def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: pass def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: pass def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> int: lowerCamelCase_ = VisionTextDualEncoderConfig.from_vision_text_configs(lowercase , lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel(lowercase ) lowerCamelCase_ = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], config.projection_dim) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> Tuple: lowerCamelCase_ = self.get_vision_text_model(lowercase , lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) lowerCamelCase_ = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> Tuple: lowerCamelCase_ = self.get_vision_text_model(lowercase , lowercase ) lowerCamelCase_ = {"vision_model": vision_model, "text_model": text_model} lowerCamelCase_ = TFVisionTextDualEncoderModel.from_vision_text_pretrained(**lowercase ) lowerCamelCase_ = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) self.assertEqual(output["text_embeds"].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["image_embeds"].shape , (pixel_values.shape[0], model.config.projection_dim) ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> List[Any]: lowerCamelCase_ = self.get_vision_text_model(lowercase , lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) lowerCamelCase_ = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) lowerCamelCase_ = output[0].numpy() with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel.from_pretrained(lowercase ) lowerCamelCase_ = model(input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase ) lowerCamelCase_ = after_output[0].numpy() lowerCamelCase_ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase , 1e-5 ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> int: lowerCamelCase_ = self.get_vision_text_model(lowercase , lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) lowerCamelCase_ = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase ) lowerCamelCase_ = output.vision_model_output.attentions self.assertEqual(len(lowercase ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) lowerCamelCase_ = to_atuple(vision_model.config.image_size ) lowerCamelCase_ = to_atuple(vision_model.config.patch_size ) lowerCamelCase_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowerCamelCase_ = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) lowerCamelCase_ = output.text_model_output.attentions self.assertEqual(len(lowercase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase ) -> Dict: lowerCamelCase_ = np.abs((a - b) ).max() self.assertLessEqual(lowercase , lowercase , f'Difference between torch and flax is {diff} (>= {tol}).' ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_model(**lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Dict: lowerCamelCase_ = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> Union[str, Any]: lowerCamelCase_ = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = self.prepare_config_and_inputs() self.check_save_load(**lowercase ) def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**lowercase ) @slow def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ = self.get_pretrained_model_and_inputs() lowerCamelCase_ = model_a(**lowercase ) lowerCamelCase_ = outputs[0].numpy() with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel.from_pretrained(lowercase ) lowerCamelCase_ = model_a(**lowercase ) lowerCamelCase_ = after_outputs[0].numpy() lowerCamelCase_ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(lowercase , 1e-5 ) @require_tf class _SCREAMING_SNAKE_CASE ( __A , unittest.TestCase ): def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "hf-internal-testing/tiny-random-vit" , "hf-internal-testing/tiny-random-bert" ) lowerCamelCase_ = 13 lowerCamelCase_ = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) lowerCamelCase_ = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) lowerCamelCase_ = random_attention_mask([batch_size, 4] ) lowerCamelCase_ = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> Optional[int]: lowerCamelCase_ = TFViTModel(lowercase , name="vision_model" ) lowerCamelCase_ = TFBertModel(lowercase , name="text_model" ) return vision_model, text_model def SCREAMING_SNAKE_CASE_( self ) -> List[str]: lowerCamelCase_ = TFViTModelTester(self ) lowerCamelCase_ = TFBertModelTester(self ) lowerCamelCase_ = vit_model_tester.prepare_config_and_inputs() lowerCamelCase_ = bert_model_tester.prepare_config_and_inputs() lowerCamelCase_ = vision_config_and_inputs ( lowerCamelCase_ ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class _SCREAMING_SNAKE_CASE ( __A , unittest.TestCase ): def SCREAMING_SNAKE_CASE_( self ) -> Any: lowerCamelCase_ = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-deit-tf" , "hf-internal-testing/tiny-random-roberta" ) lowerCamelCase_ = 13 lowerCamelCase_ = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) lowerCamelCase_ = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) lowerCamelCase_ = random_attention_mask([batch_size, 4] ) lowerCamelCase_ = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase , lowercase , lowercase , lowercase=None , **lowercase ) -> Tuple: lowerCamelCase_ = self.get_vision_text_model(lowercase , lowercase ) lowerCamelCase_ = TFVisionTextDualEncoderModel(vision_model=lowercase , text_model=lowercase ) lowerCamelCase_ = model( input_ids=lowercase , pixel_values=lowercase , attention_mask=lowercase , output_attentions=lowercase ) lowerCamelCase_ = output.vision_model_output.attentions self.assertEqual(len(lowercase ) , vision_config.num_hidden_layers ) # in DEiT, the seq_len equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) lowerCamelCase_ = to_atuple(vision_model.config.image_size ) lowerCamelCase_ = to_atuple(vision_model.config.patch_size ) lowerCamelCase_ = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) lowerCamelCase_ = num_patches + 2 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) lowerCamelCase_ = output.text_model_output.attentions self.assertEqual(len(lowercase ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> List[Any]: lowerCamelCase_ = TFDeiTModel(lowercase , name="vision_model" ) lowerCamelCase_ = TFRobertaModel(lowercase , name="text_model" ) return vision_model, text_model def SCREAMING_SNAKE_CASE_( self ) -> Tuple: lowerCamelCase_ = TFDeiTModelTester(self ) lowerCamelCase_ = TFRobertaModelTester(self ) lowerCamelCase_ = vit_model_tester.prepare_config_and_inputs() lowerCamelCase_ = bert_model_tester.prepare_config_and_inputs() lowerCamelCase_ = vision_config_and_inputs ( lowerCamelCase_ ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_tf class _SCREAMING_SNAKE_CASE ( __A , unittest.TestCase ): def SCREAMING_SNAKE_CASE_( self ) -> List[Any]: lowerCamelCase_ = TFVisionTextDualEncoderModel.from_vision_text_pretrained( "Rocketknight1/tiny-random-clip-tf" , "hf-internal-testing/tiny-random-bert" ) lowerCamelCase_ = 13 lowerCamelCase_ = floats_tensor( [ batch_size, model.vision_model.config.num_channels, model.vision_model.config.image_size, model.vision_model.config.image_size, ] ) lowerCamelCase_ = ids_tensor([batch_size, 4] , model.text_model.config.vocab_size ) lowerCamelCase_ = random_attention_mask([batch_size, 4] ) lowerCamelCase_ = {"pixel_values": pixel_values, "input_ids": input_ids, "attention_mask": attention_mask} return model, inputs def SCREAMING_SNAKE_CASE_( self , lowercase , lowercase ) -> Dict: lowerCamelCase_ = TFCLIPVisionModel(lowercase , name="vision_model" ) lowerCamelCase_ = TFBertModel(lowercase , name="text_model" ) return vision_model, text_model def SCREAMING_SNAKE_CASE_( self ) -> Optional[Any]: lowerCamelCase_ = TFCLIPVisionModelTester(self ) lowerCamelCase_ = TFBertModelTester(self ) lowerCamelCase_ = clip_model_tester.prepare_config_and_inputs() lowerCamelCase_ = bert_model_tester.prepare_config_and_inputs() lowerCamelCase_ = vision_config_and_inputs ( lowerCamelCase_ ) = text_config_and_inputs return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": input_mask, "input_ids": input_ids, "text_token_type_ids": token_type_ids, "text_sequence_labels": sequence_labels, "text_token_labels": token_labels, "text_choice_labels": choice_labels, } @require_vision @require_tf class _SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE_( self ) -> int: lowerCamelCase_ = TFVisionTextDualEncoderModel.from_pretrained( "clip-italian/clip-italian" , logit_scale_init_value=1.0 , from_pt=lowercase ) lowerCamelCase_ = VisionTextDualEncoderProcessor.from_pretrained("clip-italian/clip-italian" ) lowerCamelCase_ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) lowerCamelCase_ = processor( text=["una foto di un gatto", "una foto di un cane"] , images=lowercase , padding=lowercase , return_tensors="np" ) lowerCamelCase_ = model(**lowercase ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) lowerCamelCase_ = np.array([[1.2_2_8_4_7_2_7, 0.3_1_0_4_1_2_2]] ) self.assertTrue(np.allclose(outputs.logits_per_image.numpy() , lowercase , atol=1e-3 ) )
19
def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(a - b) for a, b in zip(_a , _a))) def lowerCamelCase__ ( _a): if point: if isinstance(_a , _a): for item in point: if not isinstance(_a , (int, float)): SCREAMING_SNAKE_CASE : List[Any] = ( "Expected a list of numbers as input, found " f"{type(_a).__name__}" ) raise TypeError(_a) else: SCREAMING_SNAKE_CASE : List[Any] = f"Expected a list of numbers as input, found {type(_a).__name__}" raise TypeError(_a) else: raise ValueError("Missing an input") def lowerCamelCase__ ( _a , _a): _validate_point(_a) _validate_point(_a) if len(_a) != len(_a): raise ValueError("Both points must be in the same n-dimensional space") return float(sum(abs(x - y) for x, y in zip(_a , _a))) if __name__ == "__main__": import doctest doctest.testmod()
76
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, is_vision_available, ) lowerCAmelCase__ : Optional[int] = {'''processing_layoutxlm''': ['''LayoutXLMProcessor''']} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : str = ['''LayoutXLMTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCAmelCase__ : str = ['''LayoutXLMTokenizerFast'''] if TYPE_CHECKING: from .processing_layoutxlm import LayoutXLMProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm import LayoutXLMTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_layoutxlm_fast import LayoutXLMTokenizerFast else: import sys lowerCAmelCase__ : Dict = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
143
from ...configuration_utils import PretrainedConfig from ...utils import logging a_ = logging.get_logger(__name__) a_ = { 'sayakpaul/vit-msn-base': 'https://huggingface.co/sayakpaul/vit-msn-base/resolve/main/config.json', # See all ViT MSN models at https://huggingface.co/models?filter=vit_msn } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ ='vit_msn' def __init__( self : str , a : Tuple=768 , a : Tuple=12 , a : Any=12 , a : int=3072 , a : List[Any]="gelu" , a : Dict=0.0 , a : int=0.0 , a : str=0.02 , a : List[str]=1e-06 , a : List[Any]=224 , a : Union[str, Any]=16 , a : Union[str, Any]=3 , a : Tuple=True , **a : Dict , ) -> List[Any]: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Dict = hidden_size SCREAMING_SNAKE_CASE : Optional[Any] = num_hidden_layers SCREAMING_SNAKE_CASE : Optional[Any] = num_attention_heads SCREAMING_SNAKE_CASE : Optional[int] = intermediate_size SCREAMING_SNAKE_CASE : int = hidden_act SCREAMING_SNAKE_CASE : Union[str, Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE : Any = attention_probs_dropout_prob SCREAMING_SNAKE_CASE : List[Any] = initializer_range SCREAMING_SNAKE_CASE : int = layer_norm_eps SCREAMING_SNAKE_CASE : Dict = image_size SCREAMING_SNAKE_CASE : Tuple = patch_size SCREAMING_SNAKE_CASE : Optional[int] = num_channels SCREAMING_SNAKE_CASE : List[str] = qkv_bias
76
0
import json import logging import os import socket import git import numpy as np import torch logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - PID: %(process)d - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO, ) A_ :List[str] = logging.getLogger(__name__) def A ( a_ ) -> Tuple: __UpperCamelCase : Any =git.Repo(search_parent_directories=_a ) __UpperCamelCase : Optional[int] ={ "repo_id": str(_a ), "repo_sha": str(repo.head.object.hexsha ), "repo_branch": str(repo.active_branch ), } with open(os.path.join(_a ,'git_log.json' ) ,'w' ) as f: json.dump(_a ,_a ,indent=4 ) def A ( a_ ) -> Any: if params.n_gpu <= 0: __UpperCamelCase : Optional[Any] =0 __UpperCamelCase : int =-1 __UpperCamelCase : List[Any] =True __UpperCamelCase : str =False return assert torch.cuda.is_available() logger.info('Initializing GPUs' ) if params.n_gpu > 1: assert params.local_rank != -1 __UpperCamelCase : Union[str, Any] =int(os.environ['WORLD_SIZE'] ) __UpperCamelCase : Optional[Any] =int(os.environ['N_GPU_NODE'] ) __UpperCamelCase : int =int(os.environ['RANK'] ) # number of nodes / node ID __UpperCamelCase : Any =params.world_size // params.n_gpu_per_node __UpperCamelCase : Any =params.global_rank // params.n_gpu_per_node __UpperCamelCase : List[Any] =True assert params.n_nodes == int(os.environ['N_NODES'] ) assert params.node_id == int(os.environ['NODE_RANK'] ) # local job (single GPU) else: assert params.local_rank == -1 __UpperCamelCase : str =1 __UpperCamelCase : List[Any] =0 __UpperCamelCase : Tuple =0 __UpperCamelCase : Optional[Any] =0 __UpperCamelCase : Tuple =1 __UpperCamelCase : List[Any] =1 __UpperCamelCase : Tuple =False # sanity checks assert params.n_nodes >= 1 assert 0 <= params.node_id < params.n_nodes assert 0 <= params.local_rank <= params.global_rank < params.world_size assert params.world_size == params.n_nodes * params.n_gpu_per_node # define whether this is the master process / if we are in multi-node distributed mode __UpperCamelCase : Tuple =params.node_id == 0 and params.local_rank == 0 __UpperCamelCase : int =params.n_nodes > 1 # summary __UpperCamelCase : str =F'--- Global rank: {params.global_rank} - ' logger.info(PREFIX + 'Number of nodes: %i' % params.n_nodes ) logger.info(PREFIX + 'Node ID : %i' % params.node_id ) logger.info(PREFIX + 'Local rank : %i' % params.local_rank ) logger.info(PREFIX + 'World size : %i' % params.world_size ) logger.info(PREFIX + 'GPUs per node : %i' % params.n_gpu_per_node ) logger.info(PREFIX + 'Master : %s' % str(params.is_master ) ) logger.info(PREFIX + 'Multi-node : %s' % str(params.multi_node ) ) logger.info(PREFIX + 'Multi-GPU : %s' % str(params.multi_gpu ) ) logger.info(PREFIX + 'Hostname : %s' % socket.gethostname() ) # set GPU device torch.cuda.set_device(params.local_rank ) # initialize multi-GPU if params.multi_gpu: logger.info('Initializing PyTorch distributed' ) torch.distributed.init_process_group( init_method='env://' ,backend='nccl' ,) def A ( a_ ) -> Tuple: np.random.seed(args.seed ) torch.manual_seed(args.seed ) if args.n_gpu > 0: torch.cuda.manual_seed_all(args.seed )
71
import baseaa def lowerCamelCase__ ( _a): return baseaa.aaaencode(string.encode("utf-8")) def lowerCamelCase__ ( _a): return baseaa.aaadecode(_a).decode("utf-8") if __name__ == "__main__": import doctest doctest.testmod()
76
0
"""simple docstring""" def lowercase ( lowerCAmelCase__ : Tuple , lowerCAmelCase__ : Tuple ) -> List[Any]: _validate_point(_a ) _validate_point(_a ) if len(_a ) != len(_a ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(a - b ) for a, b in zip(_a , _a ) ) ) def lowercase ( lowerCAmelCase__ : Tuple ) -> str: if point: if isinstance(_a , _a ): for item in point: if not isinstance(_a , (int, float) ): __a = ( "Expected a list of numbers as input, found " f'''{type(_a ).__name__}''' ) raise TypeError(_a ) else: __a = f'''Expected a list of numbers as input, found {type(_a ).__name__}''' raise TypeError(_a ) else: raise ValueError('''Missing an input''' ) def lowercase ( lowerCAmelCase__ : Dict , lowerCAmelCase__ : Optional[Any] ) -> Optional[int]: _validate_point(_a ) _validate_point(_a ) if len(_a ) != len(_a ): raise ValueError('''Both points must be in the same n-dimensional space''' ) return float(sum(abs(x - y ) for x, y in zip(_a , _a ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
45
from datetime import datetime as dt import os from github import Github a_ = [ 'good first issue', 'good second issue', 'good difficult issue', 'feature request', 'new model', 'wip', ] def lowerCamelCase__ ( ): SCREAMING_SNAKE_CASE : int = Github(os.environ["GITHUB_TOKEN"]) SCREAMING_SNAKE_CASE : List[str] = g.get_repo("huggingface/transformers") SCREAMING_SNAKE_CASE : Optional[int] = repo.get_issues(state="open") for issue in open_issues: SCREAMING_SNAKE_CASE : List[Any] = sorted([comment for comment in issue.get_comments()] , key=lambda _a: i.created_at , reverse=_a) SCREAMING_SNAKE_CASE : str = comments[0] if len(_a) > 0 else None if ( last_comment is not None and last_comment.user.login == "github-actions[bot]" and (dt.utcnow() - issue.updated_at).days > 7 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would close issue {issue.number} since it has been 7 days of inactivity since bot mention.") issue.edit(state="closed") elif ( (dt.utcnow() - issue.updated_at).days > 23 and (dt.utcnow() - issue.created_at).days >= 30 and not any(label.name.lower() in LABELS_TO_EXEMPT for label in issue.get_labels()) ): # print(f"Would add stale comment to {issue.number}") issue.create_comment( "This issue has been automatically marked as stale because it has not had " "recent activity. If you think this still needs to be addressed " "please comment on this thread.\n\nPlease note that issues that do not follow the " "[contributing guidelines](https://github.com/huggingface/transformers/blob/main/CONTRIBUTING.md) " "are likely to be ignored.") if __name__ == "__main__": main()
76
0
import time from dataclasses import dataclass from multiprocessing import Pool from unittest import TestCase from unittest.mock import patch import multiprocess import numpy as np import pytest from datasets.utils.py_utils import ( NestedDataStructure, asdict, iflatmap_unordered, map_nested, temp_seed, temporary_assignment, zip_dict, ) from .utils import require_tf, require_torch def __lowerCamelCase ( snake_case__ ) -> int: # picklable for multiprocessing """simple docstring""" return x.sum() def __lowerCamelCase ( snake_case__ ) -> List[str]: # picklable for multiprocessing """simple docstring""" return i + 1 @dataclass class __UpperCAmelCase : __snake_case : int = 42 __snake_case : List[Any] = 42 class __UpperCAmelCase (__A ): def UpperCamelCase ( self: Tuple ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 1 _SCREAMING_SNAKE_CASE = [1, 2] _SCREAMING_SNAKE_CASE = {"a": 1, "b": 2} _SCREAMING_SNAKE_CASE = {"a": [1, 2], "b": [3, 4]} _SCREAMING_SNAKE_CASE = {"a": {"1": 1}, "b": 2} _SCREAMING_SNAKE_CASE = {"a": 1, "b": 2, "c": 3, "d": 4} _SCREAMING_SNAKE_CASE = {} _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = 2 _SCREAMING_SNAKE_CASE = [2, 3] _SCREAMING_SNAKE_CASE = {"a": 2, "b": 3} _SCREAMING_SNAKE_CASE = {"a": [2, 3], "b": [4, 5]} _SCREAMING_SNAKE_CASE = {"a": {"1": 2}, "b": 3} _SCREAMING_SNAKE_CASE = {"a": 2, "b": 3, "c": 4, "d": 5} self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ ) , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = 2 self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) _SCREAMING_SNAKE_CASE = {"a": np.eye(2 ), "b": np.zeros(3 ), "c": np.ones(2 )} _SCREAMING_SNAKE_CASE = {"a": 2, "b": 0, "c": 2} _SCREAMING_SNAKE_CASE = { "a": np.eye(2 ).astype(UpperCAmelCase_ ), "b": np.zeros(3 ).astype(UpperCAmelCase_ ), "c": np.ones(2 ).astype(UpperCAmelCase_ ), } self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , map_numpy=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(UpperCAmelCase_ , UpperCAmelCase_ , map_numpy=UpperCAmelCase_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) self.assertEqual(map_nested(UpperCAmelCase_ , UpperCAmelCase_ , map_numpy=UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) , UpperCAmelCase_ ) self.assertEqual( {k: v.tolist() for k, v in map_nested(UpperCAmelCase_ , UpperCAmelCase_ , map_numpy=UpperCAmelCase_ , num_proc=UpperCAmelCase_ ).items()} , {k: v.tolist() for k, v in expected_map_nested_sna_int.items()} , ) with self.assertRaises(UpperCAmelCase_ ): # can't pickle a local lambda map_nested(lambda UpperCAmelCase_ : x + 1 , UpperCAmelCase_ , num_proc=UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[int] ): '''simple docstring''' _SCREAMING_SNAKE_CASE = {"a": 1, "b": 2} _SCREAMING_SNAKE_CASE = {"a": 3, "b": 4} _SCREAMING_SNAKE_CASE = {"a": 5, "b": 6} _SCREAMING_SNAKE_CASE = sorted([("""a""", (1, 3, 5)), ("""b""", (2, 4, 6))] ) self.assertEqual(sorted(zip_dict(UpperCAmelCase_ , UpperCAmelCase_ , UpperCAmelCase_ ) ) , UpperCAmelCase_ ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' class __UpperCAmelCase : __snake_case : Union[str, Any] = "bar" _SCREAMING_SNAKE_CASE = Foo() self.assertEqual(foo.my_attr , """bar""" ) with temporary_assignment(UpperCAmelCase_ , """my_attr""" , """BAR""" ): self.assertEqual(foo.my_attr , """BAR""" ) self.assertEqual(foo.my_attr , """bar""" ) @pytest.mark.parametrize( """iterable_length, num_proc, expected_num_proc""" ,[ (1, None, 1), (1, 1, 1), (2, None, 1), (2, 1, 1), (2, 2, 1), (2, 3, 1), (3, 2, 1), (16, 16, 16), (16, 17, 16), (17, 16, 16), ] ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ,snake_case__ ) -> Optional[Any]: """simple docstring""" with patch("""datasets.utils.py_utils._single_map_nested""" ) as mock_single_map_nested, patch( """datasets.parallel.parallel.Pool""" ) as mock_multiprocessing_pool: _SCREAMING_SNAKE_CASE = {F'{i}': i for i in range(_a )} _SCREAMING_SNAKE_CASE = map_nested(lambda snake_case__ : x + 10 ,_a ,num_proc=_a ,parallel_min_length=16 ) if expected_num_proc == 1: assert mock_single_map_nested.called assert not mock_multiprocessing_pool.called else: assert not mock_single_map_nested.called assert mock_multiprocessing_pool.called assert mock_multiprocessing_pool.call_args[0][0] == expected_num_proc class __UpperCAmelCase (__A ): @require_tf def UpperCamelCase ( self: Union[str, Any] ): '''simple docstring''' import tensorflow as tf from tensorflow.keras import layers _SCREAMING_SNAKE_CASE = layers.Dense(2 ) def gen_random_output(): _SCREAMING_SNAKE_CASE = tf.random.uniform((1, 3) ) return model(UpperCAmelCase_ ).numpy() with temp_seed(42 , set_tensorflow=UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = gen_random_output() with temp_seed(42 , set_tensorflow=UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = gen_random_output() _SCREAMING_SNAKE_CASE = gen_random_output() np.testing.assert_equal(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @require_torch def UpperCamelCase ( self: Any ): '''simple docstring''' import torch def gen_random_output(): _SCREAMING_SNAKE_CASE = torch.nn.Linear(3 , 2 ) _SCREAMING_SNAKE_CASE = torch.rand(1 , 3 ) return model(UpperCAmelCase_ ).detach().numpy() with temp_seed(42 , set_pytorch=UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = gen_random_output() with temp_seed(42 , set_pytorch=UpperCAmelCase_ ): _SCREAMING_SNAKE_CASE = gen_random_output() _SCREAMING_SNAKE_CASE = gen_random_output() np.testing.assert_equal(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) def UpperCamelCase ( self: Optional[Any] ): '''simple docstring''' def gen_random_output(): return np.random.rand(1 , 3 ) with temp_seed(42 ): _SCREAMING_SNAKE_CASE = gen_random_output() with temp_seed(42 ): _SCREAMING_SNAKE_CASE = gen_random_output() _SCREAMING_SNAKE_CASE = gen_random_output() np.testing.assert_equal(UpperCAmelCase_ , UpperCAmelCase_ ) self.assertGreater(np.abs(outa - outa ).sum() , 0 ) @pytest.mark.parametrize("""input_data""" ,[{}] ) def __lowerCamelCase ( snake_case__ ) -> Optional[Any]: """simple docstring""" _SCREAMING_SNAKE_CASE = NestedDataStructure(_a ).data assert output_data == input_data @pytest.mark.parametrize( """data, expected_output""" ,[ ({}, []), ([], []), ("""foo""", ["""foo"""]), (["""foo""", """bar"""], ["""foo""", """bar"""]), ([["""foo""", """bar"""]], ["""foo""", """bar"""]), ([[["""foo"""], ["""bar"""]]], ["""foo""", """bar"""]), ([[["""foo"""], """bar"""]], ["""foo""", """bar"""]), ({"""a""": 1, """b""": 2}, [1, 2]), ({"""a""": [1, 2], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[1, 2]], """b""": [[3, 4]]}, [1, 2, 3, 4]), ({"""a""": [[1, 2]], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [[[3], [4]]]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [[3, 4]]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [3, 4]}, [1, 2, 3, 4]), ({"""a""": [[[1], [2]]], """b""": [3, [4]]}, [1, 2, 3, 4]), ({"""a""": {"""1""": 1}, """b""": 2}, [1, 2]), ({"""a""": {"""1""": [1]}, """b""": 2}, [1, 2]), ({"""a""": {"""1""": [1]}, """b""": [2]}, [1, 2]), ] ,) def __lowerCamelCase ( snake_case__ ,snake_case__ ) -> Any: """simple docstring""" _SCREAMING_SNAKE_CASE = NestedDataStructure(_a ).flatten() assert output == expected_output def __lowerCamelCase ( ) -> List[str]: """simple docstring""" _SCREAMING_SNAKE_CASE = A(x=1 ,y="""foobar""" ) _SCREAMING_SNAKE_CASE = {"x": 1, "y": "foobar"} assert asdict(_a ) == expected_output _SCREAMING_SNAKE_CASE = {"a": {"b": A(x=10 ,y="""foo""" )}, "c": [A(x=20 ,y="""bar""" )]} _SCREAMING_SNAKE_CASE = {"a": {"b": {"x": 10, "y": "foo"}}, "c": [{"x": 20, "y": "bar"}]} assert asdict(_a ) == expected_output with pytest.raises(_a ): asdict([1, A(x=10 ,y="""foo""" )] ) def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" return text.split() def __lowerCamelCase ( snake_case__ ) -> List[Any]: """simple docstring""" yield (time.time(), content) time.sleep(2 ) yield (time.time(), content) def __lowerCamelCase ( ) -> str: """simple docstring""" with Pool(2 ) as pool: _SCREAMING_SNAKE_CASE = list(iflatmap_unordered(_a ,_split_text ,kwargs_iterable=[{"""text""": """hello there"""}] * 10 ) ) assert out.count("""hello""" ) == 10 assert out.count("""there""" ) == 10 assert len(_a ) == 20 # check multiprocess from pathos (uses dill for pickling) with multiprocess.Pool(2 ) as pool: _SCREAMING_SNAKE_CASE = list(iflatmap_unordered(_a ,_split_text ,kwargs_iterable=[{"""text""": """hello there"""}] * 10 ) ) assert out.count("""hello""" ) == 10 assert out.count("""there""" ) == 10 assert len(_a ) == 20 # check that we get items as fast as possible with Pool(2 ) as pool: _SCREAMING_SNAKE_CASE = [] for yield_time, content in iflatmap_unordered( _a ,_aseconds_generator_of_aitems_with_timing ,kwargs_iterable=[{"""content""": """a"""}, {"""content""": """b"""}] ): assert yield_time < time.time() + 0.1, "we should each item directly after it was yielded" out.append(_a ) assert out.count("""a""" ) == 2 assert out.count("""b""" ) == 2 assert len(_a ) == 4
306
from typing import Dict, List, Optional, Union import numpy as np from transformers.utils import is_vision_available from transformers.utils.generic import TensorType from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, is_valid_image, to_numpy_array, valid_images, ) from ...utils import logging if is_vision_available(): import PIL a_ = logging.get_logger(__name__) def lowerCamelCase__ ( _a): if isinstance(_a , (list, tuple)) and isinstance(videos[0] , (list, tuple)) and is_valid_image(videos[0][0]): return videos elif isinstance(_a , (list, tuple)) and is_valid_image(videos[0]): return [videos] elif is_valid_image(_a): return [[videos]] raise ValueError(f"Could not make batched video from {videos}") class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =['pixel_values'] def __init__( self : Optional[Any] , a : bool = True , a : Dict[str, int] = None , a : PILImageResampling = PILImageResampling.BILINEAR , a : bool = True , a : Dict[str, int] = None , a : bool = True , a : Union[int, float] = 1 / 255 , a : bool = True , a : bool = True , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , **a : Tuple , ) -> None: """simple docstring""" super().__init__(**a ) SCREAMING_SNAKE_CASE : Tuple = size if size is not None else {"shortest_edge": 256} SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : List[str] = crop_size if crop_size is not None else {"height": 224, "width": 224} SCREAMING_SNAKE_CASE : str = get_size_dict(a , param_name="crop_size" ) SCREAMING_SNAKE_CASE : Dict = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : Optional[int] = do_center_crop SCREAMING_SNAKE_CASE : int = crop_size SCREAMING_SNAKE_CASE : int = resample SCREAMING_SNAKE_CASE : Any = do_rescale SCREAMING_SNAKE_CASE : int = rescale_factor SCREAMING_SNAKE_CASE : Tuple = offset SCREAMING_SNAKE_CASE : str = do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN SCREAMING_SNAKE_CASE : Dict = image_std if image_std is not None else IMAGENET_STANDARD_STD def __UpperCamelCase ( self : Optional[Any] , a : np.ndarray , a : Dict[str, int] , a : PILImageResampling = PILImageResampling.BILINEAR , a : Optional[Union[str, ChannelDimension]] = None , **a : Union[str, Any] , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = get_size_dict(a , default_to_square=a ) if "shortest_edge" in size: SCREAMING_SNAKE_CASE : str = get_resize_output_image_size(a , size["shortest_edge"] , default_to_square=a ) elif "height" in size and "width" in size: SCREAMING_SNAKE_CASE : Dict = (size["height"], size["width"]) else: raise ValueError(F"Size must have 'height' and 'width' or 'shortest_edge' as keys. Got {size.keys()}" ) return resize(a , size=a , resample=a , data_format=a , **a ) def __UpperCamelCase ( self : List[str] , a : np.ndarray , a : Dict[str, int] , a : Optional[Union[str, ChannelDimension]] = None , **a : str , ) -> np.ndarray: """simple docstring""" SCREAMING_SNAKE_CASE : str = get_size_dict(a ) if "height" not in size or "width" not in size: raise ValueError(F"Size must have 'height' and 'width' as keys. Got {size.keys()}" ) return center_crop(a , size=(size["height"], size["width"]) , data_format=a , **a ) def __UpperCamelCase ( self : List[Any] , a : np.ndarray , a : Union[int, float] , a : bool = True , a : Optional[Union[str, ChannelDimension]] = None , **a : Tuple , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : int = image.astype(np.floataa ) if offset: SCREAMING_SNAKE_CASE : Union[str, Any] = image - (scale / 2) return rescale(a , scale=a , data_format=a , **a ) def __UpperCamelCase ( self : int , a : np.ndarray , a : Union[float, List[float]] , a : Union[float, List[float]] , a : Optional[Union[str, ChannelDimension]] = None , **a : List[str] , ) -> np.ndarray: """simple docstring""" return normalize(a , mean=a , std=a , data_format=a , **a ) def __UpperCamelCase ( self : Tuple , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[ChannelDimension] = ChannelDimension.FIRST , ) -> np.ndarray: """simple docstring""" if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_center_crop and crop_size is None: raise ValueError("Crop size must be specified if do_center_crop is True." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) if offset and not do_rescale: raise ValueError("For offset, do_rescale must also be set to True." ) # All transformations expect numpy arrays. SCREAMING_SNAKE_CASE : List[str] = to_numpy_array(a ) if do_resize: SCREAMING_SNAKE_CASE : Optional[Any] = self.resize(image=a , size=a , resample=a ) if do_center_crop: SCREAMING_SNAKE_CASE : Union[str, Any] = self.center_crop(a , size=a ) if do_rescale: SCREAMING_SNAKE_CASE : Any = self.rescale(image=a , scale=a , offset=a ) if do_normalize: SCREAMING_SNAKE_CASE : Tuple = self.normalize(image=a , mean=a , std=a ) SCREAMING_SNAKE_CASE : Optional[int] = to_channel_dimension_format(a , a ) return image def __UpperCamelCase ( self : Dict , a : ImageInput , a : bool = None , a : Dict[str, int] = None , a : PILImageResampling = None , a : bool = None , a : Dict[str, int] = None , a : bool = None , a : float = None , a : bool = None , a : bool = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[float, List[float]]] = None , a : Optional[Union[str, TensorType]] = None , a : ChannelDimension = ChannelDimension.FIRST , **a : Tuple , ) -> PIL.Image.Image: """simple docstring""" SCREAMING_SNAKE_CASE : str = do_resize if do_resize is not None else self.do_resize SCREAMING_SNAKE_CASE : Union[str, Any] = resample if resample is not None else self.resample SCREAMING_SNAKE_CASE : int = do_center_crop if do_center_crop is not None else self.do_center_crop SCREAMING_SNAKE_CASE : str = do_rescale if do_rescale is not None else self.do_rescale SCREAMING_SNAKE_CASE : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor SCREAMING_SNAKE_CASE : Optional[Any] = offset if offset is not None else self.offset SCREAMING_SNAKE_CASE : str = do_normalize if do_normalize is not None else self.do_normalize SCREAMING_SNAKE_CASE : Optional[int] = image_mean if image_mean is not None else self.image_mean SCREAMING_SNAKE_CASE : Optional[Any] = image_std if image_std is not None else self.image_std SCREAMING_SNAKE_CASE : int = size if size is not None else self.size SCREAMING_SNAKE_CASE : List[Any] = get_size_dict(a , default_to_square=a ) SCREAMING_SNAKE_CASE : Tuple = crop_size if crop_size is not None else self.crop_size SCREAMING_SNAKE_CASE : Union[str, Any] = get_size_dict(a , param_name="crop_size" ) if not valid_images(a ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) SCREAMING_SNAKE_CASE : Optional[int] = make_batched(a ) SCREAMING_SNAKE_CASE : List[Any] = [ [ self._preprocess_image( image=a , do_resize=a , size=a , resample=a , do_center_crop=a , crop_size=a , do_rescale=a , rescale_factor=a , offset=a , do_normalize=a , image_mean=a , image_std=a , data_format=a , ) for img in video ] for video in videos ] SCREAMING_SNAKE_CASE : Optional[int] = {"pixel_values": videos} return BatchFeature(data=a , tensor_type=a )
76
0
import os import tempfile import unittest from transformers import DistilBertConfig, is_torch_available from transformers.testing_utils import require_torch, require_torch_gpu, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, ) class __lowerCAmelCase ( __A ): """simple docstring""" def __init__( self : List[str] , _snake_case : List[Any] , _snake_case : List[Any]=13 , _snake_case : Dict=7 , _snake_case : Optional[Any]=True , _snake_case : List[Any]=True , _snake_case : Tuple=False , _snake_case : Dict=True , _snake_case : Dict=99 , _snake_case : List[str]=32 , _snake_case : Dict=5 , _snake_case : Dict=4 , _snake_case : Tuple=37 , _snake_case : str="gelu" , _snake_case : Optional[Any]=0.1 , _snake_case : Optional[int]=0.1 , _snake_case : int=512 , _snake_case : Union[str, Any]=16 , _snake_case : Dict=2 , _snake_case : List[str]=0.02 , _snake_case : Any=3 , _snake_case : Tuple=4 , _snake_case : Union[str, Any]=None , ): __lowercase : Optional[Any] = parent __lowercase : Optional[Any] = batch_size __lowercase : List[str] = seq_length __lowercase : Optional[Any] = is_training __lowercase : str = use_input_mask __lowercase : List[str] = use_token_type_ids __lowercase : int = use_labels __lowercase : Any = vocab_size __lowercase : List[str] = hidden_size __lowercase : Optional[Any] = num_hidden_layers __lowercase : Any = num_attention_heads __lowercase : List[Any] = intermediate_size __lowercase : List[Any] = hidden_act __lowercase : Any = hidden_dropout_prob __lowercase : Union[str, Any] = attention_probs_dropout_prob __lowercase : Dict = max_position_embeddings __lowercase : str = type_vocab_size __lowercase : str = type_sequence_label_size __lowercase : Any = initializer_range __lowercase : Any = num_labels __lowercase : List[Any] = num_choices __lowercase : Dict = scope def snake_case_ ( self : str ): __lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) __lowercase : List[str] = None if self.use_input_mask: __lowercase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) __lowercase : Tuple = None __lowercase : Optional[Any] = None __lowercase : Tuple = None if self.use_labels: __lowercase : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) __lowercase : Dict = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) __lowercase : List[Any] = ids_tensor([self.batch_size] , self.num_choices ) __lowercase : List[Any] = self.get_config() return config, input_ids, input_mask, sequence_labels, token_labels, choice_labels def snake_case_ ( self : Union[str, Any] ): return DistilBertConfig( vocab_size=self.vocab_size , dim=self.hidden_size , n_layers=self.num_hidden_layers , n_heads=self.num_attention_heads , hidden_dim=self.intermediate_size , hidden_act=self.hidden_act , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , ) def snake_case_ ( self : str , _snake_case : Any , _snake_case : Optional[Any] , _snake_case : List[str] , _snake_case : List[str] , _snake_case : Union[str, Any] , _snake_case : List[str] ): __lowercase : Optional[Any] = DistilBertModel(config=_snake_case ) model.to(_snake_case ) model.eval() __lowercase : Tuple = model(_snake_case , _snake_case ) __lowercase : str = model(_snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def snake_case_ ( self : Optional[Any] , _snake_case : Dict , _snake_case : Tuple , _snake_case : Optional[int] , _snake_case : Optional[Any] , _snake_case : Any , _snake_case : int ): __lowercase : Optional[int] = DistilBertForMaskedLM(config=_snake_case ) model.to(_snake_case ) model.eval() __lowercase : Optional[Any] = model(_snake_case , attention_mask=_snake_case , labels=_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def snake_case_ ( self : str , _snake_case : Optional[Any] , _snake_case : int , _snake_case : Tuple , _snake_case : Union[str, Any] , _snake_case : int , _snake_case : Optional[Any] ): __lowercase : List[Any] = DistilBertForQuestionAnswering(config=_snake_case ) model.to(_snake_case ) model.eval() __lowercase : Union[str, Any] = model( _snake_case , attention_mask=_snake_case , start_positions=_snake_case , end_positions=_snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def snake_case_ ( self : Tuple , _snake_case : str , _snake_case : Any , _snake_case : Optional[int] , _snake_case : Dict , _snake_case : Dict , _snake_case : Union[str, Any] ): __lowercase : Any = self.num_labels __lowercase : List[Any] = DistilBertForSequenceClassification(_snake_case ) model.to(_snake_case ) model.eval() __lowercase : List[str] = model(_snake_case , attention_mask=_snake_case , labels=_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def snake_case_ ( self : Optional[int] , _snake_case : Dict , _snake_case : Any , _snake_case : List[Any] , _snake_case : List[Any] , _snake_case : Dict , _snake_case : List[str] ): __lowercase : str = self.num_labels __lowercase : Optional[int] = DistilBertForTokenClassification(config=_snake_case ) model.to(_snake_case ) model.eval() __lowercase : Dict = model(_snake_case , attention_mask=_snake_case , labels=_snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def snake_case_ ( self : str , _snake_case : Any , _snake_case : int , _snake_case : Dict , _snake_case : int , _snake_case : Optional[Any] , _snake_case : List[str] ): __lowercase : List[Any] = self.num_choices __lowercase : Optional[int] = DistilBertForMultipleChoice(config=_snake_case ) model.to(_snake_case ) model.eval() __lowercase : Tuple = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase : str = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() __lowercase : Tuple = model( _snake_case , attention_mask=_snake_case , labels=_snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def snake_case_ ( self : Optional[Any] ): __lowercase : Dict = self.prepare_config_and_inputs() (__lowercase) : Optional[Any] = config_and_inputs __lowercase : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class __lowerCAmelCase ( __A , __A , unittest.TestCase ): """simple docstring""" A__ : List[str] = ( ( DistilBertModel, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, ) if is_torch_available() else None ) A__ : List[str] = ( { '''feature-extraction''': DistilBertModel, '''fill-mask''': DistilBertForMaskedLM, '''question-answering''': DistilBertForQuestionAnswering, '''text-classification''': DistilBertForSequenceClassification, '''token-classification''': DistilBertForTokenClassification, '''zero-shot''': DistilBertForSequenceClassification, } if is_torch_available() else {} ) A__ : str = True A__ : Tuple = True A__ : List[Any] = True A__ : Optional[Any] = True def snake_case_ ( self : Dict ): __lowercase : List[Any] = DistilBertModelTester(self ) __lowercase : Union[str, Any] = ConfigTester(self , config_class=_snake_case , dim=37 ) def snake_case_ ( self : List[str] ): self.config_tester.run_common_tests() def snake_case_ ( self : str ): __lowercase : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_model(*_snake_case ) def snake_case_ ( self : Any ): __lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_masked_lm(*_snake_case ) def snake_case_ ( self : Any ): __lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_question_answering(*_snake_case ) def snake_case_ ( self : str ): __lowercase : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_sequence_classification(*_snake_case ) def snake_case_ ( self : Union[str, Any] ): __lowercase : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_token_classification(*_snake_case ) def snake_case_ ( self : List[str] ): __lowercase : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_distilbert_for_multiple_choice(*_snake_case ) @slow def snake_case_ ( self : Any ): for model_name in DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __lowercase : Dict = DistilBertModel.from_pretrained(_snake_case ) self.assertIsNotNone(_snake_case ) @slow @require_torch_gpu def snake_case_ ( self : Union[str, Any] ): __lowercase : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # BertForMultipleChoice behaves incorrectly in JIT environments. if model_class == DistilBertForMultipleChoice: return __lowercase : int = True __lowercase : Any = model_class(config=_snake_case ) __lowercase : int = self._prepare_for_class(_snake_case , _snake_case ) __lowercase : Dict = torch.jit.trace( _snake_case , (inputs_dict['''input_ids'''].to('''cpu''' ), inputs_dict['''attention_mask'''].to('''cpu''' )) ) with tempfile.TemporaryDirectory() as tmp: torch.jit.save(_snake_case , os.path.join(_snake_case , '''traced_model.pt''' ) ) __lowercase : Optional[int] = torch.jit.load(os.path.join(_snake_case , '''traced_model.pt''' ) , map_location=_snake_case ) loaded(inputs_dict['''input_ids'''].to(_snake_case ) , inputs_dict['''attention_mask'''].to(_snake_case ) ) @require_torch class __lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @slow def snake_case_ ( self : Optional[int] ): __lowercase : Tuple = DistilBertModel.from_pretrained('''distilbert-base-uncased''' ) __lowercase : Union[str, Any] = torch.tensor([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) __lowercase : int = torch.tensor([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) with torch.no_grad(): __lowercase : Optional[Any] = model(_snake_case , attention_mask=_snake_case )[0] __lowercase : Union[str, Any] = torch.Size((1, 11, 768) ) self.assertEqual(output.shape , _snake_case ) __lowercase : Any = torch.tensor( [[[-0.16_39, 0.32_99, 0.16_48], [-0.17_46, 0.32_89, 0.17_10], [-0.18_84, 0.33_57, 0.18_10]]] ) self.assertTrue(torch.allclose(output[:, 1:4, 1:4] , _snake_case , atol=1E-4 ) )
156
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_convbert import ConvBertTokenizer a_ = logging.get_logger(__name__) a_ = {'vocab_file': 'vocab.txt'} a_ = { 'vocab_file': { 'YituTech/conv-bert-base': 'https://huggingface.co/YituTech/conv-bert-base/resolve/main/vocab.txt', 'YituTech/conv-bert-medium-small': ( 'https://huggingface.co/YituTech/conv-bert-medium-small/resolve/main/vocab.txt' ), 'YituTech/conv-bert-small': 'https://huggingface.co/YituTech/conv-bert-small/resolve/main/vocab.txt', } } a_ = { 'YituTech/conv-bert-base': 512, 'YituTech/conv-bert-medium-small': 512, 'YituTech/conv-bert-small': 512, } a_ = { 'YituTech/conv-bert-base': {'do_lower_case': True}, 'YituTech/conv-bert-medium-small': {'do_lower_case': True}, 'YituTech/conv-bert-small': {'do_lower_case': True}, } class _UpperCamelCase ( __A ): '''simple docstring''' lowerCamelCase__ =VOCAB_FILES_NAMES lowerCamelCase__ =PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ =PRETRAINED_INIT_CONFIGURATION lowerCamelCase__ =PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ =ConvBertTokenizer def __init__( self : List[str] , a : Union[str, Any]=None , a : Optional[int]=None , a : int=True , a : Tuple="[UNK]" , a : Dict="[SEP]" , a : Dict="[PAD]" , a : List[Any]="[CLS]" , a : Tuple="[MASK]" , a : Dict=True , a : Optional[Any]=None , **a : str , ) -> Dict: """simple docstring""" super().__init__( a , tokenizer_file=a , do_lower_case=a , unk_token=a , sep_token=a , pad_token=a , cls_token=a , mask_token=a , tokenize_chinese_chars=a , strip_accents=a , **a , ) SCREAMING_SNAKE_CASE : Optional[Any] = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , a ) != do_lower_case or normalizer_state.get("strip_accents" , a ) != strip_accents or normalizer_state.get("handle_chinese_chars" , a ) != tokenize_chinese_chars ): SCREAMING_SNAKE_CASE : List[str] = getattr(a , normalizer_state.pop("type" ) ) SCREAMING_SNAKE_CASE : Optional[Any] = do_lower_case SCREAMING_SNAKE_CASE : Any = strip_accents SCREAMING_SNAKE_CASE : Optional[int] = tokenize_chinese_chars SCREAMING_SNAKE_CASE : List[str] = normalizer_class(**a ) SCREAMING_SNAKE_CASE : str = do_lower_case def __UpperCamelCase ( self : Union[str, Any] , a : List[Any] , a : int=None ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def __UpperCamelCase ( self : Dict , a : List[int] , a : Optional[List[int]] = None ) -> List[int]: """simple docstring""" SCREAMING_SNAKE_CASE : List[str] = [self.sep_token_id] SCREAMING_SNAKE_CASE : Union[str, Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def __UpperCamelCase ( self : Tuple , a : str , a : Optional[str] = None ) -> Tuple[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self._tokenizer.model.save(a , name=a ) return tuple(a )
76
0
from maths.is_square_free import is_square_free from maths.prime_factors import prime_factors def _snake_case( SCREAMING_SNAKE_CASE__ : Any ) -> str: '''simple docstring''' A__ = prime_factors(_a ) if is_square_free(_a ): return -1 if len(_a ) % 2 else 1 return 0 if __name__ == "__main__": import doctest doctest.testmod()
7
# tests directory-specific settings - this file is run automatically # by pytest before any tests are run import sys import warnings from os.path import abspath, dirname, join # allow having multiple repository checkouts and not needing to remember to rerun # 'pip install -e .[dev]' when switching between checkouts and running tests. a_ = abspath(join(dirname(dirname(__file__)), 'src')) sys.path.insert(1, git_repo_path) # silence FutureWarning warnings in tests since often we can't act on them until # they become normal warnings - i.e. the tests still need to test the current functionality warnings.simplefilter(action='ignore', category=FutureWarning) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_addoption_shared pytest_addoption_shared(_a) def lowerCamelCase__ ( _a): from diffusers.utils.testing_utils import pytest_terminal_summary_main SCREAMING_SNAKE_CASE : Union[str, Any] = terminalreporter.config.getoption("--make-reports") if make_reports: pytest_terminal_summary_main(_a , id=_a)
76
0
# Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from packaging import version from .. import __version__ from .constants import IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD from .doc import ( add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, copy_func, replace_return_docstrings, ) from .generic import ( ContextManagers, ExplicitEnum, ModelOutput, PaddingStrategy, TensorType, add_model_info_to_auto_map, cached_property, can_return_loss, expand_dims, find_labels, flatten_dict, infer_framework, is_jax_tensor, is_numpy_array, is_tensor, is_tf_symbolic_tensor, is_tf_tensor, is_torch_device, is_torch_dtype, is_torch_tensor, reshape, squeeze, strtobool, tensor_size, to_numpy, to_py_obj, transpose, working_or_temp_dir, ) from .hub import ( CLOUDFRONT_DISTRIB_PREFIX, DISABLE_TELEMETRY, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, EntryNotFoundError, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, cached_file, default_cache_path, define_sagemaker_information, download_url, extract_commit_hash, get_cached_models, get_file_from_repo, get_full_repo_name, has_file, http_user_agent, is_offline_mode, is_remote_url, move_cache, send_example_telemetry, try_to_load_from_cache, ) from .import_utils import ( ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, TORCH_FX_REQUIRED_VERSION, USE_JAX, USE_TF, USE_TORCH, DummyObject, OptionalDependencyNotAvailable, _LazyModule, ccl_version, direct_transformers_import, get_torch_version, is_accelerate_available, is_apex_available, is_bitsandbytes_available, is_bsa_available, is_coloredlogs_available, is_cython_available, is_datasets_available, is_decord_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_jieba_available, is_jumanpp_available, is_kenlm_available, is_keras_nlp_available, is_librosa_available, is_natten_available, is_ninja_available, is_onnx_available, is_openai_available, is_optimum_available, is_pandas_available, is_peft_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytest_available, is_pytorch_quantization_available, is_rjieba_available, is_sacremoses_available, is_safetensors_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_sudachi_available, is_tensorflow_probability_available, is_tensorflow_text_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_bfaa_cpu_available, is_torch_bfaa_gpu_available, is_torch_compile_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_neuroncore_available, is_torch_tensorrt_fx_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_torchdistx_available, is_torchdynamo_available, is_torchvision_available, is_training_run_on_sagemaker, is_vision_available, requires_backends, torch_only_method, ) __A : List[Any] = "pytorch_model.bin" __A : str = "pytorch_model.bin.index.json" __A : int = "adapter_config.json" __A : Tuple = "adapter_model.bin" __A : str = "adapter_model.safetensors" __A : Union[str, Any] = "tf_model.h5" __A : Optional[int] = "tf_model.h5.index.json" __A : List[str] = "model.ckpt" __A : Union[str, Any] = "flax_model.msgpack" __A : Optional[int] = "flax_model.msgpack.index.json" __A : str = "model.safetensors" __A : Dict = "model.safetensors.index.json" __A : Any = "config.json" __A : Dict = "preprocessor_config.json" __A : Tuple = FEATURE_EXTRACTOR_NAME __A : List[str] = "generation_config.json" __A : Dict = "modelcard.json" __A : List[str] = "▁" __A : List[Any] = SENTENCEPIECE_UNDERLINE # Kept for backward compatibility __A : str = [ [[0, 1, 0, 1], [1, 0, 0, 1]] ] * 2 # Needs to have 0s and 1s only since XLM uses it for langs too. __A : List[str] = [[7, 6, 0, 0, 1], [1, 2, 3, 0, 0], [0, 0, 0, 4, 5]] __A : Optional[Any] = [[1, 1, 1, 1, 1], [1, 1, 1, 0, 0], [0, 0, 0, 1, 1]] def __SCREAMING_SNAKE_CASE ( UpperCamelCase__ ) -> Optional[int]: '''simple docstring''' if version.parse(_a ) < version.parse(_a ): if "dev" in min_version: UpperCAmelCase = ( "This example requires a source install from HuggingFace Transformers (see " "`https://huggingface.co/docs/transformers/installation#install-from-source`)," ) else: UpperCAmelCase = F"""This example requires a minimum version of {min_version},""" error_message += F""" but the version found is {__version__}.\n""" raise ImportError( error_message + '''Check out https://github.com/huggingface/transformers/tree/main/examples#important-note for the examples corresponding to other ''' '''versions of HuggingFace Transformers.''' )
273
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : Tuple , a : int , a : Optional[int]=13 , a : Optional[int]=3 , a : int=224 , a : Optional[int]=30 , a : int=400 , a : Union[str, Any]=True , a : int=None , a : Tuple=True , a : Tuple=[0.5, 0.5, 0.5] , a : Optional[int]=[0.5, 0.5, 0.5] , ) -> Optional[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : str = size if size is not None else {"height": 18, "width": 18} SCREAMING_SNAKE_CASE : Union[str, Any] = parent SCREAMING_SNAKE_CASE : int = batch_size SCREAMING_SNAKE_CASE : int = num_channels SCREAMING_SNAKE_CASE : Any = image_size SCREAMING_SNAKE_CASE : Tuple = min_resolution SCREAMING_SNAKE_CASE : str = max_resolution SCREAMING_SNAKE_CASE : int = do_resize SCREAMING_SNAKE_CASE : List[Any] = size SCREAMING_SNAKE_CASE : int = do_normalize SCREAMING_SNAKE_CASE : Tuple = image_mean SCREAMING_SNAKE_CASE : Tuple = image_std def __UpperCamelCase ( self : Any ) -> Optional[int]: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCamelCase ( __A , unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =ViTImageProcessor if is_vision_available() else None def __UpperCamelCase ( self : Union[str, Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = EfficientFormerImageProcessorTester(self ) @property def __UpperCamelCase ( self : Any ) -> List[str]: """simple docstring""" return self.image_proc_tester.prepare_image_processor_dict() def __UpperCamelCase ( self : List[Any] ) -> Dict: """simple docstring""" SCREAMING_SNAKE_CASE : List[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a , "image_mean" ) ) self.assertTrue(hasattr(a , "image_std" ) ) self.assertTrue(hasattr(a , "do_normalize" ) ) self.assertTrue(hasattr(a , "do_resize" ) ) self.assertTrue(hasattr(a , "size" ) ) def __UpperCamelCase ( self : int ) -> str: """simple docstring""" pass def __UpperCamelCase ( self : Optional[Any] ) -> List[Any]: """simple docstring""" SCREAMING_SNAKE_CASE : Tuple = self.image_processing_class(**self.image_processor_dict ) # create random PIL images SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a ) for image in image_inputs: self.assertIsInstance(a , Image.Image ) # Test not batched input SCREAMING_SNAKE_CASE : List[str] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : str = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> Any: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors SCREAMING_SNAKE_CASE : int = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , numpify=a ) for image in image_inputs: self.assertIsInstance(a , np.ndarray ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Any = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def __UpperCamelCase ( self : List[str] ) -> List[str]: """simple docstring""" SCREAMING_SNAKE_CASE : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors SCREAMING_SNAKE_CASE : Any = prepare_image_inputs(self.image_proc_tester , equal_resolution=a , torchify=a ) for image in image_inputs: self.assertIsInstance(a , torch.Tensor ) # Test not batched input SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched SCREAMING_SNAKE_CASE : Optional[Any] = image_processor(a , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
76
0
from __future__ import annotations from decimal import Decimal from numpy import array def UpperCamelCase ( _A ): """simple docstring""" __magic_name__ : Union[str, Any] = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(_a ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix __magic_name__ : Tuple = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError("""This matrix has no inverse.""" ) # Creates a copy of the matrix with swapped positions of the elements __magic_name__ : int = [[0.0, 0.0], [0.0, 0.0]] __magic_name__ : Tuple = matrix[1][1], matrix[0][0] __magic_name__ : int = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(_a ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(_a ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule __magic_name__ : Optional[Any] = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError("""This matrix has no inverse.""" ) # Creating cofactor matrix __magic_name__ : List[str] = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] __magic_name__ : Any = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) __magic_name__ : Tuple = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) __magic_name__ : List[str] = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) __magic_name__ : List[Any] = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) __magic_name__ : Optional[int] = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) __magic_name__ : Union[str, Any] = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) __magic_name__ : Dict = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) __magic_name__ : str = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) __magic_name__ : List[Any] = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) __magic_name__ : Tuple = array(_a ) for i in range(3 ): for j in range(3 ): __magic_name__ : Union[str, Any] = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix __magic_name__ : List[str] = array(_a ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(_a ) # Calculate the inverse of the matrix return [[float(d(_a ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError("""Please provide a matrix of size 2x2 or 3x3.""" )
342
import multiprocessing import time from arguments import PretokenizationArguments from datasets import load_dataset from transformers import AutoTokenizer, HfArgumentParser def lowerCamelCase__ ( _a): SCREAMING_SNAKE_CASE : int = {} SCREAMING_SNAKE_CASE : Any = tokenizer(example["content"] , truncation=_a)["input_ids"] SCREAMING_SNAKE_CASE : Dict = len(example["content"]) / len(output["input_ids"]) return output a_ = HfArgumentParser(PretokenizationArguments) a_ = parser.parse_args() if args.num_workers is None: a_ = multiprocessing.cpu_count() a_ = AutoTokenizer.from_pretrained(args.tokenizer_dir) a_ = time.time() a_ = load_dataset(args.dataset_name, split='train') print(F'''Dataset loaded in {time.time()-t_start:.2f}s''') a_ = time.time() a_ = ds.map( tokenize, num_proc=args.num_workers, remove_columns=[ 'repo_name', 'path', 'copies', 'size', 'content', 'license', 'hash', 'line_mean', 'line_max', 'alpha_frac', 'autogenerated', ], ) print(F'''Dataset tokenized in {time.time()-t_start:.2f}s''') a_ = time.time() ds.push_to_hub(args.tokenized_data_repo) print(F'''Data pushed to the hub in {time.time()-t_start:.2f}s''')
76
0
"""simple docstring""" import random def a_ ( _lowerCAmelCase : list , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ , lowercase__ , lowercase__ : List[Any] = [], [], [] for element in data: if element < pivot: less.append(_lowerCAmelCase ) elif element > pivot: greater.append(_lowerCAmelCase ) else: equal.append(_lowerCAmelCase ) return less, equal, greater def a_ ( _lowerCAmelCase : list , _lowerCAmelCase : int ): '''simple docstring''' if index >= len(_lowerCAmelCase ) or index < 0: return None lowercase__ : List[str] = items[random.randint(0 , len(_lowerCAmelCase ) - 1 )] lowercase__ : str = 0 lowercase__ , lowercase__ , lowercase__ : str = _partition(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Tuple = len(_lowerCAmelCase ) lowercase__ : Any = len(_lowerCAmelCase ) # index is the pivot if m <= index < m + count: return pivot # must be in smaller elif m > index: return quick_select(_lowerCAmelCase , _lowerCAmelCase ) # must be in larger else: return quick_select(_lowerCAmelCase , index - (m + count) )
77
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import PIL from PIL import Image from ...utils import ( BaseOutput, OptionalDependencyNotAvailable, is_flax_available, is_k_diffusion_available, is_k_diffusion_version, is_onnx_available, is_torch_available, is_transformers_available, is_transformers_version, ) @dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[List[PIL.Image.Image], np.ndarray] lowerCamelCase__ : Optional[List[bool]] try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_cycle_diffusion import CycleDiffusionPipeline from .pipeline_stable_diffusion import StableDiffusionPipeline from .pipeline_stable_diffusion_attend_and_excite import StableDiffusionAttendAndExcitePipeline from .pipeline_stable_diffusion_imgaimg import StableDiffusionImgaImgPipeline from .pipeline_stable_diffusion_inpaint import StableDiffusionInpaintPipeline from .pipeline_stable_diffusion_inpaint_legacy import StableDiffusionInpaintPipelineLegacy from .pipeline_stable_diffusion_instruct_pixapix import StableDiffusionInstructPixaPixPipeline from .pipeline_stable_diffusion_latent_upscale import StableDiffusionLatentUpscalePipeline from .pipeline_stable_diffusion_ldmad import StableDiffusionLDMaDPipeline from .pipeline_stable_diffusion_model_editing import StableDiffusionModelEditingPipeline from .pipeline_stable_diffusion_panorama import StableDiffusionPanoramaPipeline from .pipeline_stable_diffusion_paradigms import StableDiffusionParadigmsPipeline from .pipeline_stable_diffusion_sag import StableDiffusionSAGPipeline from .pipeline_stable_diffusion_upscale import StableDiffusionUpscalePipeline from .pipeline_stable_unclip import StableUnCLIPPipeline from .pipeline_stable_unclip_imgaimg import StableUnCLIPImgaImgPipeline from .safety_checker import StableDiffusionSafetyChecker from .stable_unclip_image_normalizer import StableUnCLIPImageNormalizer try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.25.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import StableDiffusionImageVariationPipeline else: from .pipeline_stable_diffusion_image_variation import StableDiffusionImageVariationPipeline try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(">=", "4.26.0")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ( StableDiffusionDepthaImgPipeline, StableDiffusionDiffEditPipeline, StableDiffusionPixaPixZeroPipeline, ) else: from .pipeline_stable_diffusion_depthaimg import StableDiffusionDepthaImgPipeline from .pipeline_stable_diffusion_diffedit import StableDiffusionDiffEditPipeline from .pipeline_stable_diffusion_pixapix_zero import StableDiffusionPixaPixZeroPipeline try: if not ( is_torch_available() and is_transformers_available() and is_k_diffusion_available() and is_k_diffusion_version(">=", "0.0.12") ): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_and_k_diffusion_objects import * # noqa F403 else: from .pipeline_stable_diffusion_k_diffusion import StableDiffusionKDiffusionPipeline try: if not (is_transformers_available() and is_onnx_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_onnx_objects import * # noqa F403 else: from .pipeline_onnx_stable_diffusion import OnnxStableDiffusionPipeline, StableDiffusionOnnxPipeline from .pipeline_onnx_stable_diffusion_imgaimg import OnnxStableDiffusionImgaImgPipeline from .pipeline_onnx_stable_diffusion_inpaint import OnnxStableDiffusionInpaintPipeline from .pipeline_onnx_stable_diffusion_inpaint_legacy import OnnxStableDiffusionInpaintPipelineLegacy from .pipeline_onnx_stable_diffusion_upscale import OnnxStableDiffusionUpscalePipeline if is_transformers_available() and is_flax_available(): import flax @flax.struct.dataclass class UpperCAmelCase_ ( _a): lowerCamelCase__ : np.ndarray lowerCamelCase__ : List[bool] from ...schedulers.scheduling_pndm_flax import PNDMSchedulerState from .pipeline_flax_stable_diffusion import FlaxStableDiffusionPipeline from .pipeline_flax_stable_diffusion_imgaimg import FlaxStableDiffusionImgaImgPipeline from .pipeline_flax_stable_diffusion_inpaint import FlaxStableDiffusionInpaintPipeline from .safety_checker_flax import FlaxStableDiffusionSafetyChecker
77
1
"""simple docstring""" import math class UpperCAmelCase_ : def _UpperCAmelCase ( self , a , a ) -> int: lowercase__ : Tuple = 0.0 lowercase__ : str = 0.0 for i in range(len(a ) ): da += math.pow((sample[i] - weights[0][i]) , 2 ) da += math.pow((sample[i] - weights[1][i]) , 2 ) return 0 if da > da else 1 return 0 def _UpperCAmelCase ( self , a , a , a , a ) -> list[list[int | float]]: for i in range(len(a ) ): weights[j][i] += alpha * (sample[i] - weights[j][i]) return weights def a_ ( ): '''simple docstring''' lowercase__ : str = [[1, 1, 0, 0], [0, 0, 0, 1], [1, 0, 0, 0], [0, 0, 1, 1]] # weight initialization ( n, C ) lowercase__ : int = [[0.2, 0.6, 0.5, 0.9], [0.8, 0.4, 0.7, 0.3]] # training lowercase__ : Optional[int] = SelfOrganizingMap() lowercase__ : Dict = 3 lowercase__ : List[Any] = 0.5 for _ in range(_lowerCAmelCase ): for j in range(len(_lowerCAmelCase ) ): # training sample lowercase__ : Tuple = training_samples[j] # Compute the winning vector lowercase__ : str = self_organizing_map.get_winner(_lowerCAmelCase , _lowerCAmelCase ) # Update the winning vector lowercase__ : List[Any] = self_organizing_map.update(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # classify test sample lowercase__ : int = [0, 0, 0, 1] lowercase__ : Union[str, Any] = self_organizing_map.get_winner(_lowerCAmelCase , _lowerCAmelCase ) # results print(f"""Clusters that the test sample belongs to : {winner}""" ) print(f"""Weights that have been trained : {weights}""" ) # running the main() function if __name__ == "__main__": main()
77
"""simple docstring""" import pickle import numpy as np from matplotlib import pyplot as plt class UpperCAmelCase_ : def __init__( self , a , a , a , a , a , a=0.2 , a=0.2 ) -> Dict: lowercase__ : Any = bp_numa lowercase__ : Optional[int] = bp_numa lowercase__ : Tuple = bp_numa lowercase__ : Optional[Any] = conva_get[:2] lowercase__ : Optional[int] = conva_get[2] lowercase__ : Optional[Any] = size_pa lowercase__ : Union[str, Any] = rate_w lowercase__ : Union[str, Any] = rate_t lowercase__ : List[Any] = [ np.mat(-1 * np.random.rand(self.conva[0] , self.conva[0] ) + 0.5 ) for i in range(self.conva[1] ) ] lowercase__ : Optional[Any] = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : int = np.mat(-1 * np.random.rand(self.num_bpa , self.num_bpa ) + 0.5 ) lowercase__ : Any = -2 * np.random.rand(self.conva[1] ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 lowercase__ : int = -2 * np.random.rand(self.num_bpa ) + 1 def _UpperCAmelCase ( self , a ) -> Union[str, Any]: # save model dict with pickle lowercase__ : Optional[Any] = { 'num_bp1': self.num_bpa, 'num_bp2': self.num_bpa, 'num_bp3': self.num_bpa, 'conv1': self.conva, 'step_conv1': self.step_conva, 'size_pooling1': self.size_poolinga, 'rate_weight': self.rate_weight, 'rate_thre': self.rate_thre, 'w_conv1': self.w_conva, 'wkj': self.wkj, 'vji': self.vji, 'thre_conv1': self.thre_conva, 'thre_bp2': self.thre_bpa, 'thre_bp3': self.thre_bpa, } with open(a , 'wb' ) as f: pickle.dump(a , a ) print(f"""Model saved: {save_path}""" ) @classmethod def _UpperCAmelCase ( cls , a ) -> Any: # read saved model with open(a , 'rb' ) as f: lowercase__ : Optional[int] = pickle.load(a ) # noqa: S301 lowercase__ : Optional[int] = model_dic.get('conv1' ) conv_get.append(model_dic.get('step_conv1' ) ) lowercase__ : List[Any] = model_dic.get('size_pooling1' ) lowercase__ : Tuple = model_dic.get('num_bp1' ) lowercase__ : int = model_dic.get('num_bp2' ) lowercase__ : int = model_dic.get('num_bp3' ) lowercase__ : Union[str, Any] = model_dic.get('rate_weight' ) lowercase__ : Tuple = model_dic.get('rate_thre' ) # create model instance lowercase__ : Tuple = CNN(a , a , a , a , a , a , a ) # modify model parameter lowercase__ : str = model_dic.get('w_conv1' ) lowercase__ : Optional[int] = model_dic.get('wkj' ) lowercase__ : Tuple = model_dic.get('vji' ) lowercase__ : str = model_dic.get('thre_conv1' ) lowercase__ : Union[str, Any] = model_dic.get('thre_bp2' ) lowercase__ : List[str] = model_dic.get('thre_bp3' ) return conv_ins def _UpperCAmelCase ( self , a ) -> str: return 1 / (1 + np.exp(-1 * x )) def _UpperCAmelCase ( self , a ) -> Any: return round(a , 3 ) def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[str]: # convolution process lowercase__ : int = convs[0] lowercase__ : Optional[Any] = convs[1] lowercase__ : int = np.shape(a )[0] # get the data slice of original image data, data_focus lowercase__ : Optional[Any] = [] for i_focus in range(0 , size_data - size_conv + 1 , a ): for j_focus in range(0 , size_data - size_conv + 1 , a ): lowercase__ : Optional[int] = data[ i_focus : i_focus + size_conv, j_focus : j_focus + size_conv ] data_focus.append(a ) # calculate the feature map of every single kernel, and saved as list of matrix lowercase__ : Union[str, Any] = [] lowercase__ : Dict = int((size_data - size_conv) / conv_step + 1 ) for i_map in range(a ): lowercase__ : Any = [] for i_focus in range(len(a ) ): lowercase__ : Tuple = ( np.sum(np.multiply(data_focus[i_focus] , w_convs[i_map] ) ) - thre_convs[i_map] ) featuremap.append(self.sig(a ) ) lowercase__ : Optional[Any] = np.asmatrix(a ).reshape( a , a ) data_featuremap.append(a ) # expanding the data slice to One dimenssion lowercase__ : str = [] for each_focus in data_focus: focusa_list.extend(self.Expand_Mat(a ) ) lowercase__ : int = np.asarray(a ) return focus_list, data_featuremap def _UpperCAmelCase ( self , a , a , a="average_pool" ) -> str: # pooling process lowercase__ : List[str] = len(featuremaps[0] ) lowercase__ : List[str] = int(size_map / size_pooling ) lowercase__ : str = [] for i_map in range(len(a ) ): lowercase__ : List[str] = featuremaps[i_map] lowercase__ : Optional[int] = [] for i_focus in range(0 , a , a ): for j_focus in range(0 , a , a ): lowercase__ : List[Any] = feature_map[ i_focus : i_focus + size_pooling, j_focus : j_focus + size_pooling, ] if pooling_type == "average_pool": # average pooling map_pooled.append(np.average(a ) ) elif pooling_type == "max_pooling": # max pooling map_pooled.append(np.max(a ) ) lowercase__ : List[Any] = np.asmatrix(a ).reshape(a , a ) featuremap_pooled.append(a ) return featuremap_pooled def _UpperCAmelCase ( self , a ) -> List[str]: # expanding three dimension data to one dimension list lowercase__ : Any = [] for i in range(len(a ) ): lowercase__ : Optional[int] = np.shape(data[i] ) lowercase__ : int = data[i].reshape(1 , shapes[0] * shapes[1] ) lowercase__ : str = data_listed.getA().tolist()[0] data_expanded.extend(a ) lowercase__ : int = np.asarray(a ) return data_expanded def _UpperCAmelCase ( self , a ) -> Dict: # expanding matrix to one dimension list lowercase__ : Dict = np.asarray(a ) lowercase__ : Union[str, Any] = np.shape(a ) lowercase__ : Optional[Any] = data_mat.reshape(1 , shapes[0] * shapes[1] ) return data_expanded def _UpperCAmelCase ( self , a , a , a , a , a ) -> List[Any]: lowercase__ : Dict = [] lowercase__ : int = 0 for i_map in range(a ): lowercase__ : str = np.ones((size_map, size_map) ) for i in range(0 , a , a ): for j in range(0 , a , a ): lowercase__ : Optional[Any] = pd_pool[ i_pool ] lowercase__ : Union[str, Any] = i_pool + 1 lowercase__ : List[Any] = np.multiply( a , np.multiply(out_map[i_map] , (1 - out_map[i_map]) ) ) pd_all.append(a ) return pd_all def _UpperCAmelCase ( self , a , a , a , a , a , a=bool ) -> str: # model traning print('----------------------Start Training-------------------------' ) print((' - - Shape: Train_Data ', np.shape(a )) ) print((' - - Shape: Teach_Data ', np.shape(a )) ) lowercase__ : int = 0 lowercase__ : List[Any] = [] lowercase__ : Union[str, Any] = 1_0_0_0_0 while rp < n_repeat and mse >= error_accuracy: lowercase__ : List[Any] = 0 print(f"""-------------Learning Time {rp}--------------""" ) for p in range(len(a ) ): # print('------------Learning Image: %d--------------'%p) lowercase__ : Optional[int] = np.asmatrix(datas_train[p] ) lowercase__ : int = np.asarray(datas_teach[p] ) lowercase__ , lowercase__ : Union[str, Any] = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Optional[Any] = self.pooling(a , self.size_poolinga ) lowercase__ : Tuple = np.shape(a ) lowercase__ : List[str] = self._expand(a ) lowercase__ : Optional[int] = data_bp_input lowercase__ : Optional[Any] = np.dot(a , self.vji.T ) - self.thre_bpa lowercase__ : str = self.sig(a ) lowercase__ : Tuple = np.dot(a , self.wkj.T ) - self.thre_bpa lowercase__ : Any = self.sig(a ) # --------------Model Leaning ------------------------ # calculate error and gradient--------------- lowercase__ : int = np.multiply( (data_teach - bp_outa) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Any = np.multiply( np.dot(a , self.wkj ) , np.multiply(a , (1 - bp_outa) ) ) lowercase__ : Optional[int] = np.dot(a , self.vji ) lowercase__ : Union[str, Any] = pd_i_all / (self.size_poolinga * self.size_poolinga) lowercase__ : Any = pd_conva_pooled.T.getA().tolist() lowercase__ : List[str] = self._calculate_gradient_from_pool( a , a , shape_featuremapa[0] , shape_featuremapa[1] , self.size_poolinga , ) # weight and threshold learning process--------- # convolution layer for k_conv in range(self.conva[1] ): lowercase__ : Optional[int] = self._expand_mat(pd_conva_all[k_conv] ) lowercase__ : Tuple = self.rate_weight * np.dot(a , a ) lowercase__ : Union[str, Any] = self.w_conva[k_conv] + delta_w.reshape( (self.conva[0], self.conva[0]) ) lowercase__ : Any = ( self.thre_conva[k_conv] - np.sum(pd_conva_all[k_conv] ) * self.rate_thre ) # all connected layer lowercase__ : Tuple = self.wkj + pd_k_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.vji + pd_j_all.T * bp_outa * self.rate_weight lowercase__ : Tuple = self.thre_bpa - pd_k_all * self.rate_thre lowercase__ : Optional[Any] = self.thre_bpa - pd_j_all * self.rate_thre # calculate the sum error of all single image lowercase__ : Dict = np.sum(abs(data_teach - bp_outa ) ) error_count += errors # print(' ----Teach ',data_teach) # print(' ----BP_output ',bp_out3) lowercase__ : str = rp + 1 lowercase__ : List[str] = error_count / patterns all_mse.append(a ) def draw_error(): lowercase__ : Any = [error_accuracy for i in range(int(n_repeat * 1.2 ) )] plt.plot(a , '+-' ) plt.plot(a , 'r--' ) plt.xlabel('Learning Times' ) plt.ylabel('All_mse' ) plt.grid(a , alpha=0.5 ) plt.show() print('------------------Training Complished---------------------' ) print((' - - Training epoch: ', rp, f""" - - Mse: {mse:.6f}""") ) if draw_e: draw_error() return mse def _UpperCAmelCase ( self , a ) -> List[Any]: # model predict lowercase__ : Optional[int] = [] print('-------------------Start Testing-------------------------' ) print((' - - Shape: Test_Data ', np.shape(a )) ) for p in range(len(a ) ): lowercase__ : List[str] = np.asmatrix(datas_test[p] ) lowercase__ , lowercase__ : Tuple = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Any = self.pooling(a , self.size_poolinga ) lowercase__ : Union[str, Any] = self._expand(a ) lowercase__ : Optional[Any] = data_bp_input lowercase__ : str = bp_outa * self.vji.T - self.thre_bpa lowercase__ : Optional[Any] = self.sig(a ) lowercase__ : Dict = bp_outa * self.wkj.T - self.thre_bpa lowercase__ : List[str] = self.sig(a ) produce_out.extend(bp_outa.getA().tolist() ) lowercase__ : Optional[int] = [list(map(self.do_round , a ) ) for each in produce_out] return np.asarray(a ) def _UpperCAmelCase ( self , a ) -> List[str]: # return the data of image after convoluting process so we can check it out lowercase__ : Any = np.asmatrix(a ) lowercase__ , lowercase__ : str = self.convolute( a , self.conva , self.w_conva , self.thre_conva , conv_step=self.step_conva , ) lowercase__ : Tuple = self.pooling(a , self.size_poolinga ) return data_conveda, data_pooleda if __name__ == "__main__": pass
77
1
"""simple docstring""" _UpperCamelCase : Union[str, Any] = [ [0, 16, 13, 0, 0, 0], [0, 0, 10, 12, 0, 0], [0, 4, 0, 0, 14, 0], [0, 0, 9, 0, 0, 20], [0, 0, 0, 7, 0, 4], [0, 0, 0, 0, 0, 0], ] def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Optional[int] = [False] * len(_lowerCAmelCase ) lowercase__ : List[str] = [s] lowercase__ : List[Any] = True while queue: lowercase__ : int = queue.pop(0 ) for ind in range(len(graph[u] ) ): if visited[ind] is False and graph[u][ind] > 0: queue.append(_lowerCAmelCase ) lowercase__ : Union[str, Any] = True lowercase__ : Union[str, Any] = u return visited[t] def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Any = [-1] * (len(_lowerCAmelCase )) lowercase__ : str = 0 lowercase__ : int = [] lowercase__ : Tuple = [i[:] for i in graph] # Record original cut, copy. while bfs(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): lowercase__ : Union[str, Any] = float('Inf' ) lowercase__ : Tuple = sink while s != source: # Find the minimum value in select path lowercase__ : Optional[Any] = min(_lowerCAmelCase , graph[parent[s]][s] ) lowercase__ : List[Any] = parent[s] max_flow += path_flow lowercase__ : List[Any] = sink while v != source: lowercase__ : List[str] = parent[v] graph[u][v] -= path_flow graph[v][u] += path_flow lowercase__ : Optional[int] = parent[v] for i in range(len(_lowerCAmelCase ) ): for j in range(len(graph[0] ) ): if graph[i][j] == 0 and temp[i][j] > 0: res.append((i, j) ) return res if __name__ == "__main__": print(mincut(test_graph, source=0, sink=5))
77
"""simple docstring""" from collections.abc import Generator def a_ ( ): '''simple docstring''' lowercase__ , lowercase__ : List[str] = 0, 1 while True: lowercase__ , lowercase__ : Optional[int] = b, a + b yield b def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : List[Any] = 1 lowercase__ : Any = fibonacci_generator() while len(str(next(_lowerCAmelCase ) ) ) < n: answer += 1 return answer + 1 if __name__ == "__main__": print(solution(int(str(input()).strip())))
77
1
"""simple docstring""" import torch from diffusers import DDPMScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[Any] = (DDPMScheduler,) def _UpperCAmelCase ( self , **a ) -> str: lowercase__ : int = { 'num_train_timesteps': 1_0_0_0, 'beta_start': 0.0_001, 'beta_end': 0.02, 'beta_schedule': 'linear', 'variance_type': 'fixed_small', 'clip_sample': True, } config.update(**a ) return config def _UpperCAmelCase ( self ) -> Any: for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=a ) def _UpperCAmelCase ( self ) -> Dict: for beta_start, beta_end in zip([0.0_001, 0.001, 0.01, 0.1] , [0.002, 0.02, 0.2, 2] ): self.check_over_configs(beta_start=a , beta_end=a ) def _UpperCAmelCase ( self ) -> Any: for schedule in ["linear", "squaredcos_cap_v2"]: self.check_over_configs(beta_schedule=a ) def _UpperCAmelCase ( self ) -> str: for variance in ["fixed_small", "fixed_large", "other"]: self.check_over_configs(variance_type=a ) def _UpperCAmelCase ( self ) -> Any: for clip_sample in [True, False]: self.check_over_configs(clip_sample=a ) def _UpperCAmelCase ( self ) -> Optional[int]: self.check_over_configs(thresholding=a ) for threshold in [0.5, 1.0, 2.0]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs( thresholding=a , prediction_type=a , sample_max_value=a , ) def _UpperCAmelCase ( self ) -> List[str]: for prediction_type in ["epsilon", "sample", "v_prediction"]: self.check_over_configs(prediction_type=a ) def _UpperCAmelCase ( self ) -> List[str]: for t in [0, 5_0_0, 9_9_9]: self.check_over_forward(time_step=a ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Optional[Any] = self.scheduler_classes[0] lowercase__ : List[Any] = self.get_scheduler_config() lowercase__ : Union[str, Any] = scheduler_class(**a ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 0.0 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.00_979 ) ) < 1e-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.02 ) ) < 1e-5 def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.scheduler_classes[0] lowercase__ : Optional[Any] = self.get_scheduler_config() lowercase__ : Union[str, Any] = scheduler_class(**a ) lowercase__ : List[Any] = len(a ) lowercase__ : List[Any] = self.dummy_model() lowercase__ : str = self.dummy_sample_deter lowercase__ : List[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowercase__ : Dict = model(a , a ) # 2. predict previous mean of sample x_t-1 lowercase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance lowercase__ : List[str] = pred_prev_sample lowercase__ : Any = torch.sum(torch.abs(a ) ) lowercase__ : str = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 258.9_606 ) < 1e-2 assert abs(result_mean.item() - 0.3_372 ) < 1e-3 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Any = self.scheduler_classes[0] lowercase__ : Optional[int] = self.get_scheduler_config(prediction_type='v_prediction' ) lowercase__ : Optional[Any] = scheduler_class(**a ) lowercase__ : int = len(a ) lowercase__ : Dict = self.dummy_model() lowercase__ : Optional[int] = self.dummy_sample_deter lowercase__ : List[Any] = torch.manual_seed(0 ) for t in reversed(range(a ) ): # 1. predict noise residual lowercase__ : Optional[Any] = model(a , a ) # 2. predict previous mean of sample x_t-1 lowercase__ : Optional[int] = scheduler.step(a , a , a , generator=a ).prev_sample # if t > 0: # noise = self.dummy_sample_deter # variance = scheduler.get_variance(t) ** (0.5) * noise # # sample = pred_prev_sample + variance lowercase__ : List[Any] = pred_prev_sample lowercase__ : Dict = torch.sum(torch.abs(a ) ) lowercase__ : List[str] = torch.mean(torch.abs(a ) ) assert abs(result_sum.item() - 202.0_296 ) < 1e-2 assert abs(result_mean.item() - 0.2_631 ) < 1e-3 def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Optional[Any] = self.scheduler_classes[0] lowercase__ : Any = self.get_scheduler_config() lowercase__ : List[Any] = scheduler_class(**a ) lowercase__ : Any = [1_0_0, 8_7, 5_0, 1, 0] scheduler.set_timesteps(timesteps=a ) lowercase__ : List[Any] = scheduler.timesteps for i, timestep in enumerate(a ): if i == len(a ) - 1: lowercase__ : Union[str, Any] = -1 else: lowercase__ : int = timesteps[i + 1] lowercase__ : Tuple = scheduler.previous_timestep(a ) lowercase__ : List[Any] = prev_t.item() self.assertEqual(a , a ) def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Union[str, Any] = self.scheduler_classes[0] lowercase__ : List[str] = self.get_scheduler_config() lowercase__ : Any = scheduler_class(**a ) lowercase__ : Optional[Any] = [1_0_0, 8_7, 5_0, 5_1, 0] with self.assertRaises(a , msg='`custom_timesteps` must be in descending order.' ): scheduler.set_timesteps(timesteps=a ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Tuple = self.scheduler_classes[0] lowercase__ : int = self.get_scheduler_config() lowercase__ : int = scheduler_class(**a ) lowercase__ : Dict = [1_0_0, 8_7, 5_0, 1, 0] lowercase__ : Dict = len(a ) with self.assertRaises(a , msg='Can only pass one of `num_inference_steps` or `custom_timesteps`.' ): scheduler.set_timesteps(num_inference_steps=a , timesteps=a ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Union[str, Any] = self.scheduler_classes[0] lowercase__ : str = self.get_scheduler_config() lowercase__ : List[Any] = scheduler_class(**a ) lowercase__ : Union[str, Any] = [scheduler.config.num_train_timesteps] with self.assertRaises( a , msg='`timesteps` must start before `self.config.train_timesteps`: {scheduler.config.num_train_timesteps}}' , ): scheduler.set_timesteps(timesteps=a )
77
"""simple docstring""" import baseaa import io import json import os from copy import deepcopy from ..optimizer import AcceleratedOptimizer from ..scheduler import AcceleratedScheduler class UpperCAmelCase_ : def __init__( self , a ) -> List[str]: if isinstance(a , a ): # Don't modify user's data should they want to reuse it (e.g. in tests), because once we # modified it, it will not be accepted here again, since `auto` values would have been overridden lowercase__ : Optional[Any] = deepcopy(a ) elif os.path.exists(a ): with io.open(a , 'r' , encoding='utf-8' ) as f: lowercase__ : List[Any] = json.load(a ) else: try: lowercase__ : Optional[int] = baseaa.urlsafe_baadecode(a ).decode('utf-8' ) lowercase__ : List[str] = json.loads(a ) except (UnicodeDecodeError, AttributeError, ValueError): raise ValueError( f"""Expected a string path to an existing deepspeed config, or a dictionary, or a base64 encoded string. Received: {config_file_or_dict}""" ) lowercase__ : Any = config self.set_stage_and_offload() def _UpperCAmelCase ( self ) -> Dict: # zero stage - this is done as early as possible, before model is created, to allow # ``is_deepspeed_zero3_enabled`` query and getting to the early deepspeed config object # during ``zero.Init()`` which needs to know the dtype, and some other hparams. lowercase__ : Tuple = self.get_value('zero_optimization.stage' , -1 ) # offload lowercase__ : int = False if self.is_zeroa() or self.is_zeroa(): lowercase__ : str = set(['cpu', 'nvme'] ) lowercase__ : Optional[Any] = set( [ self.get_value('zero_optimization.offload_optimizer.device' ), self.get_value('zero_optimization.offload_param.device' ), ] ) if len(offload_devices & offload_devices_valid ) > 0: lowercase__ : Optional[Any] = True def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Dict = self.config # find the config node of interest if it exists lowercase__ : int = ds_key_long.split('.' ) lowercase__ : Dict = nodes.pop() for node in nodes: lowercase__ : Optional[Any] = config.get(a ) if config is None: return None, ds_key return config, ds_key def _UpperCAmelCase ( self , a , a=None ) -> Union[str, Any]: lowercase__ , lowercase__ : Tuple = self.find_config_node(a ) if config is None: return default return config.get(a , a ) def _UpperCAmelCase ( self , a , a=False ) -> Any: lowercase__ : str = self.config # find the config node of interest if it exists lowercase__ : List[Any] = ds_key_long.split('.' ) for node in nodes: lowercase__ : str = config lowercase__ : str = config.get(a ) if config is None: if must_exist: raise ValueError(f"""Can't find {ds_key_long} entry in the config: {self.config}""" ) else: return # if found remove it if parent_config is not None: parent_config.pop(a ) def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : Union[str, Any] = self.get_value(a ) return False if value is None else bool(a ) def _UpperCAmelCase ( self , a ) -> Any: lowercase__ : Any = self.get_value(a ) return False if value is None else not bool(a ) def _UpperCAmelCase ( self ) -> Tuple: return self._stage == 2 def _UpperCAmelCase ( self ) -> List[Any]: return self._stage == 3 def _UpperCAmelCase ( self ) -> str: return self._offload class UpperCAmelCase_ : def __init__( self , a ) -> str: lowercase__ : Tuple = engine def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: # runs backpropagation and handles mixed precision self.engine.backward(a , **a ) # Deepspeed's `engine.step` performs the following operations: # - gradient accumulation check # - gradient clipping # - optimizer step # - zero grad # - checking overflow # - lr_scheduler step (only if engine.lr_scheduler is not None) self.engine.step() # and this plugin overrides the above calls with no-ops when Accelerate runs under # Deepspeed, but allows normal functionality for non-Deepspeed cases thus enabling a simple # training loop that works transparently under many training regimes. class UpperCAmelCase_ ( _a): def __init__( self , a ) -> Dict: super().__init__(a , device_placement=a , scaler=a ) lowercase__ : Union[str, Any] = hasattr(self.optimizer , 'overflow' ) def _UpperCAmelCase ( self , a=None ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed def _UpperCAmelCase ( self ) -> Optional[int]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed @property def _UpperCAmelCase ( self ) -> Tuple: if self.__has_overflow__: return self.optimizer.overflow return False class UpperCAmelCase_ ( _a): def __init__( self , a , a ) -> Any: super().__init__(a , a ) def _UpperCAmelCase ( self ) -> List[Any]: pass # `accelerator.backward(loss)` is doing that automatically. Therefore, its implementation is not needed class UpperCAmelCase_ : def __init__( self , a , a=0.001 , a=0 , **a ) -> Tuple: lowercase__ : List[Any] = params lowercase__ : int = lr lowercase__ : int = weight_decay lowercase__ : Union[str, Any] = kwargs class UpperCAmelCase_ : def __init__( self , a , a=None , a=0 , **a ) -> Tuple: lowercase__ : Dict = optimizer lowercase__ : List[str] = total_num_steps lowercase__ : Optional[int] = warmup_num_steps lowercase__ : List[Any] = kwargs
77
1
"""simple docstring""" import tempfile import torch from diffusers import IPNDMScheduler from .test_schedulers import SchedulerCommonTest class UpperCAmelCase_ ( _a): lowerCamelCase__ : int = (IPNDMScheduler,) lowerCamelCase__ : str = (("num_inference_steps", 5_0),) def _UpperCAmelCase ( self , **a ) -> List[str]: lowercase__ : int = {'num_train_timesteps': 1_0_0_0} config.update(**a ) return config def _UpperCAmelCase ( self , a=0 , **a ) -> Union[str, Any]: lowercase__ : Tuple = dict(self.forward_default_kwargs ) lowercase__ : List[Any] = kwargs.pop('num_inference_steps' , a ) lowercase__ : Dict = self.dummy_sample lowercase__ : Union[str, Any] = 0.1 * sample lowercase__ : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowercase__ : List[str] = self.get_scheduler_config(**a ) lowercase__ : Union[str, Any] = scheduler_class(**a ) scheduler.set_timesteps(a ) # copy over dummy past residuals lowercase__ : Optional[int] = dummy_past_residuals[:] if time_step is None: lowercase__ : Optional[int] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a ) lowercase__ : int = scheduler_class.from_pretrained(a ) new_scheduler.set_timesteps(a ) # copy over dummy past residuals lowercase__ : Tuple = dummy_past_residuals[:] lowercase__ : Optional[Any] = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : List[Any] = new_scheduler.step(a , a , a , **a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" lowercase__ : Tuple = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : List[str] = new_scheduler.step(a , a , a , **a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def _UpperCAmelCase ( self ) -> Tuple: pass def _UpperCAmelCase ( self , a=0 , **a ) -> List[Any]: lowercase__ : List[str] = dict(self.forward_default_kwargs ) lowercase__ : List[str] = kwargs.pop('num_inference_steps' , a ) lowercase__ : Union[str, Any] = self.dummy_sample lowercase__ : Tuple = 0.1 * sample lowercase__ : Optional[Any] = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] for scheduler_class in self.scheduler_classes: lowercase__ : str = self.get_scheduler_config() lowercase__ : Union[str, Any] = scheduler_class(**a ) scheduler.set_timesteps(a ) # copy over dummy past residuals (must be after setting timesteps) lowercase__ : Optional[Any] = dummy_past_residuals[:] if time_step is None: lowercase__ : List[Any] = scheduler.timesteps[len(scheduler.timesteps ) // 2] with tempfile.TemporaryDirectory() as tmpdirname: scheduler.save_config(a ) lowercase__ : Optional[Any] = scheduler_class.from_pretrained(a ) # copy over dummy past residuals new_scheduler.set_timesteps(a ) # copy over dummy past residual (must be after setting timesteps) lowercase__ : Union[str, Any] = dummy_past_residuals[:] lowercase__ : Optional[Any] = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : int = new_scheduler.step(a , a , a , **a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" lowercase__ : List[str] = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : int = new_scheduler.step(a , a , a , **a ).prev_sample assert torch.sum(torch.abs(output - new_output ) ) < 1e-5, "Scheduler outputs are not identical" def _UpperCAmelCase ( self , **a ) -> Tuple: lowercase__ : Union[str, Any] = self.scheduler_classes[0] lowercase__ : Tuple = self.get_scheduler_config(**a ) lowercase__ : str = scheduler_class(**a ) lowercase__ : Dict = 1_0 lowercase__ : str = self.dummy_model() lowercase__ : Optional[Any] = self.dummy_sample_deter scheduler.set_timesteps(a ) for i, t in enumerate(scheduler.timesteps ): lowercase__ : Any = model(a , a ) lowercase__ : Union[str, Any] = scheduler.step(a , a , a ).prev_sample for i, t in enumerate(scheduler.timesteps ): lowercase__ : str = model(a , a ) lowercase__ : Tuple = scheduler.step(a , a , a ).prev_sample return sample def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : List[Any] = dict(self.forward_default_kwargs ) lowercase__ : Tuple = kwargs.pop('num_inference_steps' , a ) for scheduler_class in self.scheduler_classes: lowercase__ : List[Any] = self.get_scheduler_config() lowercase__ : Optional[int] = scheduler_class(**a ) lowercase__ : List[Any] = self.dummy_sample lowercase__ : int = 0.1 * sample if num_inference_steps is not None and hasattr(a , 'set_timesteps' ): scheduler.set_timesteps(a ) elif num_inference_steps is not None and not hasattr(a , 'set_timesteps' ): lowercase__ : Any = num_inference_steps # copy over dummy past residuals (must be done after set_timesteps) lowercase__ : Any = [residual + 0.2, residual + 0.15, residual + 0.1, residual + 0.05] lowercase__ : Union[str, Any] = dummy_past_residuals[:] lowercase__ : Any = scheduler.timesteps[5] lowercase__ : str = scheduler.timesteps[6] lowercase__ : Optional[int] = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : Union[str, Any] = scheduler.step(a , a , a , **a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) lowercase__ : Union[str, Any] = scheduler.step(a , a , a , **a ).prev_sample lowercase__ : Optional[Any] = scheduler.step(a , a , a , **a ).prev_sample self.assertEqual(output_a.shape , sample.shape ) self.assertEqual(output_a.shape , output_a.shape ) def _UpperCAmelCase ( self ) -> Optional[int]: for timesteps in [1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=a , time_step=a ) def _UpperCAmelCase ( self ) -> Any: for t, num_inference_steps in zip([1, 5, 1_0] , [1_0, 5_0, 1_0_0] ): self.check_over_forward(num_inference_steps=a , time_step=a ) def _UpperCAmelCase ( self ) -> Dict: lowercase__ : List[str] = self.full_loop() lowercase__ : Tuple = torch.mean(torch.abs(a ) ) assert abs(result_mean.item() - 2_5_4_0_5_2_9 ) < 1_0
77
"""simple docstring""" from typing import List, Union from ..utils import ( add_end_docstrings, is_tf_available, is_torch_available, is_vision_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_tf_available(): import tensorflow as tf from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING from ..tf_utils import stable_softmax if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING _UpperCamelCase : int = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> Union[str, Any]: super().__init__(*a , **a ) requires_backends(self , 'vision' ) self.check_model_type( TF_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING if self.framework == 'tf' else MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING ) def _UpperCAmelCase ( self , a=None ) -> Dict: lowercase__ : Any = {} if top_k is not None: lowercase__ : List[str] = top_k return {}, {}, postprocess_params def __call__( self , a , **a ) -> Tuple: return super().__call__(a , **a ) def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : List[Any] = load_image(a ) lowercase__ : Union[str, Any] = self.image_processor(images=a , return_tensors=self.framework ) return model_inputs def _UpperCAmelCase ( self , a ) -> List[str]: lowercase__ : Dict = self.model(**a ) return model_outputs def _UpperCAmelCase ( self , a , a=5 ) -> Dict: if top_k > self.model.config.num_labels: lowercase__ : List[Any] = self.model.config.num_labels if self.framework == "pt": lowercase__ : Tuple = model_outputs.logits.softmax(-1 )[0] lowercase__ , lowercase__ : Optional[Any] = probs.topk(a ) elif self.framework == "tf": lowercase__ : Union[str, Any] = stable_softmax(model_outputs.logits , axis=-1 )[0] lowercase__ : str = tf.math.top_k(a , k=a ) lowercase__ , lowercase__ : Dict = topk.values.numpy(), topk.indices.numpy() else: raise ValueError(f"""Unsupported framework: {self.framework}""" ) lowercase__ : Dict = scores.tolist() lowercase__ : Dict = ids.tolist() return [{"score": score, "label": self.model.config.idalabel[_id]} for score, _id in zip(a , a )]
77
1
"""simple docstring""" import logging import os from .state import PartialState class UpperCAmelCase_ ( logging.LoggerAdapter): @staticmethod def _UpperCAmelCase ( a ) -> Dict: lowercase__ : Any = PartialState() return not main_process_only or (main_process_only and state.is_main_process) def _UpperCAmelCase ( self , a , a , *a , **a ) -> Union[str, Any]: if PartialState._shared_state == {}: raise RuntimeError( 'You must initialize the accelerate state by calling either `PartialState()` or `Accelerator()` before using the logging utility.' ) lowercase__ : str = kwargs.pop('main_process_only' , a ) lowercase__ : Optional[int] = kwargs.pop('in_order' , a ) if self.isEnabledFor(a ): if self._should_log(a ): lowercase__ , lowercase__ : int = self.process(a , a ) self.logger.log(a , a , *a , **a ) elif in_order: lowercase__ : Dict = PartialState() for i in range(state.num_processes ): if i == state.process_index: lowercase__ , lowercase__ : Optional[Any] = self.process(a , a ) self.logger.log(a , a , *a , **a ) state.wait_for_everyone() def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : str = None ): '''simple docstring''' if log_level is None: lowercase__ : Optional[Any] = os.environ.get('ACCELERATE_LOG_LEVEL' , _lowerCAmelCase ) lowercase__ : List[Any] = logging.getLogger(_lowerCAmelCase ) if log_level is not None: logger.setLevel(log_level.upper() ) logger.root.setLevel(log_level.upper() ) return MultiProcessAdapter(_lowerCAmelCase , {} )
77
"""simple docstring""" import json import os from functools import lru_cache from typing import Dict, List, Optional, Tuple, Union import regex as re from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...tokenization_utils_base import BatchEncoding, EncodedInput from ...utils import PaddingStrategy, logging _UpperCamelCase : List[Any] = logging.get_logger(__name__) _UpperCamelCase : str = {"vocab_file": "vocab.json", "merges_file": "merges.txt"} # See all LED models at https://huggingface.co/models?filter=LED _UpperCamelCase : Optional[Any] = { "vocab_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/vocab.json", }, "merges_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/merges.txt", }, "tokenizer_file": { "allenai/led-base-16384": "https://huggingface.co/allenai/led-base-16384/resolve/main/tokenizer.json", }, } _UpperCamelCase : Optional[int] = { "allenai/led-base-16384": 1_63_84, } @lru_cache() # Copied from transformers.models.bart.tokenization_bart.bytes_to_unicode def a_ ( ): '''simple docstring''' lowercase__ : int = ( list(range(ord('!' ) , ord('~' ) + 1 ) ) + list(range(ord('¡' ) , ord('¬' ) + 1 ) ) + list(range(ord('®' ) , ord('ÿ' ) + 1 ) ) ) lowercase__ : Union[str, Any] = bs[:] lowercase__ : str = 0 for b in range(2**8 ): if b not in bs: bs.append(_lowerCAmelCase ) cs.append(2**8 + n ) n += 1 lowercase__ : str = [chr(_lowerCAmelCase ) for n in cs] return dict(zip(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Dict = set() lowercase__ : Union[str, Any] = word[0] for char in word[1:]: pairs.add((prev_char, char) ) lowercase__ : Optional[Any] = char return pairs class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = VOCAB_FILES_NAMES lowerCamelCase__ : List[str] = PRETRAINED_VOCAB_FILES_MAP lowerCamelCase__ : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCamelCase__ : Union[str, Any] = ["input_ids", "attention_mask"] def __init__( self , a , a , a="replace" , a="<s>" , a="</s>" , a="</s>" , a="<s>" , a="<unk>" , a="<pad>" , a="<mask>" , a=False , **a , ) -> Any: lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else bos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else eos_token lowercase__ : List[str] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else sep_token lowercase__ : Dict = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else cls_token lowercase__ : Any = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else unk_token lowercase__ : Tuple = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else pad_token # Mask token behave like a normal word, i.e. include the space before it lowercase__ : Optional[int] = AddedToken(a , lstrip=a , rstrip=a ) if isinstance(a , a ) else mask_token super().__init__( errors=a , bos_token=a , eos_token=a , unk_token=a , sep_token=a , cls_token=a , pad_token=a , mask_token=a , add_prefix_space=a , **a , ) with open(a , encoding='utf-8' ) as vocab_handle: lowercase__ : Tuple = json.load(a ) lowercase__ : Dict = {v: k for k, v in self.encoder.items()} lowercase__ : str = errors # how to handle errors in decoding lowercase__ : Optional[Any] = bytes_to_unicode() lowercase__ : Optional[Any] = {v: k for k, v in self.byte_encoder.items()} with open(a , encoding='utf-8' ) as merges_handle: lowercase__ : Optional[Any] = merges_handle.read().split('\n' )[1:-1] lowercase__ : Optional[int] = [tuple(merge.split() ) for merge in bpe_merges] lowercase__ : Union[str, Any] = dict(zip(a , range(len(a ) ) ) ) lowercase__ : Tuple = {} lowercase__ : List[str] = add_prefix_space # Should have added re.IGNORECASE so BPE merges can happen for capitalized versions of contractions lowercase__ : List[Any] = re.compile(R'\'s|\'t|\'re|\'ve|\'m|\'ll|\'d| ?\p{L}+| ?\p{N}+| ?[^\s\p{L}\p{N}]+|\s+(?!\S)|\s+' ) @property # Copied from transformers.models.bart.tokenization_bart.BartTokenizer.vocab_size def _UpperCAmelCase ( self ) -> List[Any]: return len(self.encoder ) def _UpperCAmelCase ( self ) -> str: return dict(self.encoder , **self.added_tokens_encoder ) def _UpperCAmelCase ( self , a ) -> List[str]: if token in self.cache: return self.cache[token] lowercase__ : Optional[Any] = tuple(a ) lowercase__ : int = get_pairs(a ) if not pairs: return token while True: lowercase__ : List[str] = min(a , key=lambda a : self.bpe_ranks.get(a , float('inf' ) ) ) if bigram not in self.bpe_ranks: break lowercase__ , lowercase__ : List[str] = bigram lowercase__ : Union[str, Any] = [] lowercase__ : List[Any] = 0 while i < len(a ): try: lowercase__ : str = word.index(a , a ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) lowercase__ : Optional[int] = j if word[i] == first and i < len(a ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 lowercase__ : int = tuple(a ) lowercase__ : Dict = new_word if len(a ) == 1: break else: lowercase__ : Any = get_pairs(a ) lowercase__ : List[str] = ' '.join(a ) lowercase__ : Optional[Any] = word return word def _UpperCAmelCase ( self , a ) -> Union[str, Any]: lowercase__ : Tuple = [] for token in re.findall(self.pat , a ): lowercase__ : Union[str, Any] = ''.join( self.byte_encoder[b] for b in token.encode('utf-8' ) ) # Maps all our bytes to unicode strings, avoiding control tokens of the BPE (spaces in our case) bpe_tokens.extend(bpe_token for bpe_token in self.bpe(a ).split(' ' ) ) return bpe_tokens def _UpperCAmelCase ( self , a ) -> Optional[Any]: return self.encoder.get(a , self.encoder.get(self.unk_token ) ) def _UpperCAmelCase ( self , a ) -> Optional[int]: return self.decoder.get(a ) def _UpperCAmelCase ( self , a ) -> str: lowercase__ : Any = ''.join(a ) lowercase__ : Dict = bytearray([self.byte_decoder[c] for c in text] ).decode('utf-8' , errors=self.errors ) return text def _UpperCAmelCase ( self , a , a = None ) -> Tuple[str]: if not os.path.isdir(a ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return lowercase__ : Any = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['vocab_file'] ) lowercase__ : str = os.path.join( a , (filename_prefix + '-' if filename_prefix else '') + VOCAB_FILES_NAMES['merges_file'] ) with open(a , 'w' , encoding='utf-8' ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=a , ensure_ascii=a ) + '\n' ) lowercase__ : List[Any] = 0 with open(a , 'w' , encoding='utf-8' ) as writer: writer.write('#version: 0.2\n' ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda a : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.""" ' Please check that the tokenizer is not corrupted!' ) lowercase__ : Union[str, Any] = token_index writer.write(' '.join(a ) + '\n' ) index += 1 return vocab_file, merge_file def _UpperCAmelCase ( self , a , a = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] lowercase__ : Union[str, Any] = [self.cls_token_id] lowercase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _UpperCAmelCase ( self , a , a = None , a = False ) -> List[int]: if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=a , token_ids_a=a , already_has_special_tokens=a ) if token_ids_a is None: return [1] + ([0] * len(a )) + [1] return [1] + ([0] * len(a )) + [1, 1] + ([0] * len(a )) + [1] def _UpperCAmelCase ( self , a , a = None ) -> List[int]: lowercase__ : Dict = [self.sep_token_id] lowercase__ : Optional[Any] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCAmelCase ( self , a , a=False , **a ) -> Optional[int]: lowercase__ : Tuple = kwargs.pop('add_prefix_space' , self.add_prefix_space ) if (is_split_into_words or add_prefix_space) and (len(a ) > 0 and not text[0].isspace()): lowercase__ : List[str] = ' ' + text return (text, kwargs) def _UpperCAmelCase ( self , a , a = None , a = PaddingStrategy.DO_NOT_PAD , a = None , a = None , ) -> dict: lowercase__ : Dict = super()._pad( encoded_inputs=a , max_length=a , padding_strategy=a , pad_to_multiple_of=a , return_attention_mask=a , ) # Load from model defaults if return_attention_mask is None: lowercase__ : Union[str, Any] = 'attention_mask' in self.model_input_names if return_attention_mask and "global_attention_mask" in encoded_inputs: lowercase__ : Any = encoded_inputs[self.model_input_names[0]] # `global_attention_mask` need to have the same length as other (sequential) inputs. lowercase__ : Tuple = len(encoded_inputs['global_attention_mask'] ) != len(a ) if needs_to_be_padded: lowercase__ : str = len(a ) - len(encoded_inputs['global_attention_mask'] ) if self.padding_side == "right": # Use `-1` since `0` in `global_attention_mask` means `local attention` instead of `not to attend` lowercase__ : Union[str, Any] = ( encoded_inputs['global_attention_mask'] + [-1] * difference ) elif self.padding_side == "left": lowercase__ : List[str] = [-1] * difference + encoded_inputs[ 'global_attention_mask' ] else: raise ValueError('Invalid padding strategy:' + str(self.padding_side ) ) return encoded_inputs
77
1
"""simple docstring""" from copy import deepcopy import torch import torch.nn.functional as F from torch.optim import AdamW from torch.optim.lr_scheduler import LambdaLR from torch.utils.data import DataLoader from accelerate.accelerator import Accelerator from accelerate.state import GradientState from accelerate.test_utils import RegressionDataset, RegressionModel from accelerate.utils import DistributedType, is_torch_version, set_seed def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Tuple , _lowerCAmelCase : str , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for param, grad_param in zip(model_a.parameters() , model_b.parameters() ): if not param.requires_grad: continue if not did_step: # Grads should not be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is False ), f"""Gradients in sync when they should not be at iteration {iteration}:\nmodel_a grad ({param.grad}) == model_b grad ({grad_param.grad})""" else: # Grads should be in sync assert ( torch.allclose(param.grad , grad_param.grad ) is True ), f"""Gradients not in sync when they should be at iteration {iteration}:\nmodel_a grad ({param.grad}) != model_b grad ({grad_param.grad})""" def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Dict=True ): '''simple docstring''' model.train() lowercase__ : Optional[int] = model(_lowerCAmelCase ) lowercase__ : int = F.mse_loss(_lowerCAmelCase , target.to(output.device ) ) if not do_backward: loss /= accelerator.gradient_accumulation_steps loss.backward() else: accelerator.backward(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str=False ): '''simple docstring''' set_seed(42 ) lowercase__ : Union[str, Any] = RegressionModel() lowercase__ : Dict = deepcopy(_lowerCAmelCase ) lowercase__ : Dict = RegressionDataset(length=80 ) lowercase__ : int = DataLoader(_lowerCAmelCase , batch_size=16 ) model.to(accelerator.device ) if sched: lowercase__ : int = AdamW(params=model.parameters() , lr=1E-3 ) lowercase__ : Dict = AdamW(params=ddp_model.parameters() , lr=1E-3 ) lowercase__ : str = LambdaLR(_lowerCAmelCase , lr_lambda=lambda _lowerCAmelCase : epoch**0.6_5 ) lowercase__ : Optional[Any] = LambdaLR(_lowerCAmelCase , lr_lambda=lambda _lowerCAmelCase : epoch**0.6_5 ) # Make a copy of `model` if sched: lowercase__ , lowercase__ , lowercase__ , lowercase__ : Optional[int] = accelerator.prepare(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: lowercase__ , lowercase__ : str = accelerator.prepare(_lowerCAmelCase , _lowerCAmelCase ) if sched: return (model, opt, sched, dataloader, ddp_model, ddp_opt, ddp_sched) return model, ddp_model, dataloader def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ , lowercase__ , lowercase__ : Union[str, Any] = get_training_setup(_lowerCAmelCase ) # Use a single batch lowercase__ , lowercase__ : Optional[int] = next(iter(_lowerCAmelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ : Tuple = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ : Optional[Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_lowerCAmelCase ): step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: # Sync grads step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Since `no_sync` is a noop, `ddp_model` and `model` grads should always be in sync check_model_parameters(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue assert torch.allclose( param.grad , ddp_param.grad ), f"""Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ : Tuple = ddp_input[torch.randperm(len(_lowerCAmelCase ) )] def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' lowercase__ , lowercase__ , lowercase__ : List[str] = get_training_setup(_lowerCAmelCase ) # Use a single batch lowercase__ , lowercase__ : Tuple = next(iter(_lowerCAmelCase ) ).values() for iteration in range(3 ): # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ : str = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ : Dict = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Do "gradient accumulation" (noop) if iteration % 2 == 0: # Accumulate grads locally with accelerator.no_sync(_lowerCAmelCase ): step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) else: # Sync grads step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if iteration % 2 == 0: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f"""Gradients in sync when they should not be:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})""" else: # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f"""Gradients not in sync when they should be:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ : List[Any] = ddp_input[torch.randperm(len(_lowerCAmelCase ) )] def a_ ( _lowerCAmelCase : List[str]=False , _lowerCAmelCase : Optional[int]=False ): '''simple docstring''' lowercase__ : str = Accelerator( split_batches=_lowerCAmelCase , dispatch_batches=_lowerCAmelCase , gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowercase__ , lowercase__ , lowercase__ : Optional[Any] = get_training_setup(_lowerCAmelCase ) for iteration, batch in enumerate(_lowerCAmelCase ): lowercase__ , lowercase__ : Optional[int] = batch.values() # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ : Dict = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ : Union[str, Any] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Do "gradient accumulation" (noop) with accelerator.accumulate(_lowerCAmelCase ): step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # DDP model and model should only be in sync when not (iteration % 2 == 0) for param, ddp_param in zip(model.parameters() , ddp_model.parameters() ): if not param.requires_grad: continue if ((iteration + 1) % 2 == 0) or (iteration == len(_lowerCAmelCase ) - 1): # Grads should be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is True ), f"""Gradients not in sync when they should be at iteration {iteration}:\nModel grad ({param.grad}) != DDP grad ({ddp_param.grad})""" else: # Grads should not be in sync assert ( torch.allclose(param.grad , ddp_param.grad ) is False ), f"""Gradients in sync when they should not be at iteration {iteration}:\nModel grad ({param.grad}) == DDP grad ({ddp_param.grad})""" # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) lowercase__ : int = ddp_input[torch.randperm(len(_lowerCAmelCase ) )] GradientState._reset_state() def a_ ( _lowerCAmelCase : Union[str, Any]=False , _lowerCAmelCase : Optional[int]=False ): '''simple docstring''' lowercase__ : List[Any] = Accelerator( split_batches=_lowerCAmelCase , dispatch_batches=_lowerCAmelCase , gradient_accumulation_steps=2 ) # Test that context manager behaves properly lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : List[str] = get_training_setup(_lowerCAmelCase , _lowerCAmelCase ) for iteration, batch in enumerate(_lowerCAmelCase ): lowercase__ , lowercase__ : str = batch.values() # Gather the distributed inputs and targs for the base model lowercase__ , lowercase__ : List[str] = accelerator.gather((ddp_input, ddp_target) ) lowercase__ , lowercase__ : List[str] = input.to(accelerator.device ), target.to(accelerator.device ) # Perform our initial ground truth step in non "DDP" model.train() ddp_model.train() step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) opt.step() if ((iteration + 1) % 2 == 0) or ((iteration + 1) == len(_lowerCAmelCase )): if split_batches: sched.step() else: for _ in range(accelerator.num_processes ): sched.step() opt.zero_grad() # Perform gradient accumulation under wrapper with accelerator.accumulate(_lowerCAmelCase ): step_model(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) ddp_opt.step() ddp_sched.step() ddp_opt.zero_grad() # Learning rates should be the same assert ( opt.param_groups[0]["lr"] == ddp_opt.param_groups[0]["lr"] ), f"""Learning rates found in each optimizer did not align\nopt: {opt.param_groups[0]["lr"]}\nDDP opt: {ddp_opt.param_groups[0]["lr"]}\n""" lowercase__ : List[str] = (((iteration + 1) % 2) == 0) or ((iteration + 1) == len(_lowerCAmelCase )) if accelerator.num_processes > 1: check_model_parameters(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Shuffle ddp_input on each iteration torch.manual_seed(1337 + iteration ) GradientState._reset_state() def a_ ( ): '''simple docstring''' lowercase__ : List[str] = Accelerator() lowercase__ : int = RegressionDataset(length=80 ) lowercase__ : Tuple = DataLoader(_lowerCAmelCase , batch_size=16 ) lowercase__ : Optional[int] = RegressionDataset(length=96 ) lowercase__ : List[str] = DataLoader(_lowerCAmelCase , batch_size=16 ) lowercase__ , lowercase__ : Optional[int] = accelerator.prepare(_lowerCAmelCase , _lowerCAmelCase ) assert accelerator.gradient_state.active_dataloader is None for iteration, _ in enumerate(_lowerCAmelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(_lowerCAmelCase ) if iteration < len(_lowerCAmelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader if iteration == 1: for batch_num, _ in enumerate(_lowerCAmelCase ): assert id(accelerator.gradient_state.active_dataloader ) == id(_lowerCAmelCase ) if batch_num < len(_lowerCAmelCase ) - 1: assert not accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader else: assert accelerator.gradient_state.end_of_dataloader assert accelerator.gradient_state.active_dataloader is None def a_ ( ): '''simple docstring''' lowercase__ : List[str] = Accelerator() lowercase__ : str = accelerator.state if state.local_process_index == 0: print('**Test `accumulate` gradient accumulation with dataloader break**' ) test_dataloader_break() if state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print('**Test NOOP `no_sync` context manager**' ) test_noop_sync(_lowerCAmelCase ) if state.distributed_type in (DistributedType.MULTI_GPU, DistributedType.MULTI_CPU): if state.local_process_index == 0: print('**Test Distributed `no_sync` context manager**' ) test_distributed_sync(_lowerCAmelCase ) if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation, ' , f"""`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**""" , ) test_gradient_accumulation(_lowerCAmelCase , _lowerCAmelCase ) # Currently will break on torch 2.0 +, need to investigate why if is_torch_version('<' , '2.0' ) or state.distributed_type == DistributedType.NO: if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , '`split_batches=False`, `dispatch_batches=False`**' , ) test_gradient_accumulation_with_opt_and_scheduler() if state.distributed_type == DistributedType.MULTI_GPU: for split_batch in [True, False]: for dispatch_batches in [True, False]: if not split_batch and not dispatch_batches: continue if state.local_process_index == 0: print( '**Test `accumulate` gradient accumulation with optimizer and scheduler, ' , f"""`split_batches={split_batch}` and `dispatch_batches={dispatch_batches}`**""" , ) test_gradient_accumulation_with_opt_and_scheduler(_lowerCAmelCase , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : str ): '''simple docstring''' main() if __name__ == "__main__": main()
77
"""simple docstring""" _UpperCamelCase : Union[str, Any] = 8.3_1_4_4_5_9_8 def a_ ( _lowerCAmelCase : float , _lowerCAmelCase : float ): '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _UpperCamelCase : List[Any] = 3_00 _UpperCamelCase : Tuple = 28 _UpperCamelCase : Any = rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
77
1
"""simple docstring""" from typing import Dict, Iterable, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_DEFAULT_MEAN, IMAGENET_DEFAULT_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, logging _UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[int] = ["pixel_values"] def __init__( self , a = True , a = None , a = PILImageResampling.BICUBIC , a = True , a = None , a = True , a = 1 / 2_5_5 , a = True , a = IMAGENET_DEFAULT_MEAN , a = IMAGENET_DEFAULT_STD , **a , ) -> None: super().__init__(**a ) lowercase__ : int = size if size is not None else {'shortest_edge': 2_2_4} lowercase__ : List[Any] = get_size_dict(a , default_to_square=a ) lowercase__ : Tuple = crop_size if crop_size is not None else {'height': 2_2_4, 'width': 2_2_4} lowercase__ : int = get_size_dict(a , param_name='crop_size' ) lowercase__ : int = do_resize lowercase__ : Optional[Any] = size lowercase__ : int = resample lowercase__ : Union[str, Any] = do_center_crop lowercase__ : str = crop_size lowercase__ : Union[str, Any] = do_rescale lowercase__ : Optional[int] = rescale_factor lowercase__ : Tuple = do_normalize lowercase__ : Dict = image_mean if image_mean is not None else IMAGENET_DEFAULT_MEAN lowercase__ : List[str] = image_std if image_std is not None else IMAGENET_DEFAULT_STD def _UpperCAmelCase ( self , a , a , a = PILImageResampling.BICUBIC , a = None , **a , ) -> np.ndarray: lowercase__ : int = get_size_dict(a , default_to_square=a ) # size_dict is a dict with either keys "height" and "width" or "shortest_edge" if "shortest_edge" in size: lowercase__ : int = int((2_5_6 / 2_2_4) * size['shortest_edge'] ) lowercase__ : Tuple = get_resize_output_image_size(a , size=a , default_to_square=a ) lowercase__ : Any = {'height': output_size[0], 'width': output_size[1]} if "height" not in size_dict or "width" not in size_dict: raise ValueError( f"""Size dict must have keys 'height' and 'width' or 'shortest_edge'. Got {size_dict.keys()}""" ) return resize( a , size=(size_dict['height'], size_dict['width']) , resample=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a = None , **a , ) -> np.ndarray: lowercase__ : Optional[Any] = get_size_dict(a ) if "height" not in size or "width" not in size: raise ValueError(f"""Size dict must have keys 'height' and 'width'. Got {size.keys()}""" ) return center_crop(a , size=(size['height'], size['width']) , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a = None , **a , ) -> np.ndarray: return rescale(a , scale=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a , a , a = None , **a , ) -> np.ndarray: return normalize(a , mean=a , std=a , data_format=a , **a ) def _UpperCAmelCase ( self , a , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = None , a = ChannelDimension.FIRST , **a , ) -> BatchFeature: lowercase__ : int = do_resize if do_resize is not None else self.do_resize lowercase__ : Optional[Any] = resample if resample is not None else self.resample lowercase__ : int = do_center_crop if do_center_crop is not None else self.do_center_crop lowercase__ : Optional[Any] = do_rescale if do_rescale is not None else self.do_rescale lowercase__ : List[str] = rescale_factor if rescale_factor is not None else self.rescale_factor lowercase__ : Any = do_normalize if do_normalize is not None else self.do_normalize lowercase__ : str = image_mean if image_mean is not None else self.image_mean lowercase__ : Any = image_std if image_std is not None else self.image_std lowercase__ : Dict = size if size is not None else self.size lowercase__ : Tuple = get_size_dict(a , default_to_square=a ) lowercase__ : Optional[Any] = crop_size if crop_size is not None else self.crop_size lowercase__ : Tuple = get_size_dict(a , param_name='crop_size' ) lowercase__ : List[str] = make_list_of_images(a ) if not valid_images(a ): raise ValueError( 'Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, ' 'torch.Tensor, tf.Tensor or jax.ndarray.' ) if do_resize and size is None: raise ValueError('Size must be specified if do_resize is True.' ) if do_center_crop and crop_size is None: raise ValueError('Crop size must be specified if do_center_crop is True.' ) if do_rescale and rescale_factor is None: raise ValueError('Rescale factor must be specified if do_rescale is True.' ) if do_normalize and (image_mean is None or image_std is None): raise ValueError('Image mean and std must be specified if do_normalize is True.' ) # All transformations expect numpy arrays. lowercase__ : Optional[Any] = [to_numpy_array(a ) for image in images] if do_resize: lowercase__ : Optional[int] = [self.resize(a , a , a ) for image in images] if do_center_crop: lowercase__ : List[Any] = [self.center_crop(a , a ) for image in images] if do_rescale: lowercase__ : Optional[Any] = [self.rescale(a , a ) for image in images] if do_normalize: lowercase__ : Dict = [self.normalize(a , a , a ) for image in images] lowercase__ : List[Any] = [to_channel_dimension_format(a , a ) for image in images] lowercase__ : Any = {'pixel_values': images} return BatchFeature(data=a , tensor_type=a )
77
"""simple docstring""" import copy from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import ClassLabel, Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): # `task` is not a ClassVar since we want it to be part of the `asdict` output for JSON serialization lowerCamelCase__ : str = field(default="text-classification" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({"labels": ClassLabel}) lowerCamelCase__ : str = "text" lowerCamelCase__ : str = "labels" def _UpperCAmelCase ( self , a ) -> Tuple: if self.label_column not in features: raise ValueError(f"""Column {self.label_column} is not present in features.""" ) if not isinstance(features[self.label_column] , a ): raise ValueError(f"""Column {self.label_column} is not a ClassLabel.""" ) lowercase__ : Optional[Any] = copy.deepcopy(self ) lowercase__ : Optional[Any] = self.label_schema.copy() lowercase__ : Any = features[self.label_column] lowercase__ : Optional[Any] = label_schema return task_template @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return { self.text_column: "text", self.label_column: "labels", }
77
1
"""simple docstring""" import math from typing import List, Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils import SchedulerMixin, SchedulerOutput class UpperCAmelCase_ ( _a , _a): lowerCamelCase__ : Any = 1 @register_to_config def __init__( self , a = 1_0_0_0 , a = None ) -> Dict: # set `betas`, `alphas`, `timesteps` self.set_timesteps(a ) # standard deviation of the initial noise distribution lowercase__ : Optional[int] = 1.0 # For now we only support F-PNDM, i.e. the runge-kutta method # For more information on the algorithm please take a look at the paper: https://arxiv.org/pdf/2202.09778.pdf # mainly at formula (9), (12), (13) and the Algorithm 2. lowercase__ : List[Any] = 4 # running values lowercase__ : Dict = [] def _UpperCAmelCase ( self , a , a = None ) -> Optional[int]: lowercase__ : int = num_inference_steps lowercase__ : Optional[int] = torch.linspace(1 , 0 , num_inference_steps + 1 )[:-1] lowercase__ : str = torch.cat([steps, torch.tensor([0.0] )] ) if self.config.trained_betas is not None: lowercase__ : Optional[int] = torch.tensor(self.config.trained_betas , dtype=torch.floataa ) else: lowercase__ : Optional[int] = torch.sin(steps * math.pi / 2 ) ** 2 lowercase__ : int = (1.0 - self.betas**2) ** 0.5 lowercase__ : Optional[int] = (torch.atana(self.betas , self.alphas ) / math.pi * 2)[:-1] lowercase__ : int = timesteps.to(a ) lowercase__ : Any = [] def _UpperCAmelCase ( self , a , a , a , a = True , ) -> Union[SchedulerOutput, Tuple]: if self.num_inference_steps is None: raise ValueError( 'Number of inference steps is \'None\', you need to run \'set_timesteps\' after creating the scheduler' ) lowercase__ : Optional[int] = (self.timesteps == timestep).nonzero().item() lowercase__ : int = timestep_index + 1 lowercase__ : Tuple = sample * self.betas[timestep_index] + model_output * self.alphas[timestep_index] self.ets.append(a ) if len(self.ets ) == 1: lowercase__ : Tuple = self.ets[-1] elif len(self.ets ) == 2: lowercase__ : Dict = (3 * self.ets[-1] - self.ets[-2]) / 2 elif len(self.ets ) == 3: lowercase__ : List[Any] = (2_3 * self.ets[-1] - 1_6 * self.ets[-2] + 5 * self.ets[-3]) / 1_2 else: lowercase__ : Dict = (1 / 2_4) * (5_5 * self.ets[-1] - 5_9 * self.ets[-2] + 3_7 * self.ets[-3] - 9 * self.ets[-4]) lowercase__ : str = self._get_prev_sample(a , a , a , a ) if not return_dict: return (prev_sample,) return SchedulerOutput(prev_sample=a ) def _UpperCAmelCase ( self , a , *a , **a ) -> torch.FloatTensor: return sample def _UpperCAmelCase ( self , a , a , a , a ) -> int: lowercase__ : Dict = self.alphas[timestep_index] lowercase__ : Tuple = self.betas[timestep_index] lowercase__ : Dict = self.alphas[prev_timestep_index] lowercase__ : Tuple = self.betas[prev_timestep_index] lowercase__ : Optional[Any] = (sample - sigma * ets) / max(a , 1e-8 ) lowercase__ : Optional[int] = next_alpha * pred + ets * next_sigma return prev_sample def __len__( self ) -> Union[str, Any]: return self.config.num_train_timesteps
77
"""simple docstring""" import argparse import collections import json import os import re import string import sys import numpy as np _UpperCamelCase : Any = re.compile(r"\b(a|an|the)\b", re.UNICODE) _UpperCamelCase : Union[str, Any] = None def a_ ( ): '''simple docstring''' lowercase__ : Optional[int] = argparse.ArgumentParser('Official evaluation script for SQuAD version 2.0.' ) parser.add_argument('data_file' , metavar='data.json' , help='Input data JSON file.' ) parser.add_argument('pred_file' , metavar='pred.json' , help='Model predictions.' ) parser.add_argument( '--out-file' , '-o' , metavar='eval.json' , help='Write accuracy metrics to file (default is stdout).' ) parser.add_argument( '--na-prob-file' , '-n' , metavar='na_prob.json' , help='Model estimates of probability of no answer.' ) parser.add_argument( '--na-prob-thresh' , '-t' , type=_lowerCAmelCase , default=1.0 , help='Predict "" if no-answer probability exceeds this (default = 1.0).' , ) parser.add_argument( '--out-image-dir' , '-p' , metavar='out_images' , default=_lowerCAmelCase , help='Save precision-recall curves to directory.' ) parser.add_argument('--verbose' , '-v' , action='store_true' ) if len(sys.argv ) == 1: parser.print_help() sys.exit(1 ) return parser.parse_args() def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Tuple = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Optional[int] = bool(qa['answers']['text'] ) return qid_to_has_ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : int ): return ARTICLES_REGEX.sub(' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : str ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : List[Any] ): lowercase__ : int = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : List[str] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' if not s: return [] return normalize_answer(_lowerCAmelCase ).split() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : str ): '''simple docstring''' return int(normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' lowercase__ : Dict = get_tokens(_lowerCAmelCase ) lowercase__ : List[str] = get_tokens(_lowerCAmelCase ) lowercase__ : List[Any] = collections.Counter(_lowerCAmelCase ) & collections.Counter(_lowerCAmelCase ) lowercase__ : int = sum(common.values() ) if len(_lowerCAmelCase ) == 0 or len(_lowerCAmelCase ) == 0: # If either is no-answer, then F1 is 1 if they agree, 0 otherwise return int(gold_toks == pred_toks ) if num_same == 0: return 0 lowercase__ : Any = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Dict = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : Any = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Optional[int] = {} lowercase__ : Union[str, Any] = {} for article in dataset: for p in article["paragraphs"]: for qa in p["qas"]: lowercase__ : Any = qa['id'] lowercase__ : Union[str, Any] = [t for t in qa['answers']['text'] if normalize_answer(_lowerCAmelCase )] if not gold_answers: # For unanswerable questions, only correct answer is empty string lowercase__ : Dict = [''] if qid not in preds: print(f"""Missing prediction for {qid}""" ) continue lowercase__ : Optional[int] = preds[qid] # Take max over all gold answers lowercase__ : int = max(compute_exact(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) lowercase__ : Optional[Any] = max(compute_fa(_lowerCAmelCase , _lowerCAmelCase ) for a in gold_answers ) return exact_scores, fa_scores def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[str] ): '''simple docstring''' lowercase__ : str = {} for qid, s in scores.items(): lowercase__ : int = na_probs[qid] > na_prob_thresh if pred_na: lowercase__ : Optional[Any] = float(not qid_to_has_ans[qid] ) else: lowercase__ : Optional[Any] = s return new_scores def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : List[Any] , _lowerCAmelCase : str=None ): '''simple docstring''' if not qid_list: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores.values() ) / total), ('f1', 1_0_0.0 * sum(fa_scores.values() ) / total), ('total', total), ] ) else: lowercase__ : Optional[Any] = len(_lowerCAmelCase ) return collections.OrderedDict( [ ('exact', 1_0_0.0 * sum(exact_scores[k] for k in qid_list ) / total), ('f1', 1_0_0.0 * sum(fa_scores[k] for k in qid_list ) / total), ('total', total), ] ) def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' for k in new_eval: lowercase__ : int = new_eval[k] def a_ ( _lowerCAmelCase : str , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : int , _lowerCAmelCase : Tuple ): '''simple docstring''' plt.step(_lowerCAmelCase , _lowerCAmelCase , color='b' , alpha=0.2 , where='post' ) plt.fill_between(_lowerCAmelCase , _lowerCAmelCase , step='post' , alpha=0.2 , color='b' ) plt.xlabel('Recall' ) plt.ylabel('Precision' ) plt.xlim([0.0, 1.0_5] ) plt.ylim([0.0, 1.0_5] ) plt.title(_lowerCAmelCase ) plt.savefig(_lowerCAmelCase ) plt.clf() def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str , _lowerCAmelCase : int , _lowerCAmelCase : str , _lowerCAmelCase : Any=None , _lowerCAmelCase : List[str]=None ): '''simple docstring''' lowercase__ : Optional[int] = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) lowercase__ : Tuple = 0.0 lowercase__ : List[str] = 1.0 lowercase__ : List[str] = 0.0 lowercase__ : Union[str, Any] = [1.0] lowercase__ : List[Any] = [0.0] lowercase__ : Optional[int] = 0.0 for i, qid in enumerate(_lowerCAmelCase ): if qid_to_has_ans[qid]: true_pos += scores[qid] lowercase__ : Tuple = true_pos / float(i + 1 ) lowercase__ : Union[str, Any] = true_pos / float(_lowerCAmelCase ) if i == len(_lowerCAmelCase ) - 1 or na_probs[qid] != na_probs[qid_list[i + 1]]: # i.e., if we can put a threshold after this point avg_prec += cur_p * (cur_r - recalls[-1]) precisions.append(_lowerCAmelCase ) recalls.append(_lowerCAmelCase ) if out_image: plot_pr_curve(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) return {"ap": 1_0_0.0 * avg_prec} def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Tuple ): '''simple docstring''' if out_image_dir and not os.path.exists(_lowerCAmelCase ): os.makedirs(_lowerCAmelCase ) lowercase__ : List[str] = sum(1 for v in qid_to_has_ans.values() if v ) if num_true_pos == 0: return lowercase__ : Dict = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_exact.png' ) , title='Precision-Recall curve for Exact Match score' , ) lowercase__ : Tuple = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_f1.png' ) , title='Precision-Recall curve for F1 score' , ) lowercase__ : List[Any] = {k: float(_lowerCAmelCase ) for k, v in qid_to_has_ans.items()} lowercase__ : Any = make_precision_recall_eval( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , out_image=os.path.join(_lowerCAmelCase , 'pr_oracle.png' ) , title='Oracle Precision-Recall curve (binary task of HasAns vs. NoAns)' , ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_exact' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_f1' ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'pr_oracle' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Any , _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' if not qid_list: return lowercase__ : List[str] = [na_probs[k] for k in qid_list] lowercase__ : Tuple = np.ones_like(_lowerCAmelCase ) / float(len(_lowerCAmelCase ) ) plt.hist(_lowerCAmelCase , weights=_lowerCAmelCase , bins=20 , range=(0.0, 1.0) ) plt.xlabel('Model probability of no-answer' ) plt.ylabel('Proportion of dataset' ) plt.title(f"""Histogram of no-answer probability: {name}""" ) plt.savefig(os.path.join(_lowerCAmelCase , f"""na_prob_hist_{name}.png""" ) ) plt.clf() def a_ ( _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : Tuple = sum(1 for k in qid_to_has_ans if not qid_to_has_ans[k] ) lowercase__ : int = num_no_ans lowercase__ : Optional[int] = cur_score lowercase__ : Tuple = 0.0 lowercase__ : Dict = sorted(_lowerCAmelCase , key=lambda _lowerCAmelCase : na_probs[k] ) for i, qid in enumerate(_lowerCAmelCase ): if qid not in scores: continue if qid_to_has_ans[qid]: lowercase__ : Optional[int] = scores[qid] else: if preds[qid]: lowercase__ : List[Any] = -1 else: lowercase__ : Optional[int] = 0 cur_score += diff if cur_score > best_score: lowercase__ : Dict = cur_score lowercase__ : Optional[int] = na_probs[qid] return 1_0_0.0 * best_score / len(_lowerCAmelCase ), best_thresh def a_ ( _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : Tuple , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : str ): '''simple docstring''' lowercase__ , lowercase__ : List[Any] = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ , lowercase__ : Dict = find_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Any = best_exact lowercase__ : Tuple = exact_thresh lowercase__ : Optional[Any] = best_fa lowercase__ : Any = fa_thresh def a_ ( ): '''simple docstring''' with open(OPTS.data_file ) as f: lowercase__ : List[Any] = json.load(_lowerCAmelCase ) lowercase__ : Union[str, Any] = dataset_json['data'] with open(OPTS.pred_file ) as f: lowercase__ : str = json.load(_lowerCAmelCase ) if OPTS.na_prob_file: with open(OPTS.na_prob_file ) as f: lowercase__ : Union[str, Any] = json.load(_lowerCAmelCase ) else: lowercase__ : str = {k: 0.0 for k in preds} lowercase__ : int = make_qid_to_has_ans(_lowerCAmelCase ) # maps qid to True/False lowercase__ : List[str] = [k for k, v in qid_to_has_ans.items() if v] lowercase__ : Any = [k for k, v in qid_to_has_ans.items() if not v] lowercase__ , lowercase__ : Any = get_raw_scores(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : Optional[Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Union[str, Any] = apply_no_ans_threshold(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.na_prob_thresh ) lowercase__ : Tuple = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase ) if has_ans_qids: lowercase__ : int = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'HasAns' ) if no_ans_qids: lowercase__ : Optional[Any] = make_eval_dict(_lowerCAmelCase , _lowerCAmelCase , qid_list=_lowerCAmelCase ) merge_eval(_lowerCAmelCase , _lowerCAmelCase , 'NoAns' ) if OPTS.na_prob_file: find_all_best_thresh(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) if OPTS.na_prob_file and OPTS.out_image_dir: run_precision_recall_analysis(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'hasAns' ) histogram_na_prob(_lowerCAmelCase , _lowerCAmelCase , OPTS.out_image_dir , 'noAns' ) if OPTS.out_file: with open(OPTS.out_file , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) else: print(json.dumps(_lowerCAmelCase , indent=2 ) ) if __name__ == "__main__": _UpperCamelCase : Optional[int] = parse_args() if OPTS.out_image_dir: import matplotlib matplotlib.use("Agg") import matplotlib.pyplot as plt main()
77
1
"""simple docstring""" from math import sqrt def a_ ( _lowerCAmelCase : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' must been an int and positive" lowercase__ : List[Any] = True # 0 and 1 are none primes. if number <= 1: lowercase__ : List[Any] = False for divisor in range(2 , int(round(sqrt(_lowerCAmelCase ) ) ) + 1 ): # if 'number' divisible by 'divisor' then sets 'status' # of false and break up the loop. if number % divisor == 0: lowercase__ : Any = False break # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'status' must been from type bool" return status def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" # beginList: contains all natural numbers from 2 up to N lowercase__ : int = list(range(2 , n + 1 ) ) lowercase__ : Optional[int] = [] # this list will be returns. # actual sieve of erathostenes for i in range(len(_lowerCAmelCase ) ): for j in range(i + 1 , len(_lowerCAmelCase ) ): if (begin_list[i] != 0) and (begin_list[j] % begin_list[i] == 0): lowercase__ : Optional[int] = 0 # filters actual prime numbers. lowercase__ : Optional[Any] = [x for x in begin_list if x != 0] # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : int ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n > 2), "'N' must been an int and > 2" lowercase__ : str = [] # iterates over all numbers between 2 up to N+1 # if a number is prime then appends to list 'ans' for number in range(2 , n + 1 ): if is_prime(_lowerCAmelCase ): ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and number >= 0, "'number' must been an int and >= 0" lowercase__ : Union[str, Any] = [] # this list will be returns of the function. # potential prime number factors. lowercase__ : int = 2 lowercase__ : List[Any] = number if number == 0 or number == 1: ans.append(_lowerCAmelCase ) # if 'number' not prime then builds the prime factorization of 'number' elif not is_prime(_lowerCAmelCase ): while quotient != 1: if is_prime(_lowerCAmelCase ) and (quotient % factor == 0): ans.append(_lowerCAmelCase ) quotient /= factor else: factor += 1 else: ans.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type list" return ans def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ : Optional[int] = 0 # prime factorization of 'number' lowercase__ : List[str] = prime_factorization(_lowerCAmelCase ) lowercase__ : Any = max(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def a_ ( _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number >= 0 ), "'number' bust been an int and >= 0" lowercase__ : int = 0 # prime factorization of 'number' lowercase__ : List[str] = prime_factorization(_lowerCAmelCase ) lowercase__ : Union[str, Any] = min(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'ans' must been from type int" return ans def a_ ( _lowerCAmelCase : Tuple ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 == 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 == 0 def a_ ( _lowerCAmelCase : str ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ), "'number' must been an int" assert isinstance(number % 2 != 0 , _lowerCAmelCase ), "compare bust been from type bool" return number % 2 != 0 def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (number > 2) and is_even(_lowerCAmelCase ) ), "'number' must been an int, even and > 2" lowercase__ : List[Any] = [] # this list will returned # creates a list of prime numbers between 2 up to 'number' lowercase__ : List[str] = get_prime_numbers(_lowerCAmelCase ) lowercase__ : Any = len(_lowerCAmelCase ) # run variable for while-loops. lowercase__ : Optional[int] = 0 lowercase__ : str = None # exit variable. for break up the loops lowercase__ : Optional[int] = True while i < len_pn and loop: lowercase__ : Any = i + 1 while j < len_pn and loop: if prime_numbers[i] + prime_numbers[j] == number: lowercase__ : Optional[Any] = False ans.append(prime_numbers[i] ) ans.append(prime_numbers[j] ) j += 1 i += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (len(_lowerCAmelCase ) == 2) and (ans[0] + ans[1] == number) and is_prime(ans[0] ) and is_prime(ans[1] ) ), "'ans' must contains two primes. And sum of elements must been eq 'number'" return ans def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : List[str] ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 0) and (numbera >= 0) ), "'number1' and 'number2' must been positive integer." lowercase__ : List[Any] = 0 while numbera != 0: lowercase__ : Optional[int] = numbera % numbera lowercase__ : Optional[Any] = numbera lowercase__ : Tuple = rest # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( numbera >= 0 ), "'number' must been from type int and positive" return numbera def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : int ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numbera >= 1) and (numbera >= 1) ), "'number1' and 'number2' must been positive integer." lowercase__ : List[str] = 1 # actual answer that will be return. # for kgV (x,1) if numbera > 1 and numbera > 1: # builds the prime factorization of 'number1' and 'number2' lowercase__ : Tuple = prime_factorization(_lowerCAmelCase ) lowercase__ : int = prime_factorization(_lowerCAmelCase ) elif numbera == 1 or numbera == 1: lowercase__ : Optional[Any] = [] lowercase__ : Tuple = [] lowercase__ : List[Any] = max(_lowerCAmelCase , _lowerCAmelCase ) lowercase__ : List[Any] = 0 lowercase__ : List[Any] = 0 lowercase__ : List[str] = [] # captured numbers int both 'primeFac1' and 'primeFac2' # iterates through primeFac1 for n in prime_fac_a: if n not in done: if n in prime_fac_a: lowercase__ : Any = prime_fac_a.count(_lowerCAmelCase ) lowercase__ : int = prime_fac_a.count(_lowerCAmelCase ) for _ in range(max(_lowerCAmelCase , _lowerCAmelCase ) ): ans *= n else: lowercase__ : Optional[Any] = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # iterates through primeFac2 for n in prime_fac_a: if n not in done: lowercase__ : List[Any] = prime_fac_a.count(_lowerCAmelCase ) for _ in range(_lowerCAmelCase ): ans *= n done.append(_lowerCAmelCase ) # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( ans >= 0 ), "'ans' must been from type int and positive" return ans def a_ ( _lowerCAmelCase : Dict ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'number' must been a positive int" lowercase__ : Union[str, Any] = 0 lowercase__ : int = 2 # this variable holds the answer while index < n: index += 1 ans += 1 # counts to the next number # if ans not prime then # runs to the next prime number. while not is_prime(_lowerCAmelCase ): ans += 1 # precondition assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and is_prime( _lowerCAmelCase ), "'ans' must been a prime number and from type int" return ans def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : int ): '''simple docstring''' assert ( is_prime(_lowerCAmelCase ) and is_prime(_lowerCAmelCase ) and (p_number_a < p_number_a) ), "The arguments must been prime numbers and 'pNumber1' < 'pNumber2'" lowercase__ : Any = p_number_a + 1 # jump to the next number lowercase__ : str = [] # this list will be returns. # if number is not prime then # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 while number < p_number_a: ans.append(_lowerCAmelCase ) number += 1 # fetch the next prime number. while not is_prime(_lowerCAmelCase ): number += 1 # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ans[0] != p_number_a and ans[len(_lowerCAmelCase ) - 1] != p_number_a ), "'ans' must been a list without the arguments" # 'ans' contains not 'pNumber1' and 'pNumber2' ! return ans def a_ ( _lowerCAmelCase : str ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 1), "'n' must been int and >= 1" lowercase__ : Optional[int] = [] # will be returned. for divisor in range(1 , n + 1 ): if n % divisor == 0: ans.append(_lowerCAmelCase ) # precondition assert ans[0] == 1 and ans[len(_lowerCAmelCase ) - 1] == n, "Error in function getDivisiors(...)" return ans def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and ( number > 1 ), "'number' must been an int and >= 1" lowercase__ : List[str] = get_divisors(_lowerCAmelCase ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (divisors[0] == 1) and (divisors[len(_lowerCAmelCase ) - 1] == number) ), "Error in help-function getDivisiors(...)" # summed all divisors up to 'number' (exclusive), hence [:-1] return sum(divisors[:-1] ) == number def a_ ( _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any ): '''simple docstring''' assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (denominator != 0) ), "The arguments must been from type int and 'denominator' != 0" # build the greatest common divisor of numerator and denominator. lowercase__ : Dict = gcd(abs(_lowerCAmelCase ) , abs(_lowerCAmelCase ) ) # precondition assert ( isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (numerator % gcd_of_fraction == 0) and (denominator % gcd_of_fraction == 0) ), "Error in function gcd(...,...)" return (numerator // gcd_of_fraction, denominator // gcd_of_fraction) def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been a int and >= 0" lowercase__ : Any = 1 # this will be return. for factor in range(1 , n + 1 ): ans *= factor return ans def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' assert isinstance(_lowerCAmelCase , _lowerCAmelCase ) and (n >= 0), "'n' must been an int and >= 0" lowercase__ : List[Any] = 0 lowercase__ : Union[str, Any] = 1 lowercase__ : List[str] = 1 # this will be return for _ in range(n - 1 ): lowercase__ : List[Any] = ans ans += fiba lowercase__ : Tuple = tmp return ans
77
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / "utils")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : Dict = 0 def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Tuple = AutoImageProcessor.from_pretrained('openai/clip-vit-base-patch32' ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Any: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : str = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : str = Path(a ) / 'preprocessor_config.json' lowercase__ : int = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Optional[Any]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = CLIPConfig() # Create a dummy config file with image_proceesor_type lowercase__ : Optional[int] = Path(a ) / 'preprocessor_config.json' lowercase__ : Optional[int] = Path(a ) / 'config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally lowercase__ : int = AutoImageProcessor.from_pretrained(a ).to_dict() config_dict.pop('image_processor_type' ) lowercase__ : Tuple = CLIPImageProcessor(**a ) # save in new folder model_config.save_pretrained(a ) config.save_pretrained(a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained(a ) # make sure private variable is not incorrectly saved lowercase__ : Optional[int] = json.loads(config.to_json_string() ) self.assertTrue('_processor_class' not in dict_as_saved ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> List[str]: with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Dict = Path(a ) / 'preprocessor_config.json' json.dump( {'image_processor_type': 'CLIPImageProcessor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) lowercase__ : List[str] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'clip-base is not a local folder and is not a valid model identifier' ): lowercase__ : Any = AutoImageProcessor.from_pretrained('clip-base' ) def _UpperCAmelCase ( self ) -> List[Any]: with self.assertRaisesRegex( a , R'aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)' ): lowercase__ : Dict = AutoImageProcessor.from_pretrained(a , revision='aaaaaa' ) def _UpperCAmelCase ( self ) -> Union[str, Any]: with self.assertRaisesRegex( a , 'hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json.' , ): lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/config-no-model' ) def _UpperCAmelCase ( self ) -> Optional[int]: # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(a ): lowercase__ : List[Any] = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) # If remote code is disabled, we can't load this config. with self.assertRaises(a ): lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) lowercase__ : Union[str, Any] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : str = AutoImageProcessor.from_pretrained(a , trust_remote_code=a ) self.assertEqual(reloaded_image_processor.__class__.__name__ , 'NewImageProcessor' ) def _UpperCAmelCase ( self ) -> int: try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(a ): AutoImageProcessor.register(a , a ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase__ : Optional[Any] = Path(a ) / 'preprocessor_config.json' lowercase__ : List[Any] = Path(a ) / 'config.json' json.dump( {'feature_extractor_type': 'CLIPFeatureExtractor', 'processor_class': 'CLIPProcessor'} , open(a , 'w' ) , ) json.dump({'model_type': 'clip'} , open(a , 'w' ) ) lowercase__ : Union[str, Any] = CustomImageProcessor.from_pretrained(a ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(a ) lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained(a ) self.assertIsInstance(a , a ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def _UpperCAmelCase ( self ) -> Dict: class UpperCAmelCase_ ( _a): lowerCamelCase__ : Union[str, Any] = True try: AutoConfig.register('custom' , a ) AutoImageProcessor.register(a , a ) # If remote code is not set, the default is to use local lowercase__ : int = AutoImageProcessor.from_pretrained('hf-internal-testing/test_dynamic_image_processor' ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. lowercase__ : Optional[int] = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub lowercase__ : int = AutoImageProcessor.from_pretrained( 'hf-internal-testing/test_dynamic_image_processor' , trust_remote_code=a ) self.assertEqual(image_processor.__class__.__name__ , 'NewImageProcessor' ) self.assertTrue(not hasattr(a , 'is_local' ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
77
1
"""simple docstring""" import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : int , _lowerCAmelCase : List[Any]=True , _lowerCAmelCase : Optional[Any]="pt" ): '''simple docstring''' lowercase__ : Optional[int] = {'add_prefix_space': True} if isinstance(_lowerCAmelCase , _lowerCAmelCase ) and not line.startswith(' ' ) else {} lowercase__ : str = padding_side return tokenizer( [line] , max_length=_lowerCAmelCase , padding='max_length' if pad_to_max_length else None , truncation=_lowerCAmelCase , return_tensors=_lowerCAmelCase , add_special_tokens=_lowerCAmelCase , **_lowerCAmelCase , ) def a_ ( _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : str=None , ): '''simple docstring''' lowercase__ : List[Any] = input_ids.ne(_lowerCAmelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class UpperCAmelCase_ ( _a): def __init__( self , a , a , a , a , a="train" , a=None , a=None , a=None , a="" , ) -> Any: super().__init__() lowercase__ : Any = Path(a ).joinpath(type_path + '.source' ) lowercase__ : Union[str, Any] = Path(a ).joinpath(type_path + '.target' ) lowercase__ : List[str] = self.get_char_lens(self.src_file ) lowercase__ : Any = max_source_length lowercase__ : str = max_target_length assert min(self.src_lens ) > 0, f"""found empty line in {self.src_file}""" lowercase__ : List[Any] = tokenizer lowercase__ : Optional[Any] = prefix if n_obs is not None: lowercase__ : str = self.src_lens[:n_obs] lowercase__ : Optional[Any] = src_lang lowercase__ : Tuple = tgt_lang def __len__( self ) -> List[Any]: return len(self.src_lens ) def __getitem__( self , a ) -> Dict[str, torch.Tensor]: lowercase__ : Optional[Any] = index + 1 # linecache starts at 1 lowercase__ : Tuple = self.prefix + linecache.getline(str(self.src_file ) , a ).rstrip('\n' ) lowercase__ : List[Any] = linecache.getline(str(self.tgt_file ) , a ).rstrip('\n' ) assert source_line, f"""empty source line for index {index}""" assert tgt_line, f"""empty tgt line for index {index}""" # Need to add eos token manually for T5 if isinstance(self.tokenizer , a ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right lowercase__ : str = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , a ) else self.tokenizer ) lowercase__ : List[Any] = self.tokenizer.generator if isinstance(self.tokenizer , a ) else self.tokenizer lowercase__ : Dict = encode_line(a , a , self.max_source_length , 'right' ) lowercase__ : Optional[int] = encode_line(a , a , self.max_target_length , 'right' ) lowercase__ : Dict = source_inputs['input_ids'].squeeze() lowercase__ : Any = target_inputs['input_ids'].squeeze() lowercase__ : Any = source_inputs['attention_mask'].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def _UpperCAmelCase ( a ) -> Optional[Any]: return [len(a ) for x in Path(a ).open().readlines()] def _UpperCAmelCase ( self , a ) -> Dict[str, torch.Tensor]: lowercase__ : Union[str, Any] = torch.stack([x['input_ids'] for x in batch] ) lowercase__ : Dict = torch.stack([x['attention_mask'] for x in batch] ) lowercase__ : List[str] = torch.stack([x['decoder_input_ids'] for x in batch] ) lowercase__ : Optional[Any] = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , a ) else self.tokenizer.pad_token_id ) lowercase__ : Optional[int] = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , a ) else self.tokenizer.pad_token_id ) lowercase__ : Optional[Any] = trim_batch(a , a ) lowercase__ , lowercase__ : Optional[int] = trim_batch(a , a , attention_mask=a ) lowercase__ : str = { 'input_ids': source_ids, 'attention_mask': source_mask, 'decoder_input_ids': y, } return batch _UpperCamelCase : Any = getLogger(__name__) def a_ ( _lowerCAmelCase : List[List] ): '''simple docstring''' return list(itertools.chain.from_iterable(_lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : str ): '''simple docstring''' lowercase__ : int = get_git_info() save_json(_lowerCAmelCase , os.path.join(_lowerCAmelCase , 'git_log.json' ) ) def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[Any]=4 , **_lowerCAmelCase : Dict ): '''simple docstring''' with open(_lowerCAmelCase , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase , indent=_lowerCAmelCase , **_lowerCAmelCase ) def a_ ( _lowerCAmelCase : List[str] ): '''simple docstring''' with open(_lowerCAmelCase ) as f: return json.load(_lowerCAmelCase ) def a_ ( ): '''simple docstring''' lowercase__ : int = git.Repo(search_parent_directories=_lowerCAmelCase ) lowercase__ : Union[str, Any] = { 'repo_id': str(_lowerCAmelCase ), 'repo_sha': str(repo.head.object.hexsha ), 'repo_branch': str(repo.active_branch ), 'hostname': str(socket.gethostname() ), } return repo_infos def a_ ( _lowerCAmelCase : Callable , _lowerCAmelCase : Iterable ): '''simple docstring''' return list(map(_lowerCAmelCase , _lowerCAmelCase ) ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Optional[int] ): '''simple docstring''' with open(_lowerCAmelCase , 'wb' ) as f: return pickle.dump(_lowerCAmelCase , _lowerCAmelCase ) def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' def remove_articles(_lowerCAmelCase : Union[str, Any] ): return re.sub(R'\b(a|an|the)\b' , ' ' , _lowerCAmelCase ) def white_space_fix(_lowerCAmelCase : Optional[Any] ): return " ".join(text.split() ) def remove_punc(_lowerCAmelCase : int ): lowercase__ : str = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(_lowerCAmelCase : Dict ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(_lowerCAmelCase ) ) ) ) def a_ ( _lowerCAmelCase : Dict , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' lowercase__ : Union[str, Any] = normalize_answer(_lowerCAmelCase ).split() lowercase__ : Dict = normalize_answer(_lowerCAmelCase ).split() lowercase__ : Tuple = Counter(_lowerCAmelCase ) & Counter(_lowerCAmelCase ) lowercase__ : Tuple = sum(common.values() ) if num_same == 0: return 0 lowercase__ : List[str] = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : List[Any] = 1.0 * num_same / len(_lowerCAmelCase ) lowercase__ : List[Any] = (2 * precision * recall) / (precision + recall) return fa def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : Tuple ): '''simple docstring''' return normalize_answer(_lowerCAmelCase ) == normalize_answer(_lowerCAmelCase ) def a_ ( _lowerCAmelCase : List[str] , _lowerCAmelCase : List[str] ): '''simple docstring''' assert len(_lowerCAmelCase ) == len(_lowerCAmelCase ) lowercase__ : Dict = 0 for hypo, pred in zip(_lowerCAmelCase , _lowerCAmelCase ): em += exact_match_score(_lowerCAmelCase , _lowerCAmelCase ) if len(_lowerCAmelCase ) > 0: em /= len(_lowerCAmelCase ) return {"em": em} def a_ ( _lowerCAmelCase : Any ): '''simple docstring''' return model_prefix.startswith('rag' ) def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ): '''simple docstring''' lowercase__ : Union[str, Any] = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead lowercase__ : int = 'dropout_rate' for p in extra_params: if getattr(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ): if not hasattr(_lowerCAmelCase , _lowerCAmelCase ) and not hasattr(_lowerCAmelCase , equivalent_param[p] ): logger.info('config doesn\'t have a `{}` attribute'.format(_lowerCAmelCase ) ) delattr(_lowerCAmelCase , _lowerCAmelCase ) continue lowercase__ : Any = p if hasattr(_lowerCAmelCase , _lowerCAmelCase ) else equivalent_param[p] setattr(_lowerCAmelCase , _lowerCAmelCase , getattr(_lowerCAmelCase , _lowerCAmelCase ) ) delattr(_lowerCAmelCase , _lowerCAmelCase ) return hparams, config
77
"""simple docstring""" import os from typing import List, Optional, Union from ...image_processing_utils import BatchFeature from ...image_utils import ImageInput from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType from ..auto import AutoTokenizer class UpperCAmelCase_ ( _a): lowerCamelCase__ : Dict = ["image_processor", "tokenizer"] lowerCamelCase__ : Dict = "BlipImageProcessor" lowerCamelCase__ : Union[str, Any] = "AutoTokenizer" def __init__( self , a , a , a ) -> Optional[int]: super().__init__(a , a ) # add QFormer tokenizer lowercase__ : Dict = qformer_tokenizer def __call__( self , a = None , a = None , a = True , a = False , a = None , a = None , a = 0 , a = None , a = None , a = False , a = False , a = False , a = False , a = False , a = True , a = None , **a , ) -> BatchFeature: if images is None and text is None: raise ValueError('You have to specify at least images or text.' ) lowercase__ : List[Any] = BatchFeature() if text is not None: lowercase__ : Optional[int] = self.tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) encoding.update(a ) lowercase__ : Optional[int] = self.qformer_tokenizer( text=a , add_special_tokens=a , padding=a , truncation=a , max_length=a , stride=a , pad_to_multiple_of=a , return_attention_mask=a , return_overflowing_tokens=a , return_special_tokens_mask=a , return_offsets_mapping=a , return_token_type_ids=a , return_length=a , verbose=a , return_tensors=a , **a , ) lowercase__ : List[str] = qformer_text_encoding.pop('input_ids' ) lowercase__ : Any = qformer_text_encoding.pop('attention_mask' ) if images is not None: lowercase__ : List[Any] = self.image_processor(a , return_tensors=a ) encoding.update(a ) return encoding def _UpperCAmelCase ( self , *a , **a ) -> List[str]: return self.tokenizer.batch_decode(*a , **a ) def _UpperCAmelCase ( self , *a , **a ) -> Tuple: return self.tokenizer.decode(*a , **a ) @property # Copied from transformers.models.blip.processing_blip.BlipProcessor.model_input_names def _UpperCAmelCase ( self ) -> Union[str, Any]: lowercase__ : str = self.tokenizer.model_input_names lowercase__ : List[Any] = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) def _UpperCAmelCase ( self , a , **a ) -> Optional[int]: if os.path.isfile(a ): raise ValueError(f"""Provided path ({save_directory}) should be a directory, not a file""" ) os.makedirs(a , exist_ok=a ) lowercase__ : int = os.path.join(a , 'qformer_tokenizer' ) self.qformer_tokenizer.save_pretrained(a ) return super().save_pretrained(a , **a ) @classmethod def _UpperCAmelCase ( cls , a , **a ) -> str: lowercase__ : str = AutoTokenizer.from_pretrained(a , subfolder='qformer_tokenizer' ) lowercase__ : int = cls._get_arguments_from_pretrained(a , **a ) args.append(a ) return cls(*a )
77
1
"""simple docstring""" from __future__ import annotations def a_ ( _lowerCAmelCase : list[int] ): # This function is recursive '''simple docstring''' lowercase__ : Tuple = len(_lowerCAmelCase ) # If the array contains only one element, we return it (it's the stop condition of # recursion) if array_length <= 1: return array # Else lowercase__ : List[str] = array[0] lowercase__ : Union[str, Any] = False lowercase__ : Optional[Any] = 1 lowercase__ : list[int] = [] while not is_found and i < array_length: if array[i] < pivot: lowercase__ : Any = True lowercase__ : Dict = [element for element in array[i:] if element >= array[i]] lowercase__ : Union[str, Any] = longest_subsequence(_lowerCAmelCase ) if len(_lowerCAmelCase ) > len(_lowerCAmelCase ): lowercase__ : Dict = temp_array else: i += 1 lowercase__ : int = [element for element in array[1:] if element >= pivot] lowercase__ : Union[str, Any] = [pivot, *longest_subsequence(_lowerCAmelCase )] if len(_lowerCAmelCase ) > len(_lowerCAmelCase ): return temp_array else: return longest_subseq if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" from dataclasses import dataclass, field from typing import ClassVar, Dict from ..features import Features, Value from .base import TaskTemplate @dataclass(frozen=_a) class UpperCAmelCase_ ( _a): lowerCamelCase__ : str = field(default="language-modeling" , metadata={"include_in_asdict_even_if_is_default": True}) lowerCamelCase__ : ClassVar[Features] = Features({"text": Value("string")}) lowerCamelCase__ : ClassVar[Features] = Features({}) lowerCamelCase__ : str = "text" @property def _UpperCAmelCase ( self ) -> Dict[str, str]: return {self.text_column: "text"}
77
1
"""simple docstring""" import unittest import numpy as np from transformers import is_flax_available from transformers.testing_utils import require_flax from ..test_modeling_flax_common import ids_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.generation import ( FlaxForcedBOSTokenLogitsProcessor, FlaxForcedEOSTokenLogitsProcessor, FlaxLogitsProcessorList, FlaxMinLengthLogitsProcessor, FlaxTemperatureLogitsWarper, FlaxTopKLogitsWarper, FlaxTopPLogitsWarper, ) @require_flax class UpperCAmelCase_ ( unittest.TestCase): def _UpperCAmelCase ( self , a , a ) -> Tuple: lowercase__ : Tuple = jnp.ones((batch_size, length) ) / length return scores def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : List[str] = None lowercase__ : Any = 2_0 lowercase__ : int = self._get_uniform_logits(batch_size=2 , length=a ) # tweak scores to not be uniform anymore lowercase__ : Optional[int] = scores.at[1, 5].set((1 / length) + 0.1 ) # peak, 1st batch lowercase__ : Optional[Any] = scores.at[1, 1_0].set((1 / length) - 0.4 ) # valley, 1st batch # compute softmax lowercase__ : int = jax.nn.softmax(a , axis=-1 ) lowercase__ : int = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowercase__ : Tuple = FlaxTemperatureLogitsWarper(temperature=1.3 ) lowercase__ : Union[str, Any] = jax.nn.softmax(temp_dist_warper_sharper(a , scores.copy() , cur_len=a ) , axis=-1 ) lowercase__ : str = jax.nn.softmax(temp_dist_warper_smoother(a , scores.copy() , cur_len=a ) , axis=-1 ) # uniform distribution stays uniform self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_sharp[0, :] , atol=1e-3 ) ) self.assertTrue(jnp.allclose(probs[0, :] , warped_prob_smooth[0, :] , atol=1e-3 ) ) # sharp peaks get higher, valleys get lower self.assertLess(probs[1, :].max() , warped_prob_sharp[1, :].max() ) self.assertGreater(probs[1, :].min() , warped_prob_sharp[1, :].min() ) # smooth peaks get lower, valleys get higher self.assertGreater(probs[1, :].max() , warped_prob_smooth[1, :].max() ) self.assertLess(probs[1, :].min() , warped_prob_smooth[1, :].min() ) def _UpperCAmelCase ( self ) -> str: lowercase__ : Any = None lowercase__ : List[str] = 1_0 lowercase__ : Optional[int] = 2 # create ramp distribution lowercase__ : int = np.broadcast_to(np.arange(a )[None, :] , (batch_size, vocab_size) ).copy() lowercase__ : Dict = ramp_logits[1:, : vocab_size // 2] + vocab_size lowercase__ : Optional[int] = FlaxTopKLogitsWarper(3 ) lowercase__ : Optional[Any] = top_k_warp(a , a , cur_len=a ) # check that correct tokens are filtered self.assertListEqual(jnp.isinf(scores[0] ).tolist() , 7 * [True] + 3 * [False] ) self.assertListEqual(jnp.isinf(scores[1] ).tolist() , 2 * [True] + 3 * [False] + 5 * [True] ) # check special case lowercase__ : List[Any] = 5 lowercase__ : int = FlaxTopKLogitsWarper(top_k=1 , filter_value=0.0 , min_tokens_to_keep=3 ) lowercase__ : Union[str, Any] = np.broadcast_to(np.arange(a )[None, :] , (batch_size, length) ).copy() lowercase__ : Optional[int] = top_k_warp_safety_check(a , a , cur_len=a ) # min_tokens overwrites k: 3 tokens are kept => 2 tokens are nullified self.assertListEqual((scores == 0.0).sum(axis=-1 ).tolist() , [2, 2] ) def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : Dict = None lowercase__ : Tuple = 1_0 lowercase__ : Dict = 2 # create distribution and take log (inverse to Softmax as taken in TopPLogitsWarper) lowercase__ : List[str] = np.log(np.array([[0.3, 0.1, 0.1, 0.5], [0.15, 0.3, 0.3, 0.25]] ) ) lowercase__ : Any = FlaxTopPLogitsWarper(0.8 ) lowercase__ : Optional[int] = np.exp(top_p_warp(a , a , cur_len=a ) ) # dist should be filtered to keep min num values so that sum is >= top_p # exp (-inf) => 0 lowercase__ : Dict = np.array([[0.3, 0.0, 0.0, 0.5], [0.0, 0.3, 0.3, 0.25]] ) self.assertTrue(np.allclose(a , a , atol=1e-3 ) ) # check edge cases with negative and extreme logits lowercase__ : List[Any] = np.broadcast_to(np.arange(a )[None, :] , (batch_size, vocab_size) ).copy() - ( vocab_size // 2 ) # make ramp_logits more extreme lowercase__ : str = ramp_logits[1] * 100.0 # make sure at least 2 tokens are kept lowercase__ : Optional[Any] = FlaxTopPLogitsWarper(0.9 , min_tokens_to_keep=2 , filter_value=0.0 ) lowercase__ : Optional[int] = top_p_warp(a , a , cur_len=a ) # first batch should keep three tokens, second batch would keep only 1, but due to `min_tokens_to_keep=2` keeps 2. self.assertListEqual((filtered_dist != 0.0).sum(axis=-1 ).tolist() , [3, 2] ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Union[str, Any] = 2_0 lowercase__ : Union[str, Any] = 4 lowercase__ : Tuple = 0 lowercase__ : str = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=a ) # check that min length is applied at length 5 lowercase__ : Optional[int] = ids_tensor((batch_size, 2_0) , vocab_size=2_0 ) lowercase__ : Union[str, Any] = 5 lowercase__ : Tuple = self._get_uniform_logits(a , a ) lowercase__ : Dict = min_dist_processor(a , a , cur_len=a ) self.assertListEqual(scores_before_min_length[:, eos_token_id].tolist() , 4 * [-float('inf' )] ) # check that min length is not applied anymore at length 15 lowercase__ : str = self._get_uniform_logits(a , a ) lowercase__ : Dict = 1_5 lowercase__ : Tuple = min_dist_processor(a , a , cur_len=a ) self.assertFalse(jnp.isinf(a ).any() ) def _UpperCAmelCase ( self ) -> List[Any]: lowercase__ : Union[str, Any] = 2_0 lowercase__ : str = 4 lowercase__ : List[str] = 0 lowercase__ : Tuple = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=a ) # check that all scores are -inf except the bos_token_id score lowercase__ : List[Any] = ids_tensor((batch_size, 1) , vocab_size=2_0 ) lowercase__ : Dict = 1 lowercase__ : Optional[Any] = self._get_uniform_logits(a , a ) lowercase__ : List[str] = logits_processor(a , a , cur_len=a ) self.assertTrue(jnp.isneginf(scores[:, bos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, bos_token_id].tolist() , 4 * [0] ) # score for bos_token_id shold be zero # check that bos_token_id is not forced if current length is greater than 1 lowercase__ : Tuple = 3 lowercase__ : Optional[Any] = self._get_uniform_logits(a , a ) lowercase__ : List[str] = logits_processor(a , a , cur_len=a ) self.assertFalse(jnp.isinf(a ).any() ) def _UpperCAmelCase ( self ) -> Optional[int]: lowercase__ : Optional[Any] = 2_0 lowercase__ : Union[str, Any] = 4 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = 5 lowercase__ : List[Any] = FlaxForcedEOSTokenLogitsProcessor(max_length=a , eos_token_id=a ) # check that all scores are -inf except the eos_token_id when max_length is reached lowercase__ : Union[str, Any] = ids_tensor((batch_size, 4) , vocab_size=2_0 ) lowercase__ : Optional[int] = 4 lowercase__ : List[str] = self._get_uniform_logits(a , a ) lowercase__ : Tuple = logits_processor(a , a , cur_len=a ) self.assertTrue(jnp.isneginf(scores[:, eos_token_id + 1 :] ).all() ) self.assertListEqual(scores[:, eos_token_id].tolist() , 4 * [0] ) # score for eos_token_id should be zero # check that eos_token_id is not forced if max_length is not reached lowercase__ : Optional[int] = 3 lowercase__ : int = self._get_uniform_logits(a , a ) lowercase__ : Tuple = logits_processor(a , a , cur_len=a ) self.assertFalse(jnp.isinf(a ).any() ) def _UpperCAmelCase ( self ) -> str: lowercase__ : Tuple = 4 lowercase__ : Union[str, Any] = 1_0 lowercase__ : Optional[int] = 1_5 lowercase__ : List[str] = 2 lowercase__ : Optional[Any] = 1 lowercase__ : Tuple = 1_5 # dummy input_ids and scores lowercase__ : List[str] = ids_tensor((batch_size, sequence_length) , a ) lowercase__ : int = input_ids.copy() lowercase__ : str = self._get_uniform_logits(a , a ) lowercase__ : List[Any] = scores.copy() # instantiate all dist processors lowercase__ : Dict = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowercase__ : List[Any] = FlaxTopKLogitsWarper(3 ) lowercase__ : Dict = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors lowercase__ : int = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=a ) lowercase__ : Union[str, Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=a ) lowercase__ : Optional[int] = FlaxForcedEOSTokenLogitsProcessor(max_length=a , eos_token_id=a ) lowercase__ : int = 1_0 # no processor list lowercase__ : Tuple = temp_dist_warp(a , a , cur_len=a ) lowercase__ : str = top_k_warp(a , a , cur_len=a ) lowercase__ : Optional[int] = top_p_warp(a , a , cur_len=a ) lowercase__ : Tuple = min_dist_proc(a , a , cur_len=a ) lowercase__ : Any = bos_dist_proc(a , a , cur_len=a ) lowercase__ : str = eos_dist_proc(a , a , cur_len=a ) # with processor list lowercase__ : Tuple = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) lowercase__ : Optional[Any] = processor(a , a , cur_len=a ) # scores should be equal self.assertTrue(jnp.allclose(a , a , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() ) def _UpperCAmelCase ( self ) -> Tuple: lowercase__ : Optional[Any] = 4 lowercase__ : List[str] = 1_0 lowercase__ : Union[str, Any] = 1_5 lowercase__ : Dict = 2 lowercase__ : int = 1 lowercase__ : Optional[Any] = 1_5 # dummy input_ids and scores lowercase__ : List[Any] = ids_tensor((batch_size, sequence_length) , a ) lowercase__ : Tuple = input_ids.copy() lowercase__ : int = self._get_uniform_logits(a , a ) lowercase__ : List[str] = scores.copy() # instantiate all dist processors lowercase__ : List[Any] = FlaxTemperatureLogitsWarper(temperature=0.5 ) lowercase__ : Tuple = FlaxTopKLogitsWarper(3 ) lowercase__ : Optional[Any] = FlaxTopPLogitsWarper(0.8 ) # instantiate all logits processors lowercase__ : int = FlaxMinLengthLogitsProcessor(min_length=1_0 , eos_token_id=a ) lowercase__ : List[Any] = FlaxForcedBOSTokenLogitsProcessor(bos_token_id=a ) lowercase__ : str = FlaxForcedEOSTokenLogitsProcessor(max_length=a , eos_token_id=a ) lowercase__ : int = 1_0 # no processor list def run_no_processor_list(a , a , a ): lowercase__ : Optional[Any] = temp_dist_warp(a , a , cur_len=a ) lowercase__ : List[str] = top_k_warp(a , a , cur_len=a ) lowercase__ : Tuple = top_p_warp(a , a , cur_len=a ) lowercase__ : List[Any] = min_dist_proc(a , a , cur_len=a ) lowercase__ : Union[str, Any] = bos_dist_proc(a , a , cur_len=a ) lowercase__ : List[Any] = eos_dist_proc(a , a , cur_len=a ) return scores # with processor list def run_processor_list(a , a , a ): lowercase__ : Optional[int] = FlaxLogitsProcessorList( [temp_dist_warp, top_k_warp, top_p_warp, min_dist_proc, bos_dist_proc, eos_dist_proc] ) lowercase__ : int = processor(a , a , cur_len=a ) return scores lowercase__ : Dict = jax.jit(a ) lowercase__ : Union[str, Any] = jax.jit(a ) lowercase__ : Any = jitted_run_no_processor_list(a , a , a ) lowercase__ : str = jitted_run_processor_list(a , a , a ) # scores should be equal self.assertTrue(jnp.allclose(a , a , atol=1e-3 ) ) # input_ids should never be changed self.assertListEqual(input_ids.tolist() , input_ids_comp.tolist() )
77
"""simple docstring""" from . import __version__ # Backward compatibility imports, to make sure all those objects can be found in file_utils from .utils import ( CLOUDFRONT_DISTRIB_PREFIX, CONFIG_NAME, DISABLE_TELEMETRY, DUMMY_INPUTS, DUMMY_MASK, ENV_VARS_TRUE_AND_AUTO_VALUES, ENV_VARS_TRUE_VALUES, FEATURE_EXTRACTOR_NAME, FLAX_WEIGHTS_NAME, HF_MODULES_CACHE, HUGGINGFACE_CO_PREFIX, HUGGINGFACE_CO_RESOLVE_ENDPOINT, MODEL_CARD_NAME, MULTIPLE_CHOICE_DUMMY_INPUTS, PYTORCH_PRETRAINED_BERT_CACHE, PYTORCH_TRANSFORMERS_CACHE, S3_BUCKET_PREFIX, SENTENCEPIECE_UNDERLINE, SPIECE_UNDERLINE, TF2_WEIGHTS_NAME, TF_WEIGHTS_NAME, TORCH_FX_REQUIRED_VERSION, TRANSFORMERS_CACHE, TRANSFORMERS_DYNAMIC_MODULE_NAME, USE_JAX, USE_TF, USE_TORCH, WEIGHTS_INDEX_NAME, WEIGHTS_NAME, ContextManagers, DummyObject, EntryNotFoundError, ExplicitEnum, ModelOutput, PaddingStrategy, PushToHubMixin, RepositoryNotFoundError, RevisionNotFoundError, TensorType, _LazyModule, add_code_sample_docstrings, add_end_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, cached_property, copy_func, default_cache_path, define_sagemaker_information, get_cached_models, get_file_from_repo, get_full_repo_name, get_torch_version, has_file, http_user_agent, is_apex_available, is_bsa_available, is_coloredlogs_available, is_datasets_available, is_detectrona_available, is_faiss_available, is_flax_available, is_ftfy_available, is_in_notebook, is_ipex_available, is_librosa_available, is_offline_mode, is_onnx_available, is_pandas_available, is_phonemizer_available, is_protobuf_available, is_psutil_available, is_pyanvml_available, is_pyctcdecode_available, is_pytesseract_available, is_pytorch_quantization_available, is_rjieba_available, is_sagemaker_dp_enabled, is_sagemaker_mp_enabled, is_scipy_available, is_sentencepiece_available, is_seqio_available, is_sklearn_available, is_soundfile_availble, is_spacy_available, is_speech_available, is_tensor, is_tensorflow_probability_available, is_tfaonnx_available, is_tf_available, is_timm_available, is_tokenizers_available, is_torch_available, is_torch_bfaa_available, is_torch_cuda_available, is_torch_fx_available, is_torch_fx_proxy, is_torch_mps_available, is_torch_tfaa_available, is_torch_tpu_available, is_torchaudio_available, is_training_run_on_sagemaker, is_vision_available, replace_return_docstrings, requires_backends, to_numpy, to_py_obj, torch_only_method, )
77
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_dpt import DPTImageProcessor _UpperCamelCase : Union[str, Any] = logging.get_logger(__name__) class UpperCAmelCase_ ( _a): def __init__( self , *a , **a ) -> None: warnings.warn( 'The class DPTFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please' ' use DPTImageProcessor instead.' , a , ) super().__init__(*a , **a )
77
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_vision_available(): from PIL import Image from ..image_utils import load_image if is_torch_available(): import torch from transformers.modeling_outputs import BaseModelOutput from ..models.auto.modeling_auto import MODEL_FOR_ZERO_SHOT_OBJECT_DETECTION_MAPPING _UpperCamelCase : Dict = logging.get_logger(__name__) @add_end_docstrings(_a) class UpperCAmelCase_ ( _a): def __init__( self , **a ) -> Dict: super().__init__(**a ) if self.framework == "tf": raise ValueError(f"""The {self.__class__} is only available in PyTorch.""" ) requires_backends(self , 'vision' ) self.check_model_type(a ) def __call__( self , a , a = None , **a , ) -> List[str]: if "text_queries" in kwargs: lowercase__ : Optional[Any] = kwargs.pop('text_queries' ) if isinstance(a , (str, Image.Image) ): lowercase__ : Optional[Any] = {'image': image, 'candidate_labels': candidate_labels} else: lowercase__ : List[str] = image lowercase__ : Optional[Any] = super().__call__(a , **a ) return results def _UpperCAmelCase ( self , **a ) -> Dict: lowercase__ : Optional[Any] = {} if "threshold" in kwargs: lowercase__ : Tuple = kwargs['threshold'] if "top_k" in kwargs: lowercase__ : List[Any] = kwargs['top_k'] return {}, {}, postprocess_params def _UpperCAmelCase ( self , a ) -> Dict: lowercase__ : Any = load_image(inputs['image'] ) lowercase__ : Optional[int] = inputs['candidate_labels'] if isinstance(a , a ): lowercase__ : Optional[int] = candidate_labels.split(',' ) lowercase__ : Optional[int] = torch.tensor([[image.height, image.width]] , dtype=torch.intaa ) for i, candidate_label in enumerate(a ): lowercase__ : List[str] = self.tokenizer(a , return_tensors=self.framework ) lowercase__ : List[Any] = self.image_processor(a , return_tensors=self.framework ) yield { "is_last": i == len(a ) - 1, "target_size": target_size, "candidate_label": candidate_label, **text_inputs, **image_features, } def _UpperCAmelCase ( self , a ) -> List[Any]: lowercase__ : List[Any] = model_inputs.pop('target_size' ) lowercase__ : Dict = model_inputs.pop('candidate_label' ) lowercase__ : Dict = model_inputs.pop('is_last' ) lowercase__ : Optional[int] = self.model(**a ) lowercase__ : Any = {'target_size': target_size, 'candidate_label': candidate_label, 'is_last': is_last, **outputs} return model_outputs def _UpperCAmelCase ( self , a , a=0.1 , a=None ) -> Union[str, Any]: lowercase__ : Dict = [] for model_output in model_outputs: lowercase__ : List[Any] = model_output['candidate_label'] lowercase__ : Optional[int] = BaseModelOutput(a ) lowercase__ : Any = self.image_processor.post_process_object_detection( outputs=a , threshold=a , target_sizes=model_output['target_size'] )[0] for index in outputs["scores"].nonzero(): lowercase__ : Union[str, Any] = outputs['scores'][index].item() lowercase__ : Tuple = self._get_bounding_box(outputs['boxes'][index][0] ) lowercase__ : Tuple = {'score': score, 'label': label, 'box': box} results.append(a ) lowercase__ : Dict = sorted(a , key=lambda a : x["score"] , reverse=a ) if top_k: lowercase__ : Dict = results[:top_k] return results def _UpperCAmelCase ( self , a ) -> Dict[str, int]: if self.framework != "pt": raise ValueError('The ZeroShotObjectDetectionPipeline is only available in PyTorch.' ) lowercase__ , lowercase__ , lowercase__ , lowercase__ : Tuple = box.int().tolist() lowercase__ : Any = { 'xmin': xmin, 'ymin': ymin, 'xmax': xmax, 'ymax': ymax, } return bbox
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int = 1000 ): '''simple docstring''' lowercase__ : Tuple = 2**power lowercase__ : Tuple = str(_lowerCAmelCase ) lowercase__ : Union[str, Any] = list(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0 for i in list_num: sum_of_num += int(_lowerCAmelCase ) return sum_of_num if __name__ == "__main__": _UpperCamelCase : Optional[int] = int(input("Enter the power of 2: ").strip()) print("2 ^ ", power, " = ", 2**power) _UpperCamelCase : List[Any] = solution(power) print("Sum of the digits is: ", result)
77
"""simple docstring""" from __future__ import annotations import unittest import numpy as np from transformers import BlipTextConfig from transformers.testing_utils import require_tf, slow from transformers.utils import is_tf_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask if is_tf_available(): import tensorflow as tf from transformers import TFBlipTextModel from transformers.models.blip.modeling_tf_blip import TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST class UpperCAmelCase_ : def __init__( self , a , a=1_2 , a=7 , a=True , a=True , a=True , a=9_9 , a=3_2 , a=3_2 , a=2 , a=4 , a=3_7 , a=0.1 , a=0.1 , a=5_1_2 , a=0.02 , a=0 , a=None , ) -> Union[str, Any]: lowercase__ : Any = parent lowercase__ : str = batch_size lowercase__ : List[Any] = seq_length lowercase__ : Union[str, Any] = is_training lowercase__ : List[str] = use_input_mask lowercase__ : int = use_labels lowercase__ : List[Any] = vocab_size lowercase__ : str = hidden_size lowercase__ : int = projection_dim lowercase__ : Optional[int] = num_hidden_layers lowercase__ : Any = num_attention_heads lowercase__ : Optional[Any] = intermediate_size lowercase__ : Optional[Any] = dropout lowercase__ : Optional[int] = attention_dropout lowercase__ : Optional[int] = max_position_embeddings lowercase__ : str = initializer_range lowercase__ : Tuple = scope lowercase__ : int = bos_token_id def _UpperCAmelCase ( self ) -> Optional[Any]: lowercase__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) lowercase__ : int = None if self.use_input_mask: lowercase__ : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) if input_mask is not None: lowercase__ : int = input_mask.numpy() lowercase__ , lowercase__ : Tuple = input_mask.shape lowercase__ : List[str] = np.random.randint(1 , seq_length - 1 , size=(batch_size,) ) for batch_idx, start_index in enumerate(a ): lowercase__ : Dict = 1 lowercase__ : Union[str, Any] = 0 lowercase__ : Tuple = self.get_config() return config, input_ids, tf.convert_to_tensor(a ) def _UpperCAmelCase ( self ) -> List[Any]: return BlipTextConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , projection_dim=self.projection_dim , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , dropout=self.dropout , attention_dropout=self.attention_dropout , max_position_embeddings=self.max_position_embeddings , initializer_range=self.initializer_range , bos_token_id=self.bos_token_id , ) def _UpperCAmelCase ( self , a , a , a ) -> Any: lowercase__ : List[Any] = TFBlipTextModel(config=a ) lowercase__ : Optional[int] = model(a , attention_mask=a , training=a ) lowercase__ : List[str] = model(a , training=a ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def _UpperCAmelCase ( self ) -> Any: lowercase__ : Optional[Any] = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ : Any = config_and_inputs lowercase__ : Any = {'input_ids': input_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class UpperCAmelCase_ ( _a , unittest.TestCase): lowerCamelCase__ : Dict = (TFBlipTextModel,) if is_tf_available() else () lowerCamelCase__ : Optional[Any] = False lowerCamelCase__ : List[str] = False lowerCamelCase__ : Any = False def _UpperCAmelCase ( self ) -> List[str]: lowercase__ : Optional[int] = BlipTextModelTester(self ) lowercase__ : int = ConfigTester(self , config_class=a , hidden_size=3_7 ) def _UpperCAmelCase ( self ) -> Tuple: self.config_tester.run_common_tests() def _UpperCAmelCase ( self ) -> int: lowercase__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a ) def _UpperCAmelCase ( self ) -> Optional[Any]: pass def _UpperCAmelCase ( self ) -> Optional[int]: pass @unittest.skip(reason='Blip does not use inputs_embeds' ) def _UpperCAmelCase ( self ) -> List[str]: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> Dict: pass @unittest.skip(reason='BlipTextModel has no base class and is not available in MODEL_MAPPING' ) def _UpperCAmelCase ( self ) -> str: pass @slow def _UpperCAmelCase ( self ) -> int: for model_name in TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ : Any = TFBlipTextModel.from_pretrained(a ) self.assertIsNotNone(a ) def _UpperCAmelCase ( self , a=True ) -> List[str]: super().test_pt_tf_model_equivalence(allow_missing_keys=a )
77
1
"""simple docstring""" from math import loga def a_ ( _lowerCAmelCase : int ): '''simple docstring''' if a < 0: raise ValueError('Input value must be a positive integer' ) elif isinstance(_lowerCAmelCase , _lowerCAmelCase ): raise TypeError('Input value must be a \'int\' type' ) return 0 if (a == 0) else int(loga(a & -a ) ) if __name__ == "__main__": import doctest doctest.testmod()
77
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _UpperCamelCase : Tuple = logging.get_logger(__name__) _UpperCamelCase : Union[str, Any] = { "alibaba-damo/mgp-str-base": "https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json", } class UpperCAmelCase_ ( _a): lowerCamelCase__ : List[Any] = "mgp-str" def __init__( self , a=[3_2, 1_2_8] , a=4 , a=3 , a=2_7 , a=3_8 , a=5_0_2_5_7 , a=3_0_5_2_2 , a=7_6_8 , a=1_2 , a=1_2 , a=4.0 , a=True , a=False , a=1e-5 , a=0.0 , a=0.0 , a=0.0 , a=False , a=0.02 , **a , ) -> Tuple: super().__init__(**a ) lowercase__ : int = image_size lowercase__ : List[Any] = patch_size lowercase__ : Optional[Any] = num_channels lowercase__ : Optional[Any] = max_token_length lowercase__ : Dict = num_character_labels lowercase__ : Optional[int] = num_bpe_labels lowercase__ : Dict = num_wordpiece_labels lowercase__ : Tuple = hidden_size lowercase__ : List[Any] = num_hidden_layers lowercase__ : Union[str, Any] = num_attention_heads lowercase__ : Dict = mlp_ratio lowercase__ : Optional[int] = distilled lowercase__ : Optional[int] = layer_norm_eps lowercase__ : Optional[int] = drop_rate lowercase__ : List[str] = qkv_bias lowercase__ : Optional[int] = attn_drop_rate lowercase__ : Any = drop_path_rate lowercase__ : List[Any] = output_aa_attentions lowercase__ : Tuple = initializer_range
77
1
"""simple docstring""" from math import acos, sin from typing import List, Tuple, Union import numpy as np import torch from PIL import Image from ...models import AutoencoderKL, UNetaDConditionModel from ...schedulers import DDIMScheduler, DDPMScheduler from ...utils import randn_tensor from ..pipeline_utils import AudioPipelineOutput, BaseOutput, DiffusionPipeline, ImagePipelineOutput from .mel import Mel class UpperCAmelCase_ ( _a): lowerCamelCase__ : Optional[int] = ["vqvae"] def __init__( self , a , a , a , a , ) -> Optional[int]: super().__init__() self.register_modules(unet=a , scheduler=a , mel=a , vqvae=a ) def _UpperCAmelCase ( self ) -> int: return 5_0 if isinstance(self.scheduler , a ) else 1_0_0_0 @torch.no_grad() def __call__( self , a = 1 , a = None , a = None , a = 0 , a = 0 , a = None , a = None , a = 0 , a = 0 , a = None , a = 0 , a = None , a = None , a=True , ) -> Union[ Union[AudioPipelineOutput, ImagePipelineOutput], Tuple[List[Image.Image], Tuple[int, List[np.ndarray]]], ]: lowercase__ : Optional[Any] = steps or self.get_default_steps() self.scheduler.set_timesteps(a ) lowercase__ : Dict = step_generator or generator # For backwards compatibility if type(self.unet.config.sample_size ) == int: lowercase__ : str = (self.unet.config.sample_size, self.unet.config.sample_size) if noise is None: lowercase__ : Union[str, Any] = randn_tensor( ( batch_size, self.unet.config.in_channels, self.unet.config.sample_size[0], self.unet.config.sample_size[1], ) , generator=a , device=self.device , ) lowercase__ : Dict = noise lowercase__ : str = None if audio_file is not None or raw_audio is not None: self.mel.load_audio(a , a ) lowercase__ : str = self.mel.audio_slice_to_image(a ) lowercase__ : Tuple = np.frombuffer(input_image.tobytes() , dtype='uint8' ).reshape( (input_image.height, input_image.width) ) lowercase__ : List[str] = (input_image / 2_5_5) * 2 - 1 lowercase__ : Optional[Any] = torch.tensor(input_image[np.newaxis, :, :] , dtype=torch.float ).to(self.device ) if self.vqvae is not None: lowercase__ : List[str] = self.vqvae.encode(torch.unsqueeze(a , 0 ) ).latent_dist.sample( generator=a )[0] lowercase__ : Tuple = self.vqvae.config.scaling_factor * input_images if start_step > 0: lowercase__ : Union[str, Any] = self.scheduler.add_noise(a , a , self.scheduler.timesteps[start_step - 1] ) lowercase__ : Any = ( self.unet.config.sample_size[1] * self.mel.get_sample_rate() / self.mel.x_res / self.mel.hop_length ) lowercase__ : Optional[int] = int(mask_start_secs * pixels_per_second ) lowercase__ : Optional[int] = int(mask_end_secs * pixels_per_second ) lowercase__ : Optional[Any] = self.scheduler.add_noise(a , a , torch.tensor(self.scheduler.timesteps[start_step:] ) ) for step, t in enumerate(self.progress_bar(self.scheduler.timesteps[start_step:] ) ): if isinstance(self.unet , a ): lowercase__ : Any = self.unet(a , a , a )['sample'] else: lowercase__ : Optional[int] = self.unet(a , a )['sample'] if isinstance(self.scheduler , a ): lowercase__ : Union[str, Any] = self.scheduler.step( model_output=a , timestep=a , sample=a , eta=a , generator=a , )['prev_sample'] else: lowercase__ : int = self.scheduler.step( model_output=a , timestep=a , sample=a , generator=a , )['prev_sample'] if mask is not None: if mask_start > 0: lowercase__ : int = mask[:, step, :, :mask_start] if mask_end > 0: lowercase__ : Optional[int] = mask[:, step, :, -mask_end:] if self.vqvae is not None: # 0.18215 was scaling factor used in training to ensure unit variance lowercase__ : Optional[Any] = 1 / self.vqvae.config.scaling_factor * images lowercase__ : Dict = self.vqvae.decode(a )['sample'] lowercase__ : Any = (images / 2 + 0.5).clamp(0 , 1 ) lowercase__ : List[str] = images.cpu().permute(0 , 2 , 3 , 1 ).numpy() lowercase__ : Optional[int] = (images * 2_5_5).round().astype('uint8' ) lowercase__ : List[str] = list( (Image.fromarray(_[:, :, 0] ) for _ in images) if images.shape[3] == 1 else (Image.fromarray(a , mode='RGB' ).convert('L' ) for _ in images) ) lowercase__ : int = [self.mel.image_to_audio(a ) for _ in images] if not return_dict: return images, (self.mel.get_sample_rate(), audios) return BaseOutput(**AudioPipelineOutput(np.array(a )[:, np.newaxis, :] ) , **ImagePipelineOutput(a ) ) @torch.no_grad() def _UpperCAmelCase ( self , a , a = 5_0 ) -> np.ndarray: assert isinstance(self.scheduler , a ) self.scheduler.set_timesteps(a ) lowercase__ : Optional[Any] = np.array( [np.frombuffer(image.tobytes() , dtype='uint8' ).reshape((1, image.height, image.width) ) for image in images] ) lowercase__ : Any = (sample / 2_5_5) * 2 - 1 lowercase__ : Optional[int] = torch.Tensor(a ).to(self.device ) for t in self.progress_bar(torch.flip(self.scheduler.timesteps , (0,) ) ): lowercase__ : Any = t - self.scheduler.config.num_train_timesteps // self.scheduler.num_inference_steps lowercase__ : List[Any] = self.scheduler.alphas_cumprod[t] lowercase__ : Tuple = ( self.scheduler.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.scheduler.final_alpha_cumprod ) lowercase__ : int = 1 - alpha_prod_t lowercase__ : int = self.unet(a , a )['sample'] lowercase__ : Optional[int] = (1 - alpha_prod_t_prev) ** 0.5 * model_output lowercase__ : Optional[Any] = (sample - pred_sample_direction) * alpha_prod_t_prev ** (-0.5) lowercase__ : Union[str, Any] = sample * alpha_prod_t ** 0.5 + beta_prod_t ** 0.5 * model_output return sample @staticmethod def _UpperCAmelCase ( a , a , a ) -> torch.Tensor: lowercase__ : List[Any] = acos(torch.dot(torch.flatten(a ) , torch.flatten(a ) ) / torch.norm(a ) / torch.norm(a ) ) return sin((1 - alpha) * theta ) * xa / sin(a ) + sin(alpha * theta ) * xa / sin(a )
77
"""simple docstring""" import argparse import torch from transformers import FunnelBaseModel, FunnelConfig, FunnelModel, load_tf_weights_in_funnel from transformers.utils import logging logging.set_verbosity_info() def a_ ( _lowerCAmelCase : Tuple , _lowerCAmelCase : Optional[int] , _lowerCAmelCase : List[str] , _lowerCAmelCase : Union[str, Any] ): '''simple docstring''' lowercase__ : int = FunnelConfig.from_json_file(_lowerCAmelCase ) print(f"""Building PyTorch model from configuration: {config}""" ) lowercase__ : List[Any] = FunnelBaseModel(_lowerCAmelCase ) if base_model else FunnelModel(_lowerCAmelCase ) # Load weights from tf checkpoint load_tf_weights_in_funnel(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict() , _lowerCAmelCase ) if __name__ == "__main__": _UpperCamelCase : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( "--tf_checkpoint_path", default=None, type=str, required=True, help="Path to the TensorFlow checkpoint path." ) parser.add_argument( "--config_file", default=None, type=str, required=True, help="The config json file corresponding to the pre-trained model. \nThis specifies the model architecture.", ) parser.add_argument( "--pytorch_dump_path", default=None, type=str, required=True, help="Path to the output PyTorch model." ) parser.add_argument( "--base_model", action="store_true", help="Whether you want just the base model (no decoder) or not." ) _UpperCamelCase : List[str] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path, args.base_model )
77
1
"""simple docstring""" import argparse import json import os import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils.deepspeed import DummyOptim, DummyScheduler _UpperCamelCase : Any = 16 _UpperCamelCase : Tuple = 32 def a_ ( _lowerCAmelCase : Accelerator , _lowerCAmelCase : int = 16 , _lowerCAmelCase : str = "bert-base-cased" ): '''simple docstring''' lowercase__ : int = AutoTokenizer.from_pretrained(_lowerCAmelCase ) lowercase__ : Dict = load_dataset('glue' , 'mrpc' ) def tokenize_function(_lowerCAmelCase : Optional[Any] ): # max_length=None => use the model max length (it's actually the default) lowercase__ : int = tokenizer(examples['sentence1'] , examples['sentence2'] , truncation=_lowerCAmelCase , max_length=_lowerCAmelCase ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset lowercase__ : List[str] = datasets.map( _lowerCAmelCase , batched=_lowerCAmelCase , remove_columns=['idx', 'sentence1', 'sentence2'] , load_from_cache_file=_lowerCAmelCase ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowercase__ : str = tokenized_datasets.rename_column('label' , 'labels' ) def collate_fn(_lowerCAmelCase : Optional[Any] ): # On TPU it's best to pad everything to the same length or training will be very slow. if accelerator.distributed_type == DistributedType.TPU: return tokenizer.pad(_lowerCAmelCase , padding='max_length' , max_length=128 , return_tensors='pt' ) return tokenizer.pad(_lowerCAmelCase , padding='longest' , return_tensors='pt' ) # Instantiate dataloaders. lowercase__ : List[Any] = DataLoader( tokenized_datasets['train'] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase ) lowercase__ : Optional[int] = DataLoader( tokenized_datasets['validation'] , shuffle=_lowerCAmelCase , collate_fn=_lowerCAmelCase , batch_size=_lowerCAmelCase ) return train_dataloader, eval_dataloader def a_ ( _lowerCAmelCase : int , _lowerCAmelCase : Tuple , _lowerCAmelCase : List[Any] , _lowerCAmelCase : Optional[Any] ): '''simple docstring''' model.eval() lowercase__ : List[str] = 0 for step, batch in enumerate(_lowerCAmelCase ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowercase__ : int = model(**_lowerCAmelCase ) lowercase__ : List[Any] = outputs.logits.argmax(dim=-1 ) # It is slightly faster to call this once, than multiple times lowercase__ , lowercase__ : Dict = accelerator.gather( (predictions, batch['labels']) ) # If we are in a multiprocess environment, the last batch has duplicates if accelerator.use_distributed: if step == len(_lowerCAmelCase ) - 1: lowercase__ : str = predictions[: len(eval_dataloader.dataset ) - samples_seen] lowercase__ : Optional[Any] = references[: len(eval_dataloader.dataset ) - samples_seen] else: samples_seen += references.shape[0] metric.add_batch( predictions=_lowerCAmelCase , references=_lowerCAmelCase , ) lowercase__ : Any = metric.compute() return eval_metric["accuracy"] def a_ ( _lowerCAmelCase : Any , _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : Tuple = Accelerator() # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowercase__ : str = config['lr'] lowercase__ : Any = int(config['num_epochs'] ) lowercase__ : Optional[Any] = int(config['seed'] ) lowercase__ : str = int(config['batch_size'] ) lowercase__ : Optional[int] = args.model_name_or_path set_seed(_lowerCAmelCase ) lowercase__ , lowercase__ : List[str] = get_dataloaders(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowercase__ : Dict = AutoModelForSequenceClassification.from_pretrained(_lowerCAmelCase , return_dict=_lowerCAmelCase ) # Instantiate optimizer lowercase__ : Dict = ( AdamW if accelerator.state.deepspeed_plugin is None or 'optimizer' not in accelerator.state.deepspeed_plugin.deepspeed_config else DummyOptim ) lowercase__ : Optional[Any] = optimizer_cls(params=model.parameters() , lr=_lowerCAmelCase ) if accelerator.state.deepspeed_plugin is not None: lowercase__ : Dict = accelerator.state.deepspeed_plugin.deepspeed_config[ 'gradient_accumulation_steps' ] else: lowercase__ : List[str] = 1 lowercase__ : int = (len(_lowerCAmelCase ) * num_epochs) // gradient_accumulation_steps # Instantiate scheduler if ( accelerator.state.deepspeed_plugin is None or "scheduler" not in accelerator.state.deepspeed_plugin.deepspeed_config ): lowercase__ : List[Any] = get_linear_schedule_with_warmup( optimizer=_lowerCAmelCase , num_warmup_steps=0 , num_training_steps=_lowerCAmelCase , ) else: lowercase__ : Union[str, Any] = DummyScheduler(_lowerCAmelCase , total_num_steps=_lowerCAmelCase , warmup_num_steps=0 ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowercase__ , lowercase__ , lowercase__ , lowercase__ , lowercase__ : Dict = accelerator.prepare( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) # We need to keep track of how many total steps we have iterated over lowercase__ : int = 0 # We also need to keep track of the stating epoch so files are named properly lowercase__ : Union[str, Any] = 0 lowercase__ : str = evaluate.load('glue' , 'mrpc' ) lowercase__ : Any = num_epochs if args.partial_train_epoch is not None: lowercase__ : int = args.partial_train_epoch if args.resume_from_checkpoint: accelerator.load_state(args.resume_from_checkpoint ) lowercase__ : List[Any] = args.resume_from_checkpoint.split('epoch_' )[1] lowercase__ : List[str] = '' for char in epoch_string: if char.isdigit(): state_epoch_num += char else: break lowercase__ : str = int(_lowerCAmelCase ) + 1 lowercase__ : str = evaluation_loop(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) accelerator.print('resumed checkpoint performance:' , _lowerCAmelCase ) accelerator.print('resumed checkpoint\'s scheduler\'s lr:' , lr_scheduler.get_lr()[0] ) accelerator.print('resumed optimizers\'s lr:' , optimizer.param_groups[0]['lr'] ) with open(os.path.join(args.output_dir , f"""state_{starting_epoch-1}.json""" ) , 'r' ) as f: lowercase__ : Optional[Any] = json.load(_lowerCAmelCase ) assert resumed_state["accuracy"] == accuracy, "Accuracy mismatch, loading from checkpoint failed" assert ( resumed_state["lr"] == lr_scheduler.get_lr()[0] ), "Scheduler learning rate mismatch, loading from checkpoint failed" assert ( resumed_state["optimizer_lr"] == optimizer.param_groups[0]["lr"] ), "Optimizer learning rate mismatch, loading from checkpoint failed" assert resumed_state["epoch"] == starting_epoch - 1, "Epoch mismatch, loading from checkpoint failed" return # Now we train the model lowercase__ : Optional[Any] = {} for epoch in range(_lowerCAmelCase , _lowerCAmelCase ): model.train() for step, batch in enumerate(_lowerCAmelCase ): lowercase__ : Optional[int] = model(**_lowerCAmelCase ) lowercase__ : Tuple = outputs.loss lowercase__ : Any = loss / gradient_accumulation_steps accelerator.backward(_lowerCAmelCase ) if step % gradient_accumulation_steps == 0: optimizer.step() lr_scheduler.step() optimizer.zero_grad() overall_step += 1 lowercase__ : Optional[Any] = f"""epoch_{epoch}""" lowercase__ : Optional[int] = os.path.join(args.output_dir , _lowerCAmelCase ) accelerator.save_state(_lowerCAmelCase ) lowercase__ : Union[str, Any] = evaluation_loop(_lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) lowercase__ : int = accuracy lowercase__ : int = lr_scheduler.get_lr()[0] lowercase__ : Dict = optimizer.param_groups[0]['lr'] lowercase__ : str = epoch lowercase__ : int = overall_step accelerator.print(f"""epoch {epoch}:""" , _lowerCAmelCase ) accelerator.wait_for_everyone() if accelerator.is_main_process: with open(os.path.join(args.output_dir , f"""state_{epoch}.json""" ) , 'w' ) as f: json.dump(_lowerCAmelCase , _lowerCAmelCase ) def a_ ( ): '''simple docstring''' lowercase__ : List[Any] = argparse.ArgumentParser(description='Simple example of training script tracking peak GPU memory usage.' ) parser.add_argument( '--model_name_or_path' , type=_lowerCAmelCase , default='bert-base-cased' , help='Path to pretrained model or model identifier from huggingface.co/models.' , required=_lowerCAmelCase , ) parser.add_argument( '--output_dir' , type=_lowerCAmelCase , default='.' , help='Optional save directory where all checkpoint folders will be stored. Default is the current working directory.' , ) parser.add_argument( '--resume_from_checkpoint' , type=_lowerCAmelCase , default=_lowerCAmelCase , help='If the training should continue from a checkpoint folder.' , ) parser.add_argument( '--partial_train_epoch' , type=_lowerCAmelCase , default=_lowerCAmelCase , help='If passed, the training will stop after this number of epochs.' , ) parser.add_argument( '--num_epochs' , type=_lowerCAmelCase , default=2 , help='Number of train epochs.' , ) lowercase__ : Optional[int] = parser.parse_args() lowercase__ : List[str] = {'lr': 2E-5, 'num_epochs': args.num_epochs, 'seed': 42, 'batch_size': 16} training_function(_lowerCAmelCase , _lowerCAmelCase ) if __name__ == "__main__": main()
77
"""simple docstring""" import logging import math import os from dataclasses import dataclass, field from glob import glob from typing import Optional from torch.utils.data import ConcatDataset import transformers from transformers import ( CONFIG_MAPPING, MODEL_WITH_LM_HEAD_MAPPING, AutoConfig, AutoModelWithLMHead, AutoTokenizer, DataCollatorForLanguageModeling, DataCollatorForPermutationLanguageModeling, DataCollatorForWholeWordMask, HfArgumentParser, LineByLineTextDataset, LineByLineWithRefDataset, PreTrainedTokenizer, TextDataset, Trainer, TrainingArguments, set_seed, ) from transformers.trainer_utils import is_main_process _UpperCamelCase : Any = logging.getLogger(__name__) _UpperCamelCase : Optional[Any] = list(MODEL_WITH_LM_HEAD_MAPPING.keys()) _UpperCamelCase : Any = tuple(conf.model_type for conf in MODEL_CONFIG_CLASSES) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The model checkpoint for weights initialization. Leave None if you want to train a model from" " scratch." ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "If training from scratch, pass a model type from the list: " + ", ".join(_a)} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained config name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) @dataclass class UpperCAmelCase_ : lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "The input training data file (a text file)."}) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={ "help": ( "The input training data files (multiple files in glob format). " "Very often splitting large files to smaller files can prevent tokenizer going out of memory" ) } , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input evaluation data file to evaluate the perplexity on (a text file)."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input train ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : Optional[str] = field( default=_a , metadata={"help": "An optional input eval ref data file for whole word mask in Chinese."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Whether distinct lines of text in the dataset are to be handled as distinct sequences."} , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Train with masked-language modeling loss instead of language modeling."}) lowerCamelCase__ : bool = field(default=_a , metadata={"help": "Whether ot not to use whole word mask."}) lowerCamelCase__ : float = field( default=0.15 , metadata={"help": "Ratio of tokens to mask for masked language modeling loss"}) lowerCamelCase__ : float = field( default=1 / 6 , metadata={ "help": ( "Ratio of length of a span of masked tokens to surrounding context length for permutation language" " modeling." ) } , ) lowerCamelCase__ : int = field( default=5 , metadata={"help": "Maximum length of a span of masked tokens for permutation language modeling."}) lowerCamelCase__ : int = field( default=-1 , metadata={ "help": ( "Optional input sequence length after tokenization." "The training dataset will be truncated in block of this size for training." "Default to the model max input length for single sentence inputs (take into account special tokens)." ) } , ) lowerCamelCase__ : bool = field( default=_a , metadata={"help": "Overwrite the cached training and evaluation sets"}) def a_ ( _lowerCAmelCase : DataTrainingArguments , _lowerCAmelCase : PreTrainedTokenizer , _lowerCAmelCase : bool = False , _lowerCAmelCase : Optional[str] = None , ): '''simple docstring''' def _dataset(_lowerCAmelCase : Any , _lowerCAmelCase : Any=None ): if args.line_by_line: if ref_path is not None: if not args.whole_word_mask or not args.mlm: raise ValueError('You need to set world whole masking and mlm to True for Chinese Whole Word Mask' ) return LineByLineWithRefDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , ref_path=_lowerCAmelCase , ) return LineByLineTextDataset(tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size ) else: return TextDataset( tokenizer=_lowerCAmelCase , file_path=_lowerCAmelCase , block_size=args.block_size , overwrite_cache=args.overwrite_cache , cache_dir=_lowerCAmelCase , ) if evaluate: return _dataset(args.eval_data_file , args.eval_ref_file ) elif args.train_data_files: return ConcatDataset([_dataset(_lowerCAmelCase ) for f in glob(args.train_data_files )] ) else: return _dataset(args.train_data_file , args.train_ref_file ) def a_ ( ): '''simple docstring''' lowercase__ : Any = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) lowercase__ , lowercase__ , lowercase__ : List[Any] = parser.parse_args_into_dataclasses() if data_args.eval_data_file is None and training_args.do_eval: raise ValueError( 'Cannot do evaluation without an evaluation data file. Either supply a file to --eval_data_file ' 'or remove the --do_eval argument.' ) if ( os.path.exists(training_args.output_dir ) and os.listdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir ): raise ValueError( f"""Output directory ({training_args.output_dir}) already exists and is not empty. Use""" ' --overwrite_output_dir to overcome.' ) # Setup logging logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s' , datefmt='%m/%d/%Y %H:%M:%S' , level=logging.INFO if training_args.local_rank in [-1, 0] else logging.WARN , ) logger.warning( 'Process rank: %s, device: %s, n_gpu: %s, distributed training: %s, 16-bits training: %s' , training_args.local_rank , training_args.device , training_args.n_gpu , bool(training_args.local_rank != -1 ) , training_args.fpaa , ) # Set the verbosity to info of the Transformers logger (on main process only): if is_main_process(training_args.local_rank ): transformers.utils.logging.set_verbosity_info() transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() logger.info('Training/evaluation parameters %s' , _lowerCAmelCase ) # Set seed set_seed(training_args.seed ) # Load pretrained model and tokenizer # # Distributed training: # The .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. if model_args.config_name: lowercase__ : List[Any] = AutoConfig.from_pretrained(model_args.config_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : Any = AutoConfig.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: lowercase__ : Union[str, Any] = CONFIG_MAPPING[model_args.model_type]() logger.warning('You are instantiating a new config instance from scratch.' ) if model_args.tokenizer_name: lowercase__ : Dict = AutoTokenizer.from_pretrained(model_args.tokenizer_name , cache_dir=model_args.cache_dir ) elif model_args.model_name_or_path: lowercase__ : int = AutoTokenizer.from_pretrained(model_args.model_name_or_path , cache_dir=model_args.cache_dir ) else: raise ValueError( 'You are instantiating a new tokenizer from scratch. This is not supported, but you can do it from another' ' script, save it,and load it from here, using --tokenizer_name' ) if model_args.model_name_or_path: lowercase__ : Optional[Any] = AutoModelWithLMHead.from_pretrained( model_args.model_name_or_path , from_tf=bool('.ckpt' in model_args.model_name_or_path ) , config=_lowerCAmelCase , cache_dir=model_args.cache_dir , ) else: logger.info('Training new model from scratch' ) lowercase__ : int = AutoModelWithLMHead.from_config(_lowerCAmelCase ) model.resize_token_embeddings(len(_lowerCAmelCase ) ) if config.model_type in ["bert", "roberta", "distilbert", "camembert"] and not data_args.mlm: raise ValueError( 'BERT and RoBERTa-like models do not have LM heads but masked LM heads. They must be run using the' '--mlm flag (masked language modeling).' ) if data_args.block_size <= 0: lowercase__ : List[str] = tokenizer.max_len # Our input block size will be the max possible for the model else: lowercase__ : int = min(data_args.block_size , tokenizer.max_len ) # Get datasets lowercase__ : Tuple = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_train else None ) lowercase__ : Optional[Any] = ( get_dataset(_lowerCAmelCase , tokenizer=_lowerCAmelCase , evaluate=_lowerCAmelCase , cache_dir=model_args.cache_dir ) if training_args.do_eval else None ) if config.model_type == "xlnet": lowercase__ : List[Any] = DataCollatorForPermutationLanguageModeling( tokenizer=_lowerCAmelCase , plm_probability=data_args.plm_probability , max_span_length=data_args.max_span_length , ) else: if data_args.mlm and data_args.whole_word_mask: lowercase__ : List[str] = DataCollatorForWholeWordMask( tokenizer=_lowerCAmelCase , mlm_probability=data_args.mlm_probability ) else: lowercase__ : str = DataCollatorForLanguageModeling( tokenizer=_lowerCAmelCase , mlm=data_args.mlm , mlm_probability=data_args.mlm_probability ) # Initialize our Trainer lowercase__ : Optional[int] = Trainer( model=_lowerCAmelCase , args=_lowerCAmelCase , data_collator=_lowerCAmelCase , train_dataset=_lowerCAmelCase , eval_dataset=_lowerCAmelCase , prediction_loss_only=_lowerCAmelCase , ) # Training if training_args.do_train: lowercase__ : List[str] = ( model_args.model_name_or_path if model_args.model_name_or_path is not None and os.path.isdir(model_args.model_name_or_path ) else None ) trainer.train(model_path=_lowerCAmelCase ) trainer.save_model() # For convenience, we also re-save the tokenizer to the same directory, # so that you can share your model easily on huggingface.co/models =) if trainer.is_world_master(): tokenizer.save_pretrained(training_args.output_dir ) # Evaluation lowercase__ : List[Any] = {} if training_args.do_eval: logger.info('*** Evaluate ***' ) lowercase__ : Dict = trainer.evaluate() lowercase__ : List[Any] = math.exp(eval_output['eval_loss'] ) lowercase__ : int = {'perplexity': perplexity} lowercase__ : int = os.path.join(training_args.output_dir , 'eval_results_lm.txt' ) if trainer.is_world_master(): with open(_lowerCAmelCase , 'w' ) as writer: logger.info('***** Eval results *****' ) for key in sorted(result.keys() ): logger.info(' %s = %s' , _lowerCAmelCase , str(result[key] ) ) writer.write('%s = %s\n' % (key, str(result[key] )) ) results.update(_lowerCAmelCase ) return results def a_ ( _lowerCAmelCase : Optional[Any] ): '''simple docstring''' main() if __name__ == "__main__": main()
77
1
"""simple docstring""" import json import os import tempfile import datasets from utils import generate_example_dataset, get_duration _UpperCamelCase : Any = 5_00_00 _UpperCamelCase : Any = 50_00 _UpperCamelCase , _UpperCamelCase : List[str] = os.path.split(__file__) _UpperCamelCase : Dict = os.path.join(RESULTS_BASEPATH, "results", RESULTS_FILENAME.replace(".py", ".json")) @get_duration def a_ ( _lowerCAmelCase : datasets.Dataset , _lowerCAmelCase : Dict ): '''simple docstring''' for i in range(_lowerCAmelCase ): lowercase__ : int = dataset[i] @get_duration def a_ ( _lowerCAmelCase : datasets.Dataset , _lowerCAmelCase : Optional[Any] , _lowerCAmelCase : Dict ): '''simple docstring''' for i in range(0 , len(_lowerCAmelCase ) , _lowerCAmelCase ): lowercase__ : Tuple = dataset[i : i + batch_size] @get_duration def a_ ( _lowerCAmelCase : datasets.Dataset , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Optional[int] ): '''simple docstring''' with dataset.formatted_as(type=_lowerCAmelCase ): for i in range(_lowerCAmelCase ): lowercase__ : Any = dataset[i] @get_duration def a_ ( _lowerCAmelCase : datasets.Dataset , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict , _lowerCAmelCase : Dict ): '''simple docstring''' with dataset.formatted_as(type=_lowerCAmelCase ): for i in range(0 , _lowerCAmelCase , _lowerCAmelCase ): lowercase__ : Optional[int] = dataset[i : i + batch_size] def a_ ( ): '''simple docstring''' lowercase__ : Any = {'num examples': SPEED_TEST_N_EXAMPLES} lowercase__ : int = [ (read, {'length': SMALL_TEST}), (read, {'length': SPEED_TEST_N_EXAMPLES}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 10}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 100}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 1000}), (read_formatted, {'type': 'numpy', 'length': SMALL_TEST}), (read_formatted, {'type': 'pandas', 'length': SMALL_TEST}), (read_formatted, {'type': 'torch', 'length': SMALL_TEST}), (read_formatted, {'type': 'tensorflow', 'length': SMALL_TEST}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 10}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 1000}), ] lowercase__ : Optional[Any] = [ (read, {'length': SMALL_TEST}), (read, {'length': SPEED_TEST_N_EXAMPLES}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 10}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 100}), (read_batch, {'length': SPEED_TEST_N_EXAMPLES, 'batch_size': 1000}), (read_formatted, {'type': 'numpy', 'length': SMALL_TEST}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 10}), (read_formatted_batch, {'type': 'numpy', 'length': SMALL_TEST, 'batch_size': 1000}), ] with tempfile.TemporaryDirectory() as tmp_dir: print('generating dataset' ) lowercase__ : List[str] = datasets.Features( {'list': datasets.Sequence(datasets.Value('float32' ) ), 'numbers': datasets.Value('float32' )} ) lowercase__ : str = generate_example_dataset( os.path.join(_lowerCAmelCase , 'dataset.arrow' ) , _lowerCAmelCase , num_examples=_lowerCAmelCase , seq_shapes={'list': (100,)} , ) print('first set of iterations' ) for func, kwargs in functions: print(func.__name__ , str(_lowerCAmelCase ) ) lowercase__ : Dict = func(_lowerCAmelCase , **_lowerCAmelCase ) print('shuffling dataset' ) lowercase__ : Union[str, Any] = dataset.shuffle() print('Second set of iterations (after shuffling' ) for func, kwargs in functions_shuffled: print('shuffled ' , func.__name__ , str(_lowerCAmelCase ) ) lowercase__ : str = func( _lowerCAmelCase , **_lowerCAmelCase ) with open(_lowerCAmelCase , 'wb' ) as f: f.write(json.dumps(_lowerCAmelCase ).encode('utf-8' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_iterating()
77
"""simple docstring""" import math import flax.linen as nn import jax.numpy as jnp def a_ ( _lowerCAmelCase : jnp.ndarray , _lowerCAmelCase : int , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1 , _lowerCAmelCase : float = 1.0E4 , _lowerCAmelCase : bool = False , _lowerCAmelCase : float = 1.0 , ): '''simple docstring''' assert timesteps.ndim == 1, "Timesteps should be a 1d-array" assert embedding_dim % 2 == 0, f"""Embedding dimension {embedding_dim} should be even""" lowercase__ : Optional[Any] = float(embedding_dim // 2 ) lowercase__ : Optional[Any] = math.log(max_timescale / min_timescale ) / (num_timescales - freq_shift) lowercase__ : Any = min_timescale * jnp.exp(jnp.arange(_lowerCAmelCase , dtype=jnp.floataa ) * -log_timescale_increment ) lowercase__ : Dict = jnp.expand_dims(_lowerCAmelCase , 1 ) * jnp.expand_dims(_lowerCAmelCase , 0 ) # scale embeddings lowercase__ : List[str] = scale * emb if flip_sin_to_cos: lowercase__ : Dict = jnp.concatenate([jnp.cos(_lowerCAmelCase ), jnp.sin(_lowerCAmelCase )] , axis=1 ) else: lowercase__ : Optional[int] = jnp.concatenate([jnp.sin(_lowerCAmelCase ), jnp.cos(_lowerCAmelCase )] , axis=1 ) lowercase__ : List[Any] = jnp.reshape(_lowerCAmelCase , [jnp.shape(_lowerCAmelCase )[0], embedding_dim] ) return signal class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : jnp.dtype = jnp.floataa @nn.compact def __call__( self , a ) -> Any: lowercase__ : str = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_1' )(a ) lowercase__ : Union[str, Any] = nn.silu(a ) lowercase__ : Optional[Any] = nn.Dense(self.time_embed_dim , dtype=self.dtype , name='linear_2' )(a ) return temb class UpperCAmelCase_ ( nn.Module): lowerCamelCase__ : int = 3_2 lowerCamelCase__ : bool = False lowerCamelCase__ : float = 1 @nn.compact def __call__( self , a ) -> str: return get_sinusoidal_embeddings( a , embedding_dim=self.dim , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.freq_shift )
77
1
"""simple docstring""" def a_ ( _lowerCAmelCase : int ): '''simple docstring''' lowercase__ : list[list[int]] = [[0 for _ in range(_lowerCAmelCase )] for _ in range(m + 1 )] for i in range(m + 1 ): lowercase__ : List[Any] = 1 for n in range(m + 1 ): for k in range(1 , _lowerCAmelCase ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: _UpperCamelCase : str = int(input("Enter a number: ").strip()) print(partition(n)) except ValueError: print("Please enter a number.") else: try: _UpperCamelCase : Union[str, Any] = int(sys.argv[1]) print(partition(n)) except ValueError: print("Please pass a number.")
77
"""simple docstring""" from __future__ import annotations import math from collections.abc import Callable def a_ ( _lowerCAmelCase : Callable[[int | float], int | float] , _lowerCAmelCase : int | float , _lowerCAmelCase : int | float , _lowerCAmelCase : int = 100 , ): '''simple docstring''' lowercase__ : Dict = x_start lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) lowercase__ : Optional[Any] = 0.0 for _ in range(_lowerCAmelCase ): # Approximates curve as a sequence of linear lines and sums their length lowercase__ : Union[str, Any] = (x_end - x_start) / steps + xa lowercase__ : Union[str, Any] = fnc(_lowerCAmelCase ) length += math.hypot(xa - xa , fxa - fxa ) # Increment step lowercase__ : Union[str, Any] = xa lowercase__ : int = fxa return length if __name__ == "__main__": def a_ ( _lowerCAmelCase : List[Any] ): '''simple docstring''' return math.sin(10 * x ) print("f(x) = sin(10 * x)") print("The length of the curve from x = -10 to x = 10 is:") _UpperCamelCase : str = 10 while i <= 10_00_00: print(f'''With {i} steps: {line_length(f, -10, 10, i)}''') i *= 10
77
1