code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
import numpy as np from nltk.translate import meteor_score import datasets from datasets.config import importlib_metadata, version lowerCamelCase__ : int = version.parse(importlib_metadata.version('nltk')) if NLTK_VERSION >= version.Version('3.6.4'): from nltk import word_tokenize lowerCamelCase__ : Optional[int] = '\\n@inproceedings{banarjee2005,\n title = {{METEOR}: An Automatic Metric for {MT} Evaluation with Improved Correlation with Human Judgments},\n author = {Banerjee, Satanjeev and Lavie, Alon},\n booktitle = {Proceedings of the {ACL} Workshop on Intrinsic and Extrinsic Evaluation Measures for Machine Translation and/or Summarization},\n month = jun,\n year = {2005},\n address = {Ann Arbor, Michigan},\n publisher = {Association for Computational Linguistics},\n url = {https://www.aclweb.org/anthology/W05-0909},\n pages = {65--72},\n}\n' lowerCamelCase__ : Union[str, Any] = '\\nMETEOR, an automatic metric for machine translation evaluation\nthat is based on a generalized concept of unigram matching between the\nmachine-produced translation and human-produced reference translations.\nUnigrams can be matched based on their surface forms, stemmed forms,\nand meanings; furthermore, METEOR can be easily extended to include more\nadvanced matching strategies. Once all generalized unigram matches\nbetween the two strings have been found, METEOR computes a score for\nthis matching using a combination of unigram-precision, unigram-recall, and\na measure of fragmentation that is designed to directly capture how\nwell-ordered the matched words in the machine translation are in relation\nto the reference.\n\nMETEOR gets an R correlation value of 0.347 with human evaluation on the Arabic\ndata and 0.331 on the Chinese data. This is shown to be an improvement on\nusing simply unigram-precision, unigram-recall and their harmonic F1\ncombination.\n' lowerCamelCase__ : Dict = '\nComputes METEOR score of translated segments against one or more references.\nArgs:\n predictions: list of predictions to score. Each prediction\n should be a string with tokens separated by spaces.\n references: list of reference for each prediction. Each\n reference should be a string with tokens separated by spaces.\n alpha: Parameter for controlling relative weights of precision and recall. default: 0.9\n beta: Parameter for controlling shape of penalty as a function of fragmentation. default: 3\n gamma: Relative weight assigned to fragmentation penalty. default: 0.5\nReturns:\n \'meteor\': meteor score.\nExamples:\n\n >>> meteor = datasets.load_metric(\'meteor\')\n >>> predictions = ["It is a guide to action which ensures that the military always obeys the commands of the party"]\n >>> references = ["It is a guide to action that ensures that the military will forever heed Party commands"]\n >>> results = meteor.compute(predictions=predictions, references=references)\n >>> print(round(results["meteor"], 4))\n 0.6944\n' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class lowerCamelCase_ ( datasets.Metric ): '''simple docstring''' def lowerCAmelCase_ ( self : str ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { 'predictions': datasets.Value('string' , id='sequence' ), 'references': datasets.Value('string' , id='sequence' ), } ) , codebase_urls=['https://github.com/nltk/nltk/blob/develop/nltk/translate/meteor_score.py'] , reference_urls=[ 'https://www.nltk.org/api/nltk.translate.html#module-nltk.translate.meteor_score', 'https://en.wikipedia.org/wiki/METEOR', ] , ) def lowerCAmelCase_ ( self : List[Any] , _lowerCAmelCase : str ): import nltk nltk.download('wordnet' ) if NLTK_VERSION >= version.Version('3.6.5' ): nltk.download('punkt' ) if NLTK_VERSION >= version.Version('3.6.6' ): nltk.download('omw-1.4' ) def lowerCAmelCase_ ( self : Optional[int] , _lowerCAmelCase : Union[str, Any] , _lowerCAmelCase : Any , _lowerCAmelCase : Union[str, Any]=0.9 , _lowerCAmelCase : Optional[int]=3 , _lowerCAmelCase : List[Any]=0.5 ): if NLTK_VERSION >= version.Version('3.6.5' ): SCREAMING_SNAKE_CASE_ = [ meteor_score.single_meteor_score( word_tokenize(_lowerCAmelCase ) , word_tokenize(_lowerCAmelCase ) , alpha=_lowerCAmelCase , beta=_lowerCAmelCase , gamma=_lowerCAmelCase ) for ref, pred in zip(_lowerCAmelCase , _lowerCAmelCase ) ] else: SCREAMING_SNAKE_CASE_ = [ meteor_score.single_meteor_score(_lowerCAmelCase , _lowerCAmelCase , alpha=_lowerCAmelCase , beta=_lowerCAmelCase , gamma=_lowerCAmelCase ) for ref, pred in zip(_lowerCAmelCase , _lowerCAmelCase ) ] return {"meteor": np.mean(_lowerCAmelCase )}
225
import itertools import json import linecache import os import pickle import re import socket import string from collections import Counter from logging import getLogger from pathlib import Path from typing import Callable, Dict, Iterable, List import git import torch from torch.utils.data import Dataset from transformers import BartTokenizer, RagTokenizer, TaTokenizer def UpperCAmelCase_ ( __UpperCAmelCase : Dict , __UpperCAmelCase : Dict , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any]=True , __UpperCAmelCase : Any="pt" ) -> List[str]: SCREAMING_SNAKE_CASE_ = {'add_prefix_space': True} if isinstance(__UpperCAmelCase , __UpperCAmelCase ) and not line.startswith(' ' ) else {} SCREAMING_SNAKE_CASE_ = padding_side return tokenizer( [line] , max_length=__UpperCAmelCase , padding='max_length' if pad_to_max_length else None , truncation=__UpperCAmelCase , return_tensors=__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , **__UpperCAmelCase , ) def UpperCAmelCase_ ( __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Dict=None , ) -> Optional[Any]: SCREAMING_SNAKE_CASE_ = input_ids.ne(__UpperCAmelCase ).any(dim=0 ) if attention_mask is None: return input_ids[:, keep_column_mask] else: return (input_ids[:, keep_column_mask], attention_mask[:, keep_column_mask]) class lowerCamelCase_ ( _SCREAMING_SNAKE_CASE ): '''simple docstring''' def __init__( self : List[Any] , _lowerCAmelCase : Tuple , _lowerCAmelCase : int , _lowerCAmelCase : int , _lowerCAmelCase : Dict , _lowerCAmelCase : Union[str, Any]="train" , _lowerCAmelCase : Optional[Any]=None , _lowerCAmelCase : List[str]=None , _lowerCAmelCase : Optional[int]=None , _lowerCAmelCase : Optional[int]="" , ): super().__init__() SCREAMING_SNAKE_CASE_ = Path(_lowerCAmelCase ).joinpath(type_path + '.source' ) SCREAMING_SNAKE_CASE_ = Path(_lowerCAmelCase ).joinpath(type_path + '.target' ) SCREAMING_SNAKE_CASE_ = self.get_char_lens(self.src_file ) SCREAMING_SNAKE_CASE_ = max_source_length SCREAMING_SNAKE_CASE_ = max_target_length assert min(self.src_lens ) > 0, F"found empty line in {self.src_file}" SCREAMING_SNAKE_CASE_ = tokenizer SCREAMING_SNAKE_CASE_ = prefix if n_obs is not None: SCREAMING_SNAKE_CASE_ = self.src_lens[:n_obs] SCREAMING_SNAKE_CASE_ = src_lang SCREAMING_SNAKE_CASE_ = tgt_lang def __len__( self : Tuple ): return len(self.src_lens ) def __getitem__( self : List[str] , _lowerCAmelCase : Any ): SCREAMING_SNAKE_CASE_ = index + 1 # linecache starts at 1 SCREAMING_SNAKE_CASE_ = self.prefix + linecache.getline(str(self.src_file ) , _lowerCAmelCase ).rstrip('\n' ) SCREAMING_SNAKE_CASE_ = linecache.getline(str(self.tgt_file ) , _lowerCAmelCase ).rstrip('\n' ) assert source_line, F"empty source line for index {index}" assert tgt_line, F"empty tgt line for index {index}" # Need to add eos token manually for T5 if isinstance(self.tokenizer , _lowerCAmelCase ): source_line += self.tokenizer.eos_token tgt_line += self.tokenizer.eos_token # Pad source and target to the right SCREAMING_SNAKE_CASE_ = ( self.tokenizer.question_encoder if isinstance(self.tokenizer , _lowerCAmelCase ) else self.tokenizer ) SCREAMING_SNAKE_CASE_ = self.tokenizer.generator if isinstance(self.tokenizer , _lowerCAmelCase ) else self.tokenizer SCREAMING_SNAKE_CASE_ = encode_line(_lowerCAmelCase , _lowerCAmelCase , self.max_source_length , 'right' ) SCREAMING_SNAKE_CASE_ = encode_line(_lowerCAmelCase , _lowerCAmelCase , self.max_target_length , 'right' ) SCREAMING_SNAKE_CASE_ = source_inputs['input_ids'].squeeze() SCREAMING_SNAKE_CASE_ = target_inputs['input_ids'].squeeze() SCREAMING_SNAKE_CASE_ = source_inputs['attention_mask'].squeeze() return { "input_ids": source_ids, "attention_mask": src_mask, "decoder_input_ids": target_ids, } @staticmethod def lowerCAmelCase_ ( _lowerCAmelCase : Optional[int] ): return [len(_lowerCAmelCase ) for x in Path(_lowerCAmelCase ).open().readlines()] def lowerCAmelCase_ ( self : int , _lowerCAmelCase : Optional[int] ): SCREAMING_SNAKE_CASE_ = torch.stack([x['input_ids'] for x in batch] ) SCREAMING_SNAKE_CASE_ = torch.stack([x['attention_mask'] for x in batch] ) SCREAMING_SNAKE_CASE_ = torch.stack([x['decoder_input_ids'] for x in batch] ) SCREAMING_SNAKE_CASE_ = ( self.tokenizer.generator.pad_token_id if isinstance(self.tokenizer , _lowerCAmelCase ) else self.tokenizer.pad_token_id ) SCREAMING_SNAKE_CASE_ = ( self.tokenizer.question_encoder.pad_token_id if isinstance(self.tokenizer , _lowerCAmelCase ) else self.tokenizer.pad_token_id ) SCREAMING_SNAKE_CASE_ = trim_batch(_lowerCAmelCase , _lowerCAmelCase ) SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ = trim_batch(_lowerCAmelCase , _lowerCAmelCase , attention_mask=_lowerCAmelCase ) SCREAMING_SNAKE_CASE_ = { 'input_ids': source_ids, 'attention_mask': source_mask, 'decoder_input_ids': y, } return batch lowerCamelCase__ : List[str] = getLogger(__name__) def UpperCAmelCase_ ( __UpperCAmelCase : List[List] ) -> Tuple: return list(itertools.chain.from_iterable(__UpperCAmelCase ) ) def UpperCAmelCase_ ( __UpperCAmelCase : str ) -> None: SCREAMING_SNAKE_CASE_ = get_git_info() save_json(__UpperCAmelCase , os.path.join(__UpperCAmelCase , 'git_log.json' ) ) def UpperCAmelCase_ ( __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Optional[int]=4 , **__UpperCAmelCase : Tuple ) -> str: with open(__UpperCAmelCase , 'w' ) as f: json.dump(__UpperCAmelCase , __UpperCAmelCase , indent=__UpperCAmelCase , **__UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : List[str] ) -> int: with open(__UpperCAmelCase ) as f: return json.load(__UpperCAmelCase ) def UpperCAmelCase_ ( ) -> Tuple: SCREAMING_SNAKE_CASE_ = git.Repo(search_parent_directories=__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = { 'repo_id': str(__UpperCAmelCase ), 'repo_sha': str(repo.head.object.hexsha ), 'repo_branch': str(repo.active_branch ), 'hostname': str(socket.gethostname() ), } return repo_infos def UpperCAmelCase_ ( __UpperCAmelCase : Callable , __UpperCAmelCase : Iterable ) -> List: return list(map(__UpperCAmelCase , __UpperCAmelCase ) ) def UpperCAmelCase_ ( __UpperCAmelCase : str , __UpperCAmelCase : Dict ) -> Dict: with open(__UpperCAmelCase , 'wb' ) as f: return pickle.dump(__UpperCAmelCase , __UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : List[Any] ) -> Any: def remove_articles(__UpperCAmelCase : Any ): return re.sub(r'\b(a|an|the)\b' , ' ' , __UpperCAmelCase ) def white_space_fix(__UpperCAmelCase : List[str] ): return " ".join(text.split() ) def remove_punc(__UpperCAmelCase : List[Any] ): SCREAMING_SNAKE_CASE_ = set(string.punctuation ) return "".join(ch for ch in text if ch not in exclude ) def lower(__UpperCAmelCase : Union[str, Any] ): return text.lower() return white_space_fix(remove_articles(remove_punc(lower(__UpperCAmelCase ) ) ) ) def UpperCAmelCase_ ( __UpperCAmelCase : int , __UpperCAmelCase : Any ) -> List[str]: SCREAMING_SNAKE_CASE_ = normalize_answer(__UpperCAmelCase ).split() SCREAMING_SNAKE_CASE_ = normalize_answer(__UpperCAmelCase ).split() SCREAMING_SNAKE_CASE_ = Counter(__UpperCAmelCase ) & Counter(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = sum(common.values() ) if num_same == 0: return 0 SCREAMING_SNAKE_CASE_ = 1.0 * num_same / len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = 1.0 * num_same / len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = (2 * precision * recall) / (precision + recall) return fa def UpperCAmelCase_ ( __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] ) -> Any: return normalize_answer(__UpperCAmelCase ) == normalize_answer(__UpperCAmelCase ) def UpperCAmelCase_ ( __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] ) -> Dict: assert len(__UpperCAmelCase ) == len(__UpperCAmelCase ) SCREAMING_SNAKE_CASE_ = 0 for hypo, pred in zip(__UpperCAmelCase , __UpperCAmelCase ): em += exact_match_score(__UpperCAmelCase , __UpperCAmelCase ) if len(__UpperCAmelCase ) > 0: em /= len(__UpperCAmelCase ) return {"em": em} def UpperCAmelCase_ ( __UpperCAmelCase : Union[str, Any] ) -> Dict: return model_prefix.startswith('rag' ) def UpperCAmelCase_ ( __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : List[str] ) -> Tuple: SCREAMING_SNAKE_CASE_ = {p: p for p in extra_params} # T5 models don't have `dropout` param, they have `dropout_rate` instead SCREAMING_SNAKE_CASE_ = 'dropout_rate' for p in extra_params: if getattr(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): if not hasattr(__UpperCAmelCase , __UpperCAmelCase ) and not hasattr(__UpperCAmelCase , equivalent_param[p] ): logger.info('config doesn\'t have a `{}` attribute'.format(__UpperCAmelCase ) ) delattr(__UpperCAmelCase , __UpperCAmelCase ) continue SCREAMING_SNAKE_CASE_ = p if hasattr(__UpperCAmelCase , __UpperCAmelCase ) else equivalent_param[p] setattr(__UpperCAmelCase , __UpperCAmelCase , getattr(__UpperCAmelCase , __UpperCAmelCase ) ) delattr(__UpperCAmelCase , __UpperCAmelCase ) return hparams, config
225
1
'''simple docstring''' from __future__ import annotations class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase ) -> None: A_ : Optional[int] = order # a_{0} ... a_{k} A_ : List[str] = [1.0] + [0.0] * order # b_{0} ... b_{k} A_ : int = [1.0] + [0.0] * order # x[n-1] ... x[n-k] A_ : Tuple = [0.0] * self.order # y[n-1] ... y[n-k] A_ : Optional[Any] = [0.0] * self.order def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase ) -> None: if len(_lowerCamelCase ) < self.order: A_ : int = [1.0, *a_coeffs] if len(_lowerCamelCase ) != self.order + 1: A_ : Optional[int] = ( F"Expected a_coeffs to have {self.order + 1} elements " F"for {self.order}-order filter, got {len(_lowerCamelCase )}" ) raise ValueError(_lowerCamelCase ) if len(_lowerCamelCase ) != self.order + 1: A_ : List[Any] = ( F"Expected b_coeffs to have {self.order + 1} elements " F"for {self.order}-order filter, got {len(_lowerCamelCase )}" ) raise ValueError(_lowerCamelCase ) A_ : List[str] = a_coeffs A_ : Optional[Any] = b_coeffs def UpperCAmelCase_ ( self , _lowerCamelCase ) -> float: A_ : Any = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) A_ : str = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] A_ : List[Any] = self.input_history[:-1] A_ : str = self.output_history[:-1] A_ : List[Any] = sample A_ : Any = result return result
164
'''simple docstring''' def UpperCAmelCase ( a_ , a_ ) -> int: """simple docstring""" A_ : int = 1 # To kept the Calculated Value # Since C(n, k) = C(n, n-k) if k > (n - k): A_ : Tuple = n - k # Calculate C(n,k) for i in range(a_ ): result *= n - i result //= i + 1 return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return binomial_coefficient(2 * node_count , a_ ) // (node_count + 1) def UpperCAmelCase ( a_ ) -> int: """simple docstring""" if n < 0: raise ValueError("""factorial() not defined for negative values""" ) A_ : Union[str, Any] = 1 for i in range(1 , n + 1 ): result *= i return result def UpperCAmelCase ( a_ ) -> int: """simple docstring""" return catalan_number(a_ ) * factorial(a_ ) if __name__ == "__main__": UpperCamelCase__ : Any = int(input('Enter the number of nodes: ').strip() or 0) if node_count <= 0: raise ValueError('We need some nodes to work with.') print( f'Given {node_count} nodes, there are {binary_tree_count(node_count)} ' f'binary trees and {catalan_number(node_count)} binary search trees.' )
164
1
UpperCAmelCase__ = {0: [2, 3], 1: [0], 2: [1], 3: [4], 4: []} UpperCAmelCase__ = {0: [1, 2, 3], 1: [2], 2: [0], 3: [4], 4: [5], 5: [3]} def A ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple ) -> list[int]: '''simple docstring''' _UpperCAmelCase = True _UpperCAmelCase = [] for neighbour in graph[vert]: if not visited[neighbour]: order += topology_sort(snake_case__ , snake_case__ , snake_case__ ) order.append(snake_case__ ) return order def A ( _UpperCAmelCase : Any , _UpperCAmelCase : Any , _UpperCAmelCase : Optional[int] ) -> list[int]: '''simple docstring''' _UpperCAmelCase = True _UpperCAmelCase = [vert] for neighbour in reversed_graph[vert]: if not visited[neighbour]: component += find_components(snake_case__ , snake_case__ , snake_case__ ) return component def A ( _UpperCAmelCase : Optional[Any] ) -> list[list[int]]: '''simple docstring''' _UpperCAmelCase = len(snake_case__ ) * [False] _UpperCAmelCase = {vert: [] for vert in range(len(snake_case__ ) )} for vert, neighbours in graph.items(): for neighbour in neighbours: reversed_graph[neighbour].append(snake_case__ ) _UpperCAmelCase = [] for i, was_visited in enumerate(snake_case__ ): if not was_visited: order += topology_sort(snake_case__ , snake_case__ , snake_case__ ) _UpperCAmelCase = [] _UpperCAmelCase = len(snake_case__ ) * [False] for i in range(len(snake_case__ ) ): _UpperCAmelCase = order[len(snake_case__ ) - i - 1] if not visited[vert]: _UpperCAmelCase = find_components(snake_case__ , snake_case__ , snake_case__ ) components_list.append(snake_case__ ) return components_list
339
"""simple docstring""" import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging lowerCAmelCase : int = logging.get_logger(__name__) lowerCAmelCase : List[str] = { """asapp/sew-d-tiny-100k""": """https://huggingface.co/asapp/sew-d-tiny-100k/resolve/main/config.json""", # See all SEW-D models at https://huggingface.co/models?filter=sew-d } class __magic_name__ ( UpperCAmelCase__ ): '''simple docstring''' __UpperCamelCase = "sew-d" def __init__( self , _a=32 , _a=768 , _a=12 , _a=12 , _a=3_072 , _a=2 , _a=512 , _a=256 , _a=True , _a=True , _a=("p2c", "c2p") , _a="layer_norm" , _a="gelu_python" , _a=0.1 , _a=0.1 , _a=0.1 , _a=0.0 , _a=0.1 , _a=0.02 , _a=1e-7 , _a=1e-5 , _a="group" , _a="gelu" , _a=(64, 128, 128, 128, 128, 256, 256, 256, 256, 512, 512, 512, 512) , _a=(5, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1, 2, 1) , _a=(10, 3, 1, 3, 1, 3, 1, 3, 1, 2, 1, 2, 1) , _a=False , _a=128 , _a=16 , _a=True , _a=0.05 , _a=10 , _a=2 , _a=0.0 , _a=10 , _a=0 , _a="mean" , _a=False , _a=False , _a=256 , _a=0 , _a=1 , _a=2 , **_a , ): """simple docstring""" super().__init__(**_a , pad_token_id=_a , bos_token_id=_a , eos_token_id=_a ) lowerCamelCase = hidden_size lowerCamelCase = feat_extract_norm lowerCamelCase = feat_extract_activation lowerCamelCase = list(_a ) lowerCamelCase = list(_a ) lowerCamelCase = list(_a ) lowerCamelCase = conv_bias lowerCamelCase = num_conv_pos_embeddings lowerCamelCase = num_conv_pos_embedding_groups lowerCamelCase = len(self.conv_dim ) lowerCamelCase = num_hidden_layers lowerCamelCase = intermediate_size lowerCamelCase = squeeze_factor lowerCamelCase = max_position_embeddings lowerCamelCase = position_buckets lowerCamelCase = share_att_key lowerCamelCase = relative_attention lowerCamelCase = norm_rel_ebd lowerCamelCase = list(_a ) lowerCamelCase = hidden_act lowerCamelCase = num_attention_heads lowerCamelCase = hidden_dropout lowerCamelCase = attention_dropout lowerCamelCase = activation_dropout lowerCamelCase = feat_proj_dropout lowerCamelCase = final_dropout lowerCamelCase = layer_norm_eps lowerCamelCase = feature_layer_norm_eps lowerCamelCase = initializer_range lowerCamelCase = vocab_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( """Configuration for convolutional layers is incorrect.""" """It is required that `len(config.conv_dim)` == `len(config.conv_stride)` == `len(config.conv_kernel)`,""" f'but is `len(config.conv_dim) = {len(self.conv_dim )}`, `len(config.conv_stride)' f'= {len(self.conv_stride )}`, `len(config.conv_kernel) = {len(self.conv_kernel )}`.' ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 lowerCamelCase = apply_spec_augment lowerCamelCase = mask_time_prob lowerCamelCase = mask_time_length lowerCamelCase = mask_time_min_masks lowerCamelCase = mask_feature_prob lowerCamelCase = mask_feature_length lowerCamelCase = mask_feature_min_masks # ctc loss lowerCamelCase = ctc_loss_reduction lowerCamelCase = ctc_zero_infinity # sequence classification lowerCamelCase = use_weighted_layer_sum lowerCamelCase = classifier_proj_size @property def _lowerCAmelCase ( self ): """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
291
0
'''simple docstring''' import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeqaSeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging _lowerCamelCase = logging.get_logger(__name__) def a__ ( _SCREAMING_SNAKE_CASE : nn.ModuleList , _SCREAMING_SNAKE_CASE : nn.ModuleList , _SCREAMING_SNAKE_CASE : List[int] ) -> None: """simple docstring""" UpperCAmelCase_ : Optional[int] = nn.ModuleList([src_layers[i] for i in layers_to_copy] ) assert len(_SCREAMING_SNAKE_CASE ) == len(_SCREAMING_SNAKE_CASE ), F'''{len(_SCREAMING_SNAKE_CASE )} != {len(_SCREAMING_SNAKE_CASE )}''' dest_layers.load_state_dict(layers_to_copy.state_dict() ) _lowerCamelCase = { # maps num layers in teacher -> num_layers in student -> which teacher layers to copy. # 12: bart, 16: pegasus, 6: marian/Helsinki-NLP 12: { 1: [0], # This says that if the teacher has 12 layers and the student has 1, copy layer 0 of the teacher 2: [0, 6], 3: [0, 6, 11], 4: [0, 4, 8, 11], 6: [0, 2, 4, 7, 9, 11], 9: [0, 1, 2, 4, 5, 7, 9, 10, 11], 12: list(range(12)), }, 16: { # maps num layers in student -> which teacher layers to copy 1: [0], 2: [0, 15], 3: [0, 8, 15], 4: [0, 5, 10, 15], 6: [0, 3, 6, 9, 12, 15], 8: [0, 2, 4, 6, 8, 10, 12, 15], 9: [0, 1, 3, 5, 7, 9, 11, 13, 15], 12: [0, 1, 2, 3, 4, 5, 6, 7, 9, 11, 13, 15], 16: list(range(16)), }, 6: {1: [0], 2: [0, 5], 3: [0, 2, 5], 4: [0, 1, 3, 5], 6: list(range(6))}, } _lowerCamelCase = { # maps num layers in student -> which teacher layers to copy. 6: {1: [5], 2: [3, 5], 3: [1, 4, 5], 4: [1, 2, 4, 5]}, 12: {1: [11], 2: [5, 11], 3: [3, 7, 11], 6: [1, 3, 5, 8, 10, 11]}, 16: {1: [15], 4: [4, 9, 12, 15], 8: [1, 3, 5, 7, 9, 11, 13, 15]}, } def a__ ( _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Optional[int] ) -> Dict: """simple docstring""" try: UpperCAmelCase_ : str = LAYERS_TO_COPY[n_teacher][n_student] return val except KeyError: if n_student != n_teacher: warnings.warn( F'''no hardcoded layers to copy for teacher {n_teacher} -> student {n_student}, defaulting to first''' F''' {n_student}''' ) return list(range(_SCREAMING_SNAKE_CASE ) ) def a__ ( _SCREAMING_SNAKE_CASE : Dict , _SCREAMING_SNAKE_CASE : Dict ) -> List[int]: """simple docstring""" if n_student > n_teacher: raise ValueError(F'''Cannot perform intermediate supervision for student {n_student} > teacher {n_teacher}''' ) elif n_teacher == n_student: return list(range(_SCREAMING_SNAKE_CASE ) ) elif n_student == 1: return [n_teacher - 1] else: return LAYERS_TO_SUPERVISE[n_teacher][n_student] def a__ ( _SCREAMING_SNAKE_CASE : Union[str, PreTrainedModel] , _SCREAMING_SNAKE_CASE : Union[str, Path] = "student" , _SCREAMING_SNAKE_CASE : Union[int, None] = None , _SCREAMING_SNAKE_CASE : Union[int, None] = None , _SCREAMING_SNAKE_CASE : List[Any]=False , _SCREAMING_SNAKE_CASE : str=None , _SCREAMING_SNAKE_CASE : Union[str, Any]=None , **_SCREAMING_SNAKE_CASE : Any , ) -> Tuple[PreTrainedModel, List[int], List[int]]: """simple docstring""" UpperCAmelCase_ : Optional[Any] = "encoder_layers and decoder_layers cannot be both None-- you would just have an identical teacher." assert (e is not None) or (d is not None), _msg if isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ): AutoTokenizer.from_pretrained(_SCREAMING_SNAKE_CASE ).save_pretrained(_SCREAMING_SNAKE_CASE ) # purely for convenience UpperCAmelCase_ : str = AutoModelForSeqaSeqLM.from_pretrained(_SCREAMING_SNAKE_CASE ).eval() else: assert isinstance(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ), F'''teacher must be a model or string got type {type(_SCREAMING_SNAKE_CASE )}''' UpperCAmelCase_ : Tuple = teacher.config.to_diff_dict() try: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = teacher.config.encoder_layers, teacher.config.decoder_layers if e is None: UpperCAmelCase_ : List[str] = teacher_e if d is None: UpperCAmelCase_ : Union[str, Any] = teacher_d init_kwargs.update({"encoder_layers": e, "decoder_layers": d} ) except AttributeError: # T5 if hasattr(teacher.config , "num_encoder_layers" ): UpperCAmelCase_ , UpperCAmelCase_ : int = teacher.config.num_encoder_layers, teacher.config.num_decoder_layers else: UpperCAmelCase_ , UpperCAmelCase_ : str = teacher.config.num_layers, teacher.config.num_decoder_layers if e is None: UpperCAmelCase_ : List[Any] = teacher_e if d is None: UpperCAmelCase_ : Optional[Any] = teacher_d if hasattr(teacher.config , "num_encoder_layers" ): init_kwargs.update({"num_encoder_layers": e, "num_decoder_layers": d} ) else: init_kwargs.update({"num_layers": e, "num_decoder_layers": d} ) # Kwargs to instantiate student: teacher kwargs with updated layer numbers + **extra_config_kwargs init_kwargs.update(_SCREAMING_SNAKE_CASE ) # Copy weights UpperCAmelCase_ : str = teacher.config_class(**_SCREAMING_SNAKE_CASE ) UpperCAmelCase_ : Optional[int] = AutoModelForSeqaSeqLM.from_config(_SCREAMING_SNAKE_CASE ) # Start by copying the full teacher state dict this will copy the first N teacher layers to the student. UpperCAmelCase_ : Tuple = student.load_state_dict(teacher.state_dict() , strict=_SCREAMING_SNAKE_CASE ) assert info.missing_keys == [], info.missing_keys # every student key should have a teacher keys. if copy_first_teacher_layers: # Our copying is done. We just log and save UpperCAmelCase_ , UpperCAmelCase_ : str = list(range(_SCREAMING_SNAKE_CASE ) ), list(range(_SCREAMING_SNAKE_CASE ) ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to''' F''' {save_path}''' ) student.save_pretrained(_SCREAMING_SNAKE_CASE ) return student, e_layers_to_copy, d_layers_to_copy # Decide which layers of the teacher to copy. Not exactly alternating -- we try to keep first and last layer. if e_layers_to_copy is None: UpperCAmelCase_ : List[int] = pick_layers_to_copy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) if d_layers_to_copy is None: UpperCAmelCase_ : List[int] = pick_layers_to_copy(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) try: if hasattr( _SCREAMING_SNAKE_CASE , "prophetnet" ): # For ProphetNet, student.model.encoder.layers is called student.prophetnet.encoder.layers copy_layers(teacher.prophetnet.encoder.layers , student.prophetnet.encoder.layers , _SCREAMING_SNAKE_CASE ) copy_layers(teacher.prophetnet.decoder.layers , student.prophetnet.decoder.layers , _SCREAMING_SNAKE_CASE ) else: copy_layers(teacher.model.encoder.layers , student.model.encoder.layers , _SCREAMING_SNAKE_CASE ) copy_layers(teacher.model.decoder.layers , student.model.decoder.layers , _SCREAMING_SNAKE_CASE ) except AttributeError: # For t5, student.model.encoder.layers is called student.encoder.block copy_layers(teacher.encoder.block , student.encoder.block , _SCREAMING_SNAKE_CASE ) copy_layers(teacher.decoder.block , student.decoder.block , _SCREAMING_SNAKE_CASE ) logger.info( F'''Copied encoder layers {e_layers_to_copy} and decoder layers {d_layers_to_copy}. Saving them to {save_path}''' ) UpperCAmelCase_ : Union[str, Any] = { "teacher_type": teacher.config.model_type, "copied_encoder_layers": e_layers_to_copy, "copied_decoder_layers": d_layers_to_copy, } student.save_pretrained(_SCREAMING_SNAKE_CASE ) # Save information about copying for easier reproducibility return student, e_layers_to_copy, d_layers_to_copy if __name__ == "__main__": fire.Fire(create_student_by_copying_alternating_layers)
67
'''simple docstring''' from __future__ import annotations from itertools import permutations from random import randint from timeit import repeat def a__ ( ) -> tuple[list[int], int]: """simple docstring""" UpperCAmelCase_ : Tuple = [randint(-10_00 , 10_00 ) for i in range(10 )] UpperCAmelCase_ : str = randint(-50_00 , 50_00 ) return (arr, r) _lowerCamelCase = make_dataset() def a__ ( _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : int ) -> tuple[int, ...]: """simple docstring""" for triplet in permutations(_SCREAMING_SNAKE_CASE , 3 ): if sum(_SCREAMING_SNAKE_CASE ) == target: return tuple(sorted(_SCREAMING_SNAKE_CASE ) ) return (0, 0, 0) def a__ ( _SCREAMING_SNAKE_CASE : list[int] , _SCREAMING_SNAKE_CASE : int ) -> tuple[int, int, int]: """simple docstring""" arr.sort() UpperCAmelCase_ : Optional[int] = len(_SCREAMING_SNAKE_CASE ) for i in range(n - 1 ): UpperCAmelCase_ , UpperCAmelCase_ : Union[str, Any] = i + 1, n - 1 while left < right: if arr[i] + arr[left] + arr[right] == target: return (arr[i], arr[left], arr[right]) elif arr[i] + arr[left] + arr[right] < target: left += 1 elif arr[i] + arr[left] + arr[right] > target: right -= 1 return (0, 0, 0) def a__ ( ) -> tuple[float, float]: """simple docstring""" UpperCAmelCase_ : Tuple = "\nfrom __main__ import dataset, triplet_sum1, triplet_sum2\n" UpperCAmelCase_ : Optional[Any] = "\ntriplet_sum1(*dataset)\n" UpperCAmelCase_ : str = "\ntriplet_sum2(*dataset)\n" UpperCAmelCase_ : Dict = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=1_00_00 ) UpperCAmelCase_ : Dict = repeat(setup=_SCREAMING_SNAKE_CASE , stmt=_SCREAMING_SNAKE_CASE , repeat=5 , number=1_00_00 ) return (min(_SCREAMING_SNAKE_CASE ), min(_SCREAMING_SNAKE_CASE )) if __name__ == "__main__": from doctest import testmod testmod() _lowerCamelCase = solution_times() print(f"""The time for naive implementation is {times[0]}.""") print(f"""The time for optimized implementation is {times[1]}.""")
67
1
'''simple docstring''' from collections import OrderedDict from typing import Any, Mapping, Optional, Union from ...configuration_utils import PretrainedConfig from ...feature_extraction_utils import FeatureExtractionMixin from ...onnx import OnnxConfig from ...onnx.utils import compute_effective_axis_dimension from ...tokenization_utils_base import PreTrainedTokenizerBase from ...utils import TensorType, logging lowercase : str = logging.get_logger(__name__) lowercase : int = { "deepmind/language-perceiver": "https://huggingface.co/deepmind/language-perceiver/resolve/main/config.json", # See all Perceiver models at https://huggingface.co/models?filter=perceiver } class __UpperCAmelCase ( _lowerCamelCase ): __lowercase = """perceiver""" def __init__( self , lowerCAmelCase_=2_56 , lowerCAmelCase_=12_80 , lowerCAmelCase_=7_68 , lowerCAmelCase_=1 , lowerCAmelCase_=26 , lowerCAmelCase_=8 , lowerCAmelCase_=8 , lowerCAmelCase_=None , lowerCAmelCase_=None , lowerCAmelCase_="kv" , lowerCAmelCase_=1 , lowerCAmelCase_=1 , lowerCAmelCase_="gelu" , lowerCAmelCase_=0.1 , lowerCAmelCase_=0.02 , lowerCAmelCase_=1E-12 , lowerCAmelCase_=True , lowerCAmelCase_=2_62 , lowerCAmelCase_=20_48 , lowerCAmelCase_=56 , lowerCAmelCase_=[3_68, 4_96] , lowerCAmelCase_=16 , lowerCAmelCase_=19_20 , lowerCAmelCase_=16 , lowerCAmelCase_=[1, 16, 2_24, 2_24] , **lowerCAmelCase_ , ): """simple docstring""" super().__init__(**lowerCAmelCase_ ) _snake_case = num_latents _snake_case = d_latents _snake_case = d_model _snake_case = num_blocks _snake_case = num_self_attends_per_block _snake_case = num_self_attention_heads _snake_case = num_cross_attention_heads _snake_case = qk_channels _snake_case = v_channels _snake_case = cross_attention_shape_for_attention _snake_case = self_attention_widening_factor _snake_case = cross_attention_widening_factor _snake_case = hidden_act _snake_case = attention_probs_dropout_prob _snake_case = initializer_range _snake_case = layer_norm_eps _snake_case = use_query_residual # masked language modeling attributes _snake_case = vocab_size _snake_case = max_position_embeddings # image classification attributes _snake_case = image_size # flow attributes _snake_case = train_size # multimodal autoencoding attributes _snake_case = num_frames _snake_case = audio_samples_per_frame _snake_case = samples_per_patch _snake_case = output_shape class __UpperCAmelCase ( _lowerCamelCase ): @property def lowerCamelCase ( self ): """simple docstring""" if self.task == "multiple-choice": _snake_case = {0: 'batch', 1: 'choice', 2: 'sequence'} else: _snake_case = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('inputs', dynamic_axis), ('attention_mask', dynamic_axis), ] ) @property def lowerCamelCase ( self ): """simple docstring""" return 1E-4 def lowerCamelCase ( self , lowerCAmelCase_ , lowerCAmelCase_ = -1 , lowerCAmelCase_ = -1 , lowerCAmelCase_ = -1 , lowerCAmelCase_ = False , lowerCAmelCase_ = None , lowerCAmelCase_ = 3 , lowerCAmelCase_ = 40 , lowerCAmelCase_ = 40 , ): """simple docstring""" if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ): # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _snake_case = compute_effective_axis_dimension( lowerCAmelCase_ , fixed_dimension=OnnxConfig.default_fixed_batch , num_token_to_add=0 ) # If dynamic axis (-1) we forward with a fixed dimension of 8 tokens to avoid optimizations made by ONNX _snake_case = preprocessor.num_special_tokens_to_add(lowerCAmelCase_ ) _snake_case = compute_effective_axis_dimension( lowerCAmelCase_ , fixed_dimension=OnnxConfig.default_fixed_sequence , num_token_to_add=lowerCAmelCase_ ) # Generate dummy inputs according to compute batch and sequence _snake_case = [' '.join(['a'] ) * seq_length] * batch_size _snake_case = dict(preprocessor(lowerCAmelCase_ , return_tensors=lowerCAmelCase_ ) ) _snake_case = inputs.pop('input_ids' ) return inputs elif isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and preprocessor.model_input_names[0] == "pixel_values": # If dynamic axis (-1) we forward with a fixed dimension of 2 samples to avoid optimizations made by ONNX _snake_case = compute_effective_axis_dimension(lowerCAmelCase_ , fixed_dimension=OnnxConfig.default_fixed_batch ) _snake_case = self._generate_dummy_images(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) _snake_case = dict(preprocessor(images=lowerCAmelCase_ , return_tensors=lowerCAmelCase_ ) ) _snake_case = inputs.pop('pixel_values' ) return inputs else: raise ValueError( 'Unable to generate dummy inputs for the model. Please provide a tokenizer or a preprocessor.' )
42
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments UpperCAmelCase = logging.getLogger(__name__) @dataclass class UpperCAmelCase_ ( _lowercase): snake_case__ = field( default=0.0 , metadata={'''help''': '''The label smoothing epsilon to apply (if not zero).'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''Whether to SortishSamler or not.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Whether to use generate to calculate generative metrics (ROUGE, BLEU).'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''whether to use adafactor'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Encoder layer dropout probability. Goes into model.config.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Decoder layer dropout probability. Goes into model.config.'''}) snake_case__ = field(default=_lowercase , metadata={'''help''': '''Dropout probability. Goes into model.config.'''}) snake_case__ = field( default=_lowercase , metadata={'''help''': '''Attention dropout probability. Goes into model.config.'''}) snake_case__ = field( default='''linear''' , metadata={'''help''': F'''Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys())}'''} , )
256
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import DebertaVaConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, TFDebertaVaModel, ) class snake_case : def __init__( self : Tuple , A : Union[str, Any] , A : str=1_3 , A : Dict=7 , A : Tuple=True , A : int=True , A : Dict=True , A : List[str]=True , A : str=9_9 , A : Optional[int]=3_2 , A : Optional[Any]=2 , A : Optional[Any]=4 , A : List[str]=3_7 , A : int="gelu" , A : str=0.1 , A : Union[str, Any]=0.1 , A : str=5_1_2 , A : str=1_6 , A : List[str]=2 , A : Tuple=0.02 , A : Any=False , A : List[Any]=True , A : int="None" , A : str=3 , A : Tuple=4 , A : str=None , ): '''simple docstring''' a : List[Any] = parent a : Any = batch_size a : List[str] = seq_length a : List[Any] = is_training a : Any = use_input_mask a : Union[str, Any] = use_token_type_ids a : Tuple = use_labels a : str = vocab_size a : str = hidden_size a : Union[str, Any] = num_hidden_layers a : Optional[Any] = num_attention_heads a : Tuple = intermediate_size a : Tuple = hidden_act a : Optional[int] = hidden_dropout_prob a : Dict = attention_probs_dropout_prob a : Optional[int] = max_position_embeddings a : Optional[int] = type_vocab_size a : int = type_sequence_label_size a : Any = initializer_range a : Dict = num_labels a : Optional[Any] = num_choices a : List[Any] = relative_attention a : Dict = position_biased_input a : Optional[Any] = pos_att_type a : Dict = scope def lowerCamelCase__ ( self : Union[str, Any] ): '''simple docstring''' a : int = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) a : Tuple = None if self.use_input_mask: a : str = random_attention_mask([self.batch_size, self.seq_length] ) a : str = None if self.use_token_type_ids: a : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) a : List[str] = None a : Union[str, Any] = None a : Union[str, Any] = None if self.use_labels: a : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) a : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) a : str = DebertaVaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , relative_attention=self.relative_attention , position_biased_input=self.position_biased_input , initializer_range=self.initializer_range , return_dict=A , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self : Optional[Any] , A : Any , A : str , A : Union[str, Any] , A : Dict , A : Optional[Any] , A : str , A : int ): '''simple docstring''' a : Dict = TFDebertaVaModel(config=A ) a : Optional[int] = {'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids} a : int = [input_ids, input_mask] a : Union[str, Any] = model(A ) a : Any = model(A ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self : List[Any] , A : Optional[int] , A : Optional[Any] , A : str , A : List[str] , A : str , A : str , A : Tuple ): '''simple docstring''' a : Any = TFDebertaVaForMaskedLM(config=A ) a : Any = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a : str = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self : Any , A : int , A : Optional[Any] , A : Tuple , A : Optional[int] , A : Union[str, Any] , A : Tuple , A : int ): '''simple docstring''' a : str = self.num_labels a : List[Any] = TFDebertaVaForSequenceClassification(config=A ) a : List[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a : str = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self : Dict , A : List[Any] , A : Optional[int] , A : List[Any] , A : Any , A : Tuple , A : Any , A : Union[str, Any] ): '''simple docstring''' a : List[Any] = self.num_labels a : List[str] = TFDebertaVaForTokenClassification(config=A ) a : Optional[Any] = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a : Optional[Any] = model(A ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self : Any , A : Dict , A : str , A : Tuple , A : int , A : Optional[int] , A : Any , A : List[str] ): '''simple docstring''' a : str = TFDebertaVaForQuestionAnswering(config=A ) a : str = { 'input_ids': input_ids, 'attention_mask': input_mask, 'token_type_ids': token_type_ids, } a : str = model(A ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self : List[Any] ): '''simple docstring''' a : int = self.prepare_config_and_inputs() ( a ) : List[Any] = config_and_inputs a : str = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_tf class snake_case ( UpperCAmelCase , UpperCAmelCase , unittest.TestCase ): __magic_name__ = ( ( TFDebertaVaModel, TFDebertaVaForMaskedLM, TFDebertaVaForQuestionAnswering, TFDebertaVaForSequenceClassification, TFDebertaVaForTokenClassification, ) if is_tf_available() else () ) __magic_name__ = ( { '''feature-extraction''': TFDebertaVaModel, '''fill-mask''': TFDebertaVaForMaskedLM, '''question-answering''': TFDebertaVaForQuestionAnswering, '''text-classification''': TFDebertaVaForSequenceClassification, '''token-classification''': TFDebertaVaForTokenClassification, '''zero-shot''': TFDebertaVaForSequenceClassification, } if is_tf_available() else {} ) __magic_name__ = False __magic_name__ = False def lowerCamelCase__ ( self : Dict ): '''simple docstring''' a : Tuple = TFDebertaVaModelTester(self ) a : Union[str, Any] = ConfigTester(self , config_class=A , hidden_size=3_7 ) def lowerCamelCase__ ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def lowerCamelCase__ ( self : Tuple ): '''simple docstring''' a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*A ) def lowerCamelCase__ ( self : str ): '''simple docstring''' a : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*A ) def lowerCamelCase__ ( self : Any ): '''simple docstring''' a : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*A ) def lowerCamelCase__ ( self : Any ): '''simple docstring''' a : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*A ) def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' a : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*A ) @slow def lowerCamelCase__ ( self : Dict ): '''simple docstring''' a : Any = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge' ) self.assertIsNotNone(A ) @require_tf class snake_case ( unittest.TestCase ): @unittest.skip(reason='Model not available yet' ) def lowerCamelCase__ ( self : str ): '''simple docstring''' pass @slow def lowerCamelCase__ ( self : List[str] ): '''simple docstring''' a : Any = TFDebertaVaModel.from_pretrained('kamalkraj/deberta-v2-xlarge' ) a : Union[str, Any] = tf.constant([[0, 3_1_4_1_4, 2_3_2, 3_2_8, 7_4_0, 1_1_4_0, 1_2_6_9_5, 6_9, 4_6_0_7_8, 1_5_8_8, 2]] ) a : List[Any] = tf.constant([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) a : Union[str, Any] = model(A , attention_mask=A )[0] a : List[Any] = tf.constant( [[[0.23_56, 0.19_48, 0.03_69], [-0.10_63, 0.35_86, -0.51_52], [-0.63_99, -0.02_59, -0.25_25]]] ) tf.debugging.assert_near(output[:, 1:4, 1:4] , A , atol=1E-4 )
354
"""simple docstring""" import logging from dataclasses import dataclass, field from pathlib import Path from typing import Optional, Union from .generation.configuration_utils import GenerationConfig from .training_args import TrainingArguments from .utils import add_start_docstrings _UpperCamelCase : Dict = logging.getLogger(__name__) @dataclass @add_start_docstrings(TrainingArguments.__doc__ ) class snake_case ( UpperCAmelCase ): __magic_name__ = field(default=UpperCAmelCase , metadata={'''help''': '''Whether to use SortishSampler or not.'''} ) __magic_name__ = field( default=UpperCAmelCase , metadata={'''help''': '''Whether to use generate to calculate generative metrics (ROUGE, BLEU).'''} ) __magic_name__ = field( default=UpperCAmelCase , metadata={ '''help''': ( '''The `max_length` to use on each evaluation loop when `predict_with_generate=True`. Will default ''' '''to the `max_length` value of the model configuration.''' ) } , ) __magic_name__ = field( default=UpperCAmelCase , metadata={ '''help''': ( '''The `num_beams` to use on each evaluation loop when `predict_with_generate=True`. Will default ''' '''to the `num_beams` value of the model configuration.''' ) } , ) __magic_name__ = field( default=UpperCAmelCase , metadata={ '''help''': '''Model id, file path or url pointing to a GenerationConfig json file, to use during prediction.''' } , ) def lowerCamelCase__ ( self : Optional[int] ): '''simple docstring''' a : List[Any] = super().to_dict() for k, v in d.items(): if isinstance(A , A ): a : str = v.to_dict() return d
186
0
import torch from diffusers import StableDiffusionPipeline lowerCamelCase_ = "path-to-your-trained-model" lowerCamelCase_ = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to('''cuda''') lowerCamelCase_ = "A photo of sks dog in a bucket" lowerCamelCase_ = pipe(prompt, num_inference_steps=50, guidance_scale=7.5).images[0] image.save('''dog-bucket.png''')
244
from __future__ import annotations import math def A_ ( snake_case : int ) -> bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True lowercase__ : int = [num for num in range(3, 1_0_0_0_0_1, 2) if not is_prime(num)] def A_ ( snake_case : int ) -> list[int]: '''simple docstring''' if not isinstance(snake_case , snake_case ): raise ValueError('''n must be an integer''' ) if n <= 0: raise ValueError('''n must be >= 0''' ) __UpperCamelCase = [] for num in range(len(snake_case ) ): __UpperCamelCase = 0 while 2 * i * i <= odd_composites[num]: __UpperCamelCase = odd_composites[num] - 2 * i * i if is_prime(snake_case ): break i += 1 else: list_nums.append(odd_composites[num] ) if len(snake_case ) == n: return list_nums return [] def A_ ( ) -> int: '''simple docstring''' return compute_nums(1 )[0] if __name__ == "__main__": print(F"{solution() = }")
328
0
"""simple docstring""" import functools import logging import os import sys import threading from logging import ( CRITICAL, # NOQA DEBUG, # NOQA ERROR, # NOQA FATAL, # NOQA INFO, # NOQA NOTSET, # NOQA WARN, # NOQA WARNING, # NOQA ) from typing import Optional import huggingface_hub.utils as hf_hub_utils from tqdm import auto as tqdm_lib a : int = threading.Lock() a : Optional[logging.Handler] = None a : List[str] = { """debug""": logging.DEBUG, """info""": logging.INFO, """warning""": logging.WARNING, """error""": logging.ERROR, """critical""": logging.CRITICAL, } a : Dict = logging.WARNING a : Tuple = True def lowercase__() ->str: """simple docstring""" lowercase__ : Any= os.getenv("TRANSFORMERS_VERBOSITY" , A ) if env_level_str: if env_level_str in log_levels: return log_levels[env_level_str] else: logging.getLogger().warning( f'''Unknown option TRANSFORMERS_VERBOSITY={env_level_str}, ''' f'''has to be one of: { ', '.join(log_levels.keys() ) }''' ) return _default_log_level def lowercase__() ->str: """simple docstring""" return __name__.split("." )[0] def lowercase__() ->logging.Logger: """simple docstring""" return logging.getLogger(_get_library_name() ) def lowercase__() ->None: """simple docstring""" global _default_handler with _lock: if _default_handler: # This library has already configured the library root logger. return lowercase__ : str= logging.StreamHandler() # Set sys.stderr as stream. lowercase__ : Optional[Any]= sys.stderr.flush # Apply our default configuration to the library root logger. lowercase__ : Dict= _get_library_root_logger() library_root_logger.addHandler(_default_handler ) library_root_logger.setLevel(_get_default_logging_level() ) lowercase__ : str= False def lowercase__() ->None: """simple docstring""" global _default_handler with _lock: if not _default_handler: return lowercase__ : List[str]= _get_library_root_logger() library_root_logger.removeHandler(_default_handler ) library_root_logger.setLevel(logging.NOTSET ) lowercase__ : Optional[Any]= None def lowercase__() ->Optional[int]: """simple docstring""" return log_levels def lowercase__(A = None ) ->logging.Logger: """simple docstring""" if name is None: lowercase__ : List[Any]= _get_library_name() _configure_library_root_logger() return logging.getLogger(A ) def lowercase__() ->int: """simple docstring""" _configure_library_root_logger() return _get_library_root_logger().getEffectiveLevel() def lowercase__(A ) ->None: """simple docstring""" _configure_library_root_logger() _get_library_root_logger().setLevel(A ) def lowercase__() ->Any: """simple docstring""" return set_verbosity(A ) def lowercase__() ->List[str]: """simple docstring""" return set_verbosity(A ) def lowercase__() ->Dict: """simple docstring""" return set_verbosity(A ) def lowercase__() ->Optional[int]: """simple docstring""" return set_verbosity(A ) def lowercase__() ->None: """simple docstring""" _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().removeHandler(_default_handler ) def lowercase__() ->None: """simple docstring""" _configure_library_root_logger() assert _default_handler is not None _get_library_root_logger().addHandler(_default_handler ) def lowercase__(A ) ->None: """simple docstring""" _configure_library_root_logger() assert handler is not None _get_library_root_logger().addHandler(A ) def lowercase__(A ) ->None: """simple docstring""" _configure_library_root_logger() assert handler is not None and handler not in _get_library_root_logger().handlers _get_library_root_logger().removeHandler(A ) def lowercase__() ->None: """simple docstring""" _configure_library_root_logger() lowercase__ : List[Any]= False def lowercase__() ->None: """simple docstring""" _configure_library_root_logger() lowercase__ : int= True def lowercase__() ->None: """simple docstring""" lowercase__ : Dict= _get_library_root_logger().handlers for handler in handlers: lowercase__ : str= logging.Formatter("[%(levelname)s|%(filename)s:%(lineno)s] %(asctime)s >> %(message)s" ) handler.setFormatter(A ) def lowercase__() ->None: """simple docstring""" lowercase__ : List[str]= _get_library_root_logger().handlers for handler in handlers: handler.setFormatter(A ) def lowercase__(self , *A , **A ) ->List[Any]: """simple docstring""" lowercase__ : Tuple= os.getenv("TRANSFORMERS_NO_ADVISORY_WARNINGS" , A ) if no_advisory_warnings: return self.warning(*A , **A ) a : Union[str, Any] = warning_advice @functools.lru_cache(A ) def lowercase__(self , *A , **A ) ->Any: """simple docstring""" self.warning(*A , **A ) a : Any = warning_once class __UpperCAmelCase: """simple docstring""" def __init__( self , *snake_case__ , **snake_case__ ): # pylint: disable=unused-argument '''simple docstring''' lowercase__ : Any= args[0] if args else None def __iter__( self ): '''simple docstring''' return iter(self._iterator ) def __getattr__( self , snake_case__ ): '''simple docstring''' def empty_fn(*snake_case__ , **snake_case__ ): # pylint: disable=unused-argument return return empty_fn def __enter__( self ): '''simple docstring''' return self def __exit__( self , snake_case__ , snake_case__ , snake_case__ ): '''simple docstring''' return class __UpperCAmelCase: """simple docstring""" def __call__( self , *snake_case__ , **snake_case__ ): '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm(*snake_case__ , **snake_case__ ) else: return EmptyTqdm(*snake_case__ , **snake_case__ ) def UpperCAmelCase_ ( self , *snake_case__ , **snake_case__ ): '''simple docstring''' lowercase__ : Dict= None if _tqdm_active: return tqdm_lib.tqdm.set_lock(*snake_case__ , **snake_case__ ) def UpperCAmelCase_ ( self ): '''simple docstring''' if _tqdm_active: return tqdm_lib.tqdm.get_lock() a : Tuple = _tqdm_cls() def lowercase__() ->bool: """simple docstring""" global _tqdm_active return bool(_tqdm_active ) def lowercase__() ->Optional[Any]: """simple docstring""" global _tqdm_active lowercase__ : List[str]= True hf_hub_utils.enable_progress_bars() def lowercase__() ->Union[str, Any]: """simple docstring""" global _tqdm_active lowercase__ : int= False hf_hub_utils.disable_progress_bars()
354
"""simple docstring""" import os from pathlib import Path def lowercase__() ->List[Any]: """simple docstring""" from torch.utils.cpp_extension import load lowercase__ : Any= Path(A ).resolve().parent.parent.parent / "kernels" / "deformable_detr" lowercase__ : Any= [ root / filename for filename in [ "vision.cpp", os.path.join("cpu" , "ms_deform_attn_cpu.cpp" ), os.path.join("cuda" , "ms_deform_attn_cuda.cu" ), ] ] load( "MultiScaleDeformableAttention" , A , with_cuda=A , extra_include_paths=[str(A )] , extra_cflags=["-DWITH_CUDA=1"] , extra_cuda_cflags=[ "-DCUDA_HAS_FP16=1", "-D__CUDA_NO_HALF_OPERATORS__", "-D__CUDA_NO_HALF_CONVERSIONS__", "-D__CUDA_NO_HALF2_OPERATORS__", ] , ) import MultiScaleDeformableAttention as MSDA return MSDA
150
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _A = logging.get_logger(__name__) _A = { '''facebook/dpr-ctx_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-reader-single-nq-base''': ( '''https://huggingface.co/facebook/dpr-reader-single-nq-base/resolve/main/config.json''' ), '''facebook/dpr-ctx_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-ctx_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-question_encoder-multiset-base''': ( '''https://huggingface.co/facebook/dpr-question_encoder-multiset-base/resolve/main/config.json''' ), '''facebook/dpr-reader-multiset-base''': ( '''https://huggingface.co/facebook/dpr-reader-multiset-base/resolve/main/config.json''' ), } class A ( __UpperCAmelCase ): __snake_case = 'dpr' def __init__( self, UpperCamelCase__=3_0522, UpperCamelCase__=768, UpperCamelCase__=12, UpperCamelCase__=12, UpperCamelCase__=3072, UpperCamelCase__="gelu", UpperCamelCase__=0.1, UpperCamelCase__=0.1, UpperCamelCase__=512, UpperCamelCase__=2, UpperCamelCase__=0.02, UpperCamelCase__=1E-12, UpperCamelCase__=0, UpperCamelCase__="absolute", UpperCamelCase__ = 0, **UpperCamelCase__, ): """simple docstring""" super().__init__(pad_token_id=UpperCamelCase__, **UpperCamelCase__ ) lowerCAmelCase_ = vocab_size lowerCAmelCase_ = hidden_size lowerCAmelCase_ = num_hidden_layers lowerCAmelCase_ = num_attention_heads lowerCAmelCase_ = hidden_act lowerCAmelCase_ = intermediate_size lowerCAmelCase_ = hidden_dropout_prob lowerCAmelCase_ = attention_probs_dropout_prob lowerCAmelCase_ = max_position_embeddings lowerCAmelCase_ = type_vocab_size lowerCAmelCase_ = initializer_range lowerCAmelCase_ = layer_norm_eps lowerCAmelCase_ = projection_dim lowerCAmelCase_ = position_embedding_type
278
import unittest from transformers import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device if is_torch_available(): import torch from transformers import AutoModelForImageClassification if is_vision_available(): from transformers import AutoImageProcessor @require_torch @require_vision class A ( unittest.TestCase ): @slow def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = AutoImageProcessor.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) lowerCAmelCase_ = AutoModelForImageClassification.from_pretrained('''microsoft/dit-base-finetuned-rvlcdip''' ) model.to(UpperCamelCase__ ) from datasets import load_dataset lowerCAmelCase_ = load_dataset('''nielsr/rvlcdip-demo''' ) lowerCAmelCase_ = dataset['''train'''][0]['''image'''].convert('''RGB''' ) lowerCAmelCase_ = image_processor(UpperCamelCase__, return_tensors='''pt''' ).to(UpperCamelCase__ ) # forward pass with torch.no_grad(): lowerCAmelCase_ = model(**UpperCamelCase__ ) lowerCAmelCase_ = outputs.logits lowerCAmelCase_ = torch.Size((1, 16) ) self.assertEqual(logits.shape, UpperCamelCase__ ) lowerCAmelCase_ = torch.tensor( [-0.4_158, -0.4_092, -0.4_347], device=UpperCamelCase__, dtype=torch.float, ) self.assertTrue(torch.allclose(logits[0, :3], UpperCamelCase__, atol=1E-4 ) )
278
1
def SCREAMING_SNAKE_CASE__ ( __a , __a ): snake_case_ : Optional[int] = int(__a ) # Initialize Result snake_case_ : Tuple = [] # Traverse through all denomination for denomination in reversed(__a ): # Find denominations while int(__a ) >= int(__a ): total_value -= int(__a ) answer.append(__a ) # Append the "answers" array return answer # Driver Code if __name__ == "__main__": _SCREAMING_SNAKE_CASE = [] _SCREAMING_SNAKE_CASE = """0""" if ( input("""Do you want to enter your denominations ? (yY/n): """).strip().lower() == "y" ): _SCREAMING_SNAKE_CASE = int(input("""Enter the number of denominations you want to add: """).strip()) for i in range(0, n): denominations.append(int(input(F'''Denomination {i}: ''').strip())) _SCREAMING_SNAKE_CASE = input("""Enter the change you want to make in Indian Currency: """).strip() else: # All denominations of Indian Currency if user does not enter _SCREAMING_SNAKE_CASE = [1, 2, 5, 10, 20, 50, 1_00, 5_00, 20_00] _SCREAMING_SNAKE_CASE = input("""Enter the change you want to make: """).strip() if int(value) == 0 or int(value) < 0: print("""The total value cannot be zero or negative.""") else: print(F'''Following is minimal change for {value}: ''') _SCREAMING_SNAKE_CASE = find_minimum_change(denominations, value) # Print result for i in range(len(answer)): print(answer[i], end=""" """)
366
_SCREAMING_SNAKE_CASE = { """Pillow""": """Pillow""", """accelerate""": """accelerate>=0.11.0""", """compel""": """compel==0.1.8""", """black""": """black~=23.1""", """datasets""": """datasets""", """filelock""": """filelock""", """flax""": """flax>=0.4.1""", """hf-doc-builder""": """hf-doc-builder>=0.3.0""", """huggingface-hub""": """huggingface-hub>=0.13.2""", """requests-mock""": """requests-mock==1.10.0""", """importlib_metadata""": """importlib_metadata""", """invisible-watermark""": """invisible-watermark""", """isort""": """isort>=5.5.4""", """jax""": """jax>=0.2.8,!=0.3.2""", """jaxlib""": """jaxlib>=0.1.65""", """Jinja2""": """Jinja2""", """k-diffusion""": """k-diffusion>=0.0.12""", """torchsde""": """torchsde""", """note_seq""": """note_seq""", """librosa""": """librosa""", """numpy""": """numpy""", """omegaconf""": """omegaconf""", """parameterized""": """parameterized""", """protobuf""": """protobuf>=3.20.3,<4""", """pytest""": """pytest""", """pytest-timeout""": """pytest-timeout""", """pytest-xdist""": """pytest-xdist""", """ruff""": """ruff>=0.0.241""", """safetensors""": """safetensors""", """sentencepiece""": """sentencepiece>=0.1.91,!=0.1.92""", """scipy""": """scipy""", """onnx""": """onnx""", """regex""": """regex!=2019.12.17""", """requests""": """requests""", """tensorboard""": """tensorboard""", """torch""": """torch>=1.4""", """torchvision""": """torchvision""", """transformers""": """transformers>=4.25.1""", """urllib3""": """urllib3<=2.0.0""", }
88
0
import argparse import torch from transformers import ( UniSpeechSatConfig, UniSpeechSatForAudioFrameClassification, UniSpeechSatForSequenceClassification, UniSpeechSatForXVector, WavaVecaFeatureExtractor, logging, ) logging.set_verbosity_info() A_ : int = logging.get_logger(__name__) def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Tuple: '''simple docstring''' __UpperCAmelCase = UniSpeechSatForSequenceClassification.from_pretrained(SCREAMING_SNAKE_CASE , config=SCREAMING_SNAKE_CASE ) __UpperCAmelCase = downstream_dict['''projector.weight'''] __UpperCAmelCase = downstream_dict['''projector.bias'''] __UpperCAmelCase = downstream_dict['''model.post_net.linear.weight'''] __UpperCAmelCase = downstream_dict['''model.post_net.linear.bias'''] return model def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> List[str]: '''simple docstring''' __UpperCAmelCase = UniSpeechSatForAudioFrameClassification.from_pretrained(SCREAMING_SNAKE_CASE , config=SCREAMING_SNAKE_CASE ) __UpperCAmelCase = downstream_dict['''model.linear.weight'''] __UpperCAmelCase = downstream_dict['''model.linear.bias'''] return model def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> int: '''simple docstring''' __UpperCAmelCase = UniSpeechSatForXVector.from_pretrained(SCREAMING_SNAKE_CASE , config=SCREAMING_SNAKE_CASE ) __UpperCAmelCase = downstream_dict['''connector.weight'''] __UpperCAmelCase = downstream_dict['''connector.bias'''] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): __UpperCAmelCase = downstream_dict[ f'''model.framelevel_feature_extractor.module.{i}.kernel.weight''' ] __UpperCAmelCase = downstream_dict[f'''model.framelevel_feature_extractor.module.{i}.kernel.bias'''] __UpperCAmelCase = downstream_dict['''model.utterancelevel_feature_extractor.linear1.weight'''] __UpperCAmelCase = downstream_dict['''model.utterancelevel_feature_extractor.linear1.bias'''] __UpperCAmelCase = downstream_dict['''model.utterancelevel_feature_extractor.linear2.weight'''] __UpperCAmelCase = downstream_dict['''model.utterancelevel_feature_extractor.linear2.bias'''] __UpperCAmelCase = downstream_dict['''objective.W'''] return model @torch.no_grad() def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> Optional[Any]: '''simple docstring''' __UpperCAmelCase = torch.load(SCREAMING_SNAKE_CASE , map_location='''cpu''' ) __UpperCAmelCase = checkpoint['''Downstream'''] __UpperCAmelCase = UniSpeechSatConfig.from_pretrained(SCREAMING_SNAKE_CASE ) __UpperCAmelCase = WavaVecaFeatureExtractor.from_pretrained( SCREAMING_SNAKE_CASE , return_attention_mask=SCREAMING_SNAKE_CASE , do_normalize=SCREAMING_SNAKE_CASE ) __UpperCAmelCase = hf_config.architectures[0] if arch.endswith('''ForSequenceClassification''' ): __UpperCAmelCase = convert_classification(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif arch.endswith('''ForAudioFrameClassification''' ): __UpperCAmelCase = convert_diarization(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) elif arch.endswith('''ForXVector''' ): __UpperCAmelCase = convert_xvector(SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) else: raise NotImplementedError(f'''S3PRL weights conversion is not supported for {arch}''' ) if hf_config.use_weighted_layer_sum: __UpperCAmelCase = checkpoint['''Featurizer''']['''weights'''] hf_feature_extractor.save_pretrained(SCREAMING_SNAKE_CASE ) hf_model.save_pretrained(SCREAMING_SNAKE_CASE ) if __name__ == "__main__": A_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument( '--base_model_name', default=None, type=str, help='Name of the huggingface pretrained base model.' ) parser.add_argument('--config_path', default=None, type=str, help='Path to the huggingface classifier config.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to the s3prl checkpoint.') parser.add_argument('--model_dump_path', default=None, type=str, help='Path to the final converted model.') A_ : int = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
333
import math def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> float: '''simple docstring''' if ( not isinstance(SCREAMING_SNAKE_CASE , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * power_factor def __a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ) -> float: '''simple docstring''' if ( not isinstance(SCREAMING_SNAKE_CASE , (int, float) ) or power_factor < -1 or power_factor > 1 ): raise ValueError('''power_factor must be a valid float value between -1 and 1.''' ) return apparent_power * math.sqrt(1 - power_factor**2 ) if __name__ == "__main__": import doctest doctest.testmod()
333
1
from __future__ import annotations def a_ ( __lowercase : Optional[Any] , __lowercase : Union[str, Any] , __lowercase : Union[str, Any] , __lowercase : Optional[int] ) -> str: # noqa: E741 while r - l > 1: _snake_case = (l + r) // 2 if v[m] >= key: _snake_case = m else: _snake_case = m # noqa: E741 return r def a_ ( __lowercase : list[int] ) -> int: if len(__lowercase ) == 0: return 0 _snake_case = [0] * len(__lowercase ) _snake_case = 1 _snake_case = v[0] for i in range(1 , len(__lowercase ) ): if v[i] < tail[0]: _snake_case = v[i] elif v[i] > tail[length - 1]: _snake_case = v[i] length += 1 else: _snake_case = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
365
import json import pathlib import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DeformableDetrImageProcessor class SCREAMING_SNAKE_CASE__ ( unittest.TestCase ): '''simple docstring''' def __init__( self : int , lowercase : Union[str, Any] , lowercase : str=7 , lowercase : Union[str, Any]=3 , lowercase : Tuple=30 , lowercase : Optional[Any]=400 , lowercase : List[Any]=True , lowercase : Any=None , lowercase : str=True , lowercase : Tuple=[0.5, 0.5, 0.5] , lowercase : List[Any]=[0.5, 0.5, 0.5] , lowercase : Union[str, Any]=True , lowercase : List[Any]=1 / 255 , lowercase : int=True , ): '''simple docstring''' _snake_case = size if size is not None else {'shortest_edge': 18, 'longest_edge': 1_333} _snake_case = parent _snake_case = batch_size _snake_case = num_channels _snake_case = min_resolution _snake_case = max_resolution _snake_case = do_resize _snake_case = size _snake_case = do_normalize _snake_case = image_mean _snake_case = image_std _snake_case = do_rescale _snake_case = rescale_factor _snake_case = do_pad def A ( self : str ): '''simple docstring''' return { "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, "image_mean": self.image_mean, "image_std": self.image_std, "do_rescale": self.do_rescale, "rescale_factor": self.rescale_factor, "do_pad": self.do_pad, } def A ( self : Optional[int] , lowercase : List[Any] , lowercase : Tuple=False ): '''simple docstring''' if not batched: _snake_case = image_inputs[0] if isinstance(lowercase , Image.Image ): _snake_case , _snake_case = image.size else: _snake_case , _snake_case = image.shape[1], image.shape[2] if w < h: _snake_case = int(self.size['shortest_edge'] * h / w ) _snake_case = self.size['shortest_edge'] elif w > h: _snake_case = self.size['shortest_edge'] _snake_case = int(self.size['shortest_edge'] * w / h ) else: _snake_case = self.size['shortest_edge'] _snake_case = self.size['shortest_edge'] else: _snake_case = [] for image in image_inputs: _snake_case , _snake_case = self.get_expected_values([image] ) expected_values.append((expected_height, expected_width) ) _snake_case = max(lowercase , key=lambda lowercase : item[0] )[0] _snake_case = max(lowercase , key=lambda lowercase : item[1] )[1] return expected_height, expected_width @require_torch @require_vision class SCREAMING_SNAKE_CASE__ ( UpperCAmelCase ,unittest.TestCase ): '''simple docstring''' _UpperCAmelCase : Dict = DeformableDetrImageProcessor if is_vision_available() else None def A ( self : List[Any] ): '''simple docstring''' _snake_case = DeformableDetrImageProcessingTester(self ) @property def A ( self : int ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def A ( self : Dict ): '''simple docstring''' _snake_case = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase , 'image_mean' ) ) self.assertTrue(hasattr(lowercase , 'image_std' ) ) self.assertTrue(hasattr(lowercase , 'do_normalize' ) ) self.assertTrue(hasattr(lowercase , 'do_resize' ) ) self.assertTrue(hasattr(lowercase , 'do_rescale' ) ) self.assertTrue(hasattr(lowercase , 'do_pad' ) ) self.assertTrue(hasattr(lowercase , 'size' ) ) def A ( self : Union[str, Any] ): '''simple docstring''' _snake_case = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {'shortest_edge': 18, 'longest_edge': 1_333} ) self.assertEqual(image_processor.do_pad , lowercase ) _snake_case = self.image_processing_class.from_dict( self.image_processor_dict , size=42 , max_size=84 , pad_and_return_pixel_mask=lowercase ) self.assertEqual(image_processor.size , {'shortest_edge': 42, 'longest_edge': 84} ) self.assertEqual(image_processor.do_pad , lowercase ) def A ( self : Dict ): '''simple docstring''' pass def A ( self : List[str] ): '''simple docstring''' _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , Image.Image ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) _snake_case = image_processing(lowercase , return_tensors='pt' ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : List[str] ): '''simple docstring''' _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , numpify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , np.ndarray ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _snake_case = image_processing(lowercase , return_tensors='pt' ).pixel_values _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case = prepare_image_inputs(self.image_processor_tester , equal_resolution=lowercase , torchify=lowercase ) for image in image_inputs: self.assertIsInstance(lowercase , torch.Tensor ) # Test not batched input _snake_case = image_processing(image_inputs[0] , return_tensors='pt' ).pixel_values _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase ) self.assertEqual( encoded_images.shape , (1, self.image_processor_tester.num_channels, expected_height, expected_width) , ) # Test batched _snake_case = image_processing(lowercase , return_tensors='pt' ).pixel_values _snake_case , _snake_case = self.image_processor_tester.get_expected_values(lowercase , batched=lowercase ) self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, expected_height, expected_width, ) , ) @slow def A ( self : List[str] ): '''simple docstring''' _snake_case = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_annotations.txt' , 'r' ) as f: _snake_case = json.loads(f.read() ) _snake_case = {'image_id': 39_769, 'annotations': target} # encode them _snake_case = DeformableDetrImageProcessor() _snake_case = image_processing(images=lowercase , annotations=lowercase , return_tensors='pt' ) # verify pixel values _snake_case = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['pixel_values'].shape , lowercase ) _snake_case = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowercase , atol=1E-4 ) ) # verify area _snake_case = torch.tensor([5887.9600, 11250.2061, 489353.8438, 837122.7500, 147967.5156, 165732.3438] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowercase ) ) # verify boxes _snake_case = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowercase ) _snake_case = torch.tensor([0.5503, 0.2765, 0.0604, 0.2215] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowercase , atol=1E-3 ) ) # verify image_id _snake_case = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowercase ) ) # verify is_crowd _snake_case = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowercase ) ) # verify class_labels _snake_case = torch.tensor([75, 75, 63, 65, 17, 17] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowercase ) ) # verify orig_size _snake_case = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowercase ) ) # verify size _snake_case = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowercase ) ) @slow def A ( self : Optional[Any] ): '''simple docstring''' _snake_case = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) with open('./tests/fixtures/tests_samples/COCO/coco_panoptic_annotations.txt' , 'r' ) as f: _snake_case = json.loads(f.read() ) _snake_case = {'file_name': '000000039769.png', 'image_id': 39_769, 'segments_info': target} _snake_case = pathlib.Path('./tests/fixtures/tests_samples/COCO/coco_panoptic' ) # encode them _snake_case = DeformableDetrImageProcessor(format='coco_panoptic' ) _snake_case = image_processing(images=lowercase , annotations=lowercase , masks_path=lowercase , return_tensors='pt' ) # verify pixel values _snake_case = torch.Size([1, 3, 800, 1_066] ) self.assertEqual(encoding['pixel_values'].shape , lowercase ) _snake_case = torch.tensor([0.2796, 0.3138, 0.3481] ) self.assertTrue(torch.allclose(encoding['pixel_values'][0, 0, 0, :3] , lowercase , atol=1E-4 ) ) # verify area _snake_case = torch.tensor([147979.6875, 165527.0469, 484638.5938, 11292.9375, 5879.6562, 7634.1147] ) self.assertTrue(torch.allclose(encoding['labels'][0]['area'] , lowercase ) ) # verify boxes _snake_case = torch.Size([6, 4] ) self.assertEqual(encoding['labels'][0]['boxes'].shape , lowercase ) _snake_case = torch.tensor([0.2625, 0.5437, 0.4688, 0.8625] ) self.assertTrue(torch.allclose(encoding['labels'][0]['boxes'][0] , lowercase , atol=1E-3 ) ) # verify image_id _snake_case = torch.tensor([39_769] ) self.assertTrue(torch.allclose(encoding['labels'][0]['image_id'] , lowercase ) ) # verify is_crowd _snake_case = torch.tensor([0, 0, 0, 0, 0, 0] ) self.assertTrue(torch.allclose(encoding['labels'][0]['iscrowd'] , lowercase ) ) # verify class_labels _snake_case = torch.tensor([17, 17, 63, 75, 75, 93] ) self.assertTrue(torch.allclose(encoding['labels'][0]['class_labels'] , lowercase ) ) # verify masks _snake_case = 822_873 self.assertEqual(encoding['labels'][0]['masks'].sum().item() , lowercase ) # verify orig_size _snake_case = torch.tensor([480, 640] ) self.assertTrue(torch.allclose(encoding['labels'][0]['orig_size'] , lowercase ) ) # verify size _snake_case = torch.tensor([800, 1_066] ) self.assertTrue(torch.allclose(encoding['labels'][0]['size'] , lowercase ) )
130
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCAmelCase_ : List[Any] = logging.get_logger(__name__) UpperCAmelCase_ : List[Any] = { """microsoft/markuplm-base""": """https://huggingface.co/microsoft/markuplm-base/resolve/main/config.json""", """microsoft/markuplm-large""": """https://huggingface.co/microsoft/markuplm-large/resolve/main/config.json""", } class lowerCAmelCase__ ( snake_case_ ): '''simple docstring''' __UpperCamelCase = "markuplm" def __init__( self : List[str] , lowercase_ : Optional[Any]=30522 , lowercase_ : str=768 , lowercase_ : Optional[int]=12 , lowercase_ : Dict=12 , lowercase_ : Any=3072 , lowercase_ : Dict="gelu" , lowercase_ : Any=0.1 , lowercase_ : str=0.1 , lowercase_ : str=512 , lowercase_ : Optional[int]=2 , lowercase_ : str=0.02 , lowercase_ : str=1e-12 , lowercase_ : str=0 , lowercase_ : int=0 , lowercase_ : Optional[int]=2 , lowercase_ : Optional[int]=256 , lowercase_ : Any=1024 , lowercase_ : Union[str, Any]=216 , lowercase_ : Any=1001 , lowercase_ : int=32 , lowercase_ : List[str]=50 , lowercase_ : str="absolute" , lowercase_ : str=True , lowercase_ : Optional[Any]=None , **lowercase_ : Optional[Any] , ): '''simple docstring''' super().__init__( pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase , ) SCREAMING_SNAKE_CASE_ : str = vocab_size SCREAMING_SNAKE_CASE_ : int = hidden_size SCREAMING_SNAKE_CASE_ : Any = num_hidden_layers SCREAMING_SNAKE_CASE_ : Dict = num_attention_heads SCREAMING_SNAKE_CASE_ : Optional[Any] = hidden_act SCREAMING_SNAKE_CASE_ : Dict = intermediate_size SCREAMING_SNAKE_CASE_ : List[Any] = hidden_dropout_prob SCREAMING_SNAKE_CASE_ : Union[str, Any] = attention_probs_dropout_prob SCREAMING_SNAKE_CASE_ : List[Any] = max_position_embeddings SCREAMING_SNAKE_CASE_ : Any = type_vocab_size SCREAMING_SNAKE_CASE_ : List[Any] = initializer_range SCREAMING_SNAKE_CASE_ : Union[str, Any] = layer_norm_eps SCREAMING_SNAKE_CASE_ : List[str] = position_embedding_type SCREAMING_SNAKE_CASE_ : Union[str, Any] = use_cache SCREAMING_SNAKE_CASE_ : List[str] = classifier_dropout # additional properties SCREAMING_SNAKE_CASE_ : Union[str, Any] = max_depth SCREAMING_SNAKE_CASE_ : int = max_xpath_tag_unit_embeddings SCREAMING_SNAKE_CASE_ : List[str] = max_xpath_subs_unit_embeddings SCREAMING_SNAKE_CASE_ : Dict = tag_pad_id SCREAMING_SNAKE_CASE_ : Any = subs_pad_id SCREAMING_SNAKE_CASE_ : Optional[int] = xpath_unit_hidden_size
91
'''simple docstring''' import argparse import logging import os from datetime import datetime import numpy as np import torch from torch import nn from torch.utils.data import DataLoader, RandomSampler, TensorDataset from tqdm import tqdm from transformers import GPTaLMHeadModel lowerCamelCase_ = logging.getLogger(__name__) def __lowercase ( __lowercase , __lowercase ) -> Optional[int]: '''simple docstring''' if os.path.exists(__lowercase ): if os.path.exists(os.path.join(__lowercase , "config.json" ) ) and os.path.isfile( os.path.join(__lowercase , "config.json" ) ): os.remove(os.path.join(__lowercase , "config.json" ) ) if os.path.exists(os.path.join(__lowercase , "pytorch_model.bin" ) ) and os.path.isfile( os.path.join(__lowercase , "pytorch_model.bin" ) ): os.remove(os.path.join(__lowercase , "pytorch_model.bin" ) ) else: os.makedirs(__lowercase ) model.save_pretrained(__lowercase ) def __lowercase ( __lowercase , __lowercase=False ) -> Optional[int]: '''simple docstring''' _A = 2 if unlogit: _A = torch.pow(__lowercase , __lowercase ) _A = p * torch.log(__lowercase ) _A = 0 return -plogp.sum(dim=-1 ) def __lowercase ( __lowercase ) -> Optional[Any]: '''simple docstring''' logger.info("lv, h >\t" + "\t".join(F'''{x + 1}''' for x in range(len(__lowercase ) ) ) ) for row in range(len(__lowercase ) ): if tensor.dtype != torch.long: logger.info(F'''layer {row + 1}:\t''' + "\t".join(F'''{x:.5f}''' for x in tensor[row].cpu().data ) ) else: logger.info(F'''layer {row + 1}:\t''' + "\t".join(F'''{x:d}''' for x in tensor[row].cpu().data ) ) def __lowercase ( __lowercase , __lowercase , __lowercase , __lowercase=True , __lowercase=True , __lowercase=None , __lowercase=False ) -> int: '''simple docstring''' _A , _A = model.config.num_hidden_layers, model.config.num_attention_heads _A = torch.zeros(__lowercase , __lowercase ).to(args.device ) _A = torch.zeros(__lowercase , __lowercase ).to(args.device ) if head_mask is None: _A = torch.ones(__lowercase , __lowercase ).to(args.device ) head_mask.requires_grad_(requires_grad=__lowercase ) # If actually pruned attention multi-head, set head mask to None to avoid shape mismatch if actually_pruned: _A = None _A = 0.0 _A = 0.0 for step, inputs in enumerate(tqdm(__lowercase , desc="Iteration" , disable=args.local_rank not in [-1, 0] ) ): _A = tuple(t.to(args.device ) for t in inputs ) ((_A) , ) = inputs # Do a forward pass (not with torch.no_grad() since we need gradients for importance score - see below) _A = model(__lowercase , labels=__lowercase , head_mask=__lowercase ) # (loss), lm_logits, presents, (all hidden_states), (attentions) _A , _A , _A = ( outputs[0], outputs[1], outputs[-1], ) # Loss and logits are the first, attention the last loss.backward() # Backpropagate to populate the gradients in the head mask total_loss += loss.detach().cpu().numpy() if compute_entropy: for layer, attn in enumerate(__lowercase ): _A = entropy(attn.detach() , __lowercase ) attn_entropy[layer] += masked_entropy.sum(-1 ).sum(0 ).sum(0 ).detach() if compute_importance: head_importance += head_mask.grad.abs().detach() tot_tokens += torch.ones_like(__lowercase ).float().detach().sum().data # Normalize attn_entropy /= tot_tokens head_importance /= tot_tokens # Layerwise importance normalization if not args.dont_normalize_importance_by_layer: _A = 2 _A = torch.pow(torch.pow(__lowercase , __lowercase ).sum(-1 ) , 1 / exponent ) head_importance /= norm_by_layer.unsqueeze(-1 ) + 1e-20 if not args.dont_normalize_global_importance: _A = (head_importance - head_importance.min()) / (head_importance.max() - head_importance.min()) # Print matrices if compute_entropy: logger.info("Attention entropies" ) print_ad_tensor(__lowercase ) if compute_importance: logger.info("Head importance scores" ) print_ad_tensor(__lowercase ) logger.info("Head ranked by importance scores" ) _A = torch.zeros(head_importance.numel() , dtype=torch.long , device=args.device ) _A = torch.arange( head_importance.numel() , device=args.device ) _A = head_ranks.view_as(__lowercase ) print_ad_tensor(__lowercase ) return attn_entropy, head_importance, total_loss def __lowercase ( __lowercase , __lowercase , __lowercase ) -> List[str]: '''simple docstring''' _A , _A , _A = compute_heads_importance(__lowercase , __lowercase , __lowercase , compute_entropy=__lowercase ) _A = 1 / loss # instead of downsteam score use the LM loss logger.info("Pruning: original score: %f, threshold: %f" , __lowercase , original_score * args.masking_threshold ) _A = torch.ones_like(__lowercase ) _A = max(1 , int(new_head_mask.numel() * args.masking_amount ) ) _A = original_score while current_score >= original_score * args.masking_threshold: _A = new_head_mask.clone().detach() # save current head mask # heads from least important to most - keep only not-masked heads _A = float("Inf" ) _A = head_importance.view(-1 ).sort()[1] if len(__lowercase ) <= num_to_mask: print("BREAK BY num_to_mask" ) break # mask heads _A = current_heads_to_mask[:num_to_mask] logger.info("Heads to mask: %s" , str(current_heads_to_mask.tolist() ) ) _A = new_head_mask.view(-1 ) _A = 0.0 _A = new_head_mask.view_as(__lowercase ) _A = new_head_mask.clone().detach() print_ad_tensor(__lowercase ) # Compute metric and head importance again _A , _A , _A = compute_heads_importance( __lowercase , __lowercase , __lowercase , compute_entropy=__lowercase , head_mask=__lowercase ) _A = 1 / loss logger.info( "Masking: current score: %f, remaining heads %d (%.1f percents)" , __lowercase , new_head_mask.sum() , new_head_mask.sum() / new_head_mask.numel() * 100 , ) logger.info("Final head mask" ) print_ad_tensor(__lowercase ) np.save(os.path.join(args.output_dir , "head_mask.npy" ) , head_mask.detach().cpu().numpy() ) return head_mask def __lowercase ( __lowercase , __lowercase , __lowercase , __lowercase ) -> List[str]: '''simple docstring''' _A = datetime.now() _A , _A , _A = compute_heads_importance( __lowercase , __lowercase , __lowercase , compute_entropy=__lowercase , compute_importance=__lowercase , head_mask=__lowercase ) _A = 1 / loss _A = datetime.now() - before_time _A = sum(p.numel() for p in model.parameters() ) _A = { layer: (1 - head_mask[layer].long()).nonzero().squeeze().tolist() for layer in range(len(__lowercase ) ) } for k, v in heads_to_prune.items(): if isinstance(__lowercase , __lowercase ): _A = [ v, ] assert sum(len(__lowercase ) for h in heads_to_prune.values() ) == (1 - head_mask.long()).sum().item() model.prune_heads(__lowercase ) _A = sum(p.numel() for p in model.parameters() ) _A = datetime.now() _A , _A , _A = compute_heads_importance( __lowercase , __lowercase , __lowercase , compute_entropy=__lowercase , compute_importance=__lowercase , head_mask=__lowercase , actually_pruned=__lowercase , ) _A = 1 / loss _A = datetime.now() - before_time logger.info( "Pruning: original num of params: %.2e, after pruning %.2e (%.1f percents)" , __lowercase , __lowercase , pruned_num_params / original_num_params * 100 , ) logger.info("Pruning: score with masking: %f score with pruning: %f" , __lowercase , __lowercase ) logger.info("Pruning: speed ratio (original timing / new timing): %f percents" , original_time / new_time * 100 ) save_model(__lowercase , args.output_dir ) def __lowercase ( ) -> Union[str, Any]: '''simple docstring''' _A = argparse.ArgumentParser() # Required parameters parser.add_argument( "--data_dir" , default=__lowercase , type=__lowercase , required=__lowercase , help="The input data dir. Should contain the .tsv files (or other data files) for the task." , ) parser.add_argument( "--model_name_or_path" , default=__lowercase , type=__lowercase , required=__lowercase , help="Path to pretrained model or model identifier from huggingface.co/models" , ) parser.add_argument( "--output_dir" , default=__lowercase , type=__lowercase , required=__lowercase , help="The output directory where the model predictions and checkpoints will be written." , ) # Other parameters parser.add_argument( "--config_name" , default="" , type=__lowercase , help="Pretrained config name or path if not the same as model_name_or_path" , ) parser.add_argument( "--tokenizer_name" , default="" , type=__lowercase , help="Pretrained tokenizer name or path if not the same as model_name_or_path" , ) parser.add_argument( "--cache_dir" , default=__lowercase , type=__lowercase , help="Where do you want to store the pre-trained models downloaded from s3" , ) parser.add_argument( "--data_subset" , type=__lowercase , default=-1 , help="If > 0: limit the data to a subset of data_subset instances." ) parser.add_argument( "--overwrite_output_dir" , action="store_true" , help="Whether to overwrite data in output directory" ) parser.add_argument( "--overwrite_cache" , action="store_true" , help="Overwrite the cached training and evaluation sets" ) parser.add_argument( "--dont_normalize_importance_by_layer" , action="store_true" , help="Don't normalize importance score by layers" ) parser.add_argument( "--dont_normalize_global_importance" , action="store_true" , help="Don't normalize all importance scores between 0 and 1" , ) parser.add_argument( "--try_masking" , action="store_true" , help="Whether to try to mask head until a threshold of accuracy." ) parser.add_argument( "--masking_threshold" , default=0.9 , type=__lowercase , help="masking threshold in term of metrics (stop masking when metric < threshold * original metric value)." , ) parser.add_argument( "--masking_amount" , default=0.1 , type=__lowercase , help="Amount to heads to masking at each masking step." ) parser.add_argument("--metric_name" , default="acc" , type=__lowercase , help="Metric to use for head masking." ) parser.add_argument( "--max_seq_length" , default=128 , type=__lowercase , help=( "The maximum total input sequence length after WordPiece tokenization. \n" "Sequences longer than this will be truncated, sequences shorter padded." ) , ) parser.add_argument("--batch_size" , default=1 , type=__lowercase , help="Batch size." ) parser.add_argument("--seed" , type=__lowercase , default=42 ) parser.add_argument("--local_rank" , type=__lowercase , default=-1 , help="local_rank for distributed training on gpus" ) parser.add_argument("--no_cuda" , action="store_true" , help="Whether not to use CUDA when available" ) parser.add_argument("--server_ip" , type=__lowercase , default="" , help="Can be used for distant debugging." ) parser.add_argument("--server_port" , type=__lowercase , default="" , help="Can be used for distant debugging." ) _A = parser.parse_args() if args.server_ip and args.server_port: # Distant debugging - see https://code.visualstudio.com/docs/python/debugging#_attach-to-a-local-script import ptvsd print("Waiting for debugger attach" ) ptvsd.enable_attach(address=(args.server_ip, args.server_port) , redirect_output=__lowercase ) ptvsd.wait_for_attach() # Setup devices and distributed training if args.local_rank == -1 or args.no_cuda: _A = torch.device("cuda" if torch.cuda.is_available() and not args.no_cuda else "cpu" ) _A = 0 if args.no_cuda else torch.cuda.device_count() else: torch.cuda.set_device(args.local_rank ) _A = torch.device("cuda" , args.local_rank ) _A = 1 torch.distributed.init_process_group(backend="nccl" ) # Initializes the distributed backend # Setup logging logging.basicConfig(level=logging.INFO if args.local_rank in [-1, 0] else logging.WARN ) logger.info("device: {} n_gpu: {}, distributed: {}".format(args.device , args.n_gpu , bool(args.local_rank != -1 ) ) ) _A = GPTaLMHeadModel.from_pretrained(args.model_name_or_path ) # Distributed and parallel training model.to(args.device ) if args.local_rank != -1: _A = nn.parallel.DistributedDataParallel( __lowercase , device_ids=[args.local_rank] , output_device=args.local_rank , find_unused_parameters=__lowercase ) elif args.n_gpu > 1: _A = nn.DataParallel(__lowercase ) # Print/save training arguments os.makedirs(args.output_dir , exist_ok=__lowercase ) torch.save(__lowercase , os.path.join(args.output_dir , "run_args.bin" ) ) logger.info("Training/evaluation parameters %s" , __lowercase ) # Prepare dataset _A = np.concatenate( [ np.loadtxt(args.data_dir , dtype=np.intaa ), ] ) _A = (torch.from_numpy(__lowercase ),) _A = TensorDataset(*__lowercase ) _A = RandomSampler(__lowercase ) _A = DataLoader(__lowercase , sampler=__lowercase , batch_size=args.batch_size ) # Compute head entropy and importance score compute_heads_importance(__lowercase , __lowercase , __lowercase ) # Try head masking (set heads to zero until the score goes under a threshole) # and head pruning (remove masked heads and see the effect on the network) if args.try_masking and args.masking_threshold > 0.0 and args.masking_threshold < 1.0: _A = mask_heads(__lowercase , __lowercase , __lowercase ) prune_heads(__lowercase , __lowercase , __lowercase , __lowercase ) if __name__ == "__main__": main()
79
0
from __future__ import annotations from typing import Any class lowercase : def __init__( self ,A__): lowercase = num_of_nodes lowercase = [] lowercase = {} def A__ ( self ,A__ ,A__ ,A__): self.m_edges.append([u_node, v_node, weight]) def A__ ( self ,A__): if self.m_component[u_node] == u_node: return u_node return self.find_component(self.m_component[u_node]) def A__ ( self ,A__): if self.m_component[u_node] != u_node: for k in self.m_component: lowercase = self.find_component(A__) def A__ ( self ,A__ ,A__ ,A__): if component_size[u_node] <= component_size[v_node]: lowercase = v_node component_size[v_node] += component_size[u_node] self.set_component(A__) elif component_size[u_node] >= component_size[v_node]: lowercase = self.find_component(A__) component_size[u_node] += component_size[v_node] self.set_component(A__) def A__ ( self): lowercase = [] lowercase = 0 lowercase = [-1] * self.m_num_of_nodes # A list of components (initialized to all of the nodes) for node in range(self.m_num_of_nodes): self.m_component.update({node: node}) component_size.append(1) lowercase = self.m_num_of_nodes while num_of_components > 1: for edge in self.m_edges: lowercase , lowercase , lowercase = edge lowercase = self.m_component[u] lowercase = self.m_component[v] if u_component != v_component: for component in (u_component, v_component): if ( minimum_weight_edge[component] == -1 or minimum_weight_edge[component][2] > w ): lowercase = [u, v, w] for edge in minimum_weight_edge: if isinstance(A__ ,A__): lowercase , lowercase , lowercase = edge lowercase = self.m_component[u] lowercase = self.m_component[v] if u_component != v_component: mst_weight += w self.union(A__ ,A__ ,A__) print(f'Added edge [{u} - {v}]\nAdded weight: {w}\n') num_of_components -= 1 lowercase = [-1] * self.m_num_of_nodes print(f'The total weight of the minimal spanning tree is: {mst_weight}') def UpperCamelCase ( ): '''simple docstring''' if __name__ == "__main__": import doctest doctest.testmod()
367
from __future__ import annotations from decimal import Decimal from numpy import array def UpperCamelCase ( lowerCAmelCase__ ): '''simple docstring''' lowercase = Decimal # Check if the provided matrix has 2 rows and 2 columns # since this implementation only works for 2x2 matrices if len(lowerCAmelCase__ ) == 2 and len(matrix[0] ) == 2 and len(matrix[1] ) == 2: # Calculate the determinant of the matrix lowercase = float( d(matrix[0][0] ) * d(matrix[1][1] ) - d(matrix[1][0] ) * d(matrix[0][1] ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creates a copy of the matrix with swapped positions of the elements lowercase = [[0.0, 0.0], [0.0, 0.0]] lowercase , lowercase = matrix[1][1], matrix[0][0] lowercase , lowercase = -matrix[1][0], -matrix[0][1] # Calculate the inverse of the matrix return [ [(float(d(lowerCAmelCase__ ) ) / determinant) or 0.0 for n in row] for row in swapped_matrix ] elif ( len(lowerCAmelCase__ ) == 3 and len(matrix[0] ) == 3 and len(matrix[1] ) == 3 and len(matrix[2] ) == 3 ): # Calculate the determinant of the matrix using Sarrus rule lowercase = float( ( (d(matrix[0][0] ) * d(matrix[1][1] ) * d(matrix[2][2] )) + (d(matrix[0][1] ) * d(matrix[1][2] ) * d(matrix[2][0] )) + (d(matrix[0][2] ) * d(matrix[1][0] ) * d(matrix[2][1] )) ) - ( (d(matrix[0][2] ) * d(matrix[1][1] ) * d(matrix[2][0] )) + (d(matrix[0][1] ) * d(matrix[1][0] ) * d(matrix[2][2] )) + (d(matrix[0][0] ) * d(matrix[1][2] ) * d(matrix[2][1] )) ) ) if determinant == 0: raise ValueError('''This matrix has no inverse.''' ) # Creating cofactor matrix lowercase = [ [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], [d(0.0 ), d(0.0 ), d(0.0 )], ] lowercase = (d(matrix[1][1] ) * d(matrix[2][2] )) - ( d(matrix[1][2] ) * d(matrix[2][1] ) ) lowercase = -( (d(matrix[1][0] ) * d(matrix[2][2] )) - (d(matrix[1][2] ) * d(matrix[2][0] )) ) lowercase = (d(matrix[1][0] ) * d(matrix[2][1] )) - ( d(matrix[1][1] ) * d(matrix[2][0] ) ) lowercase = -( (d(matrix[0][1] ) * d(matrix[2][2] )) - (d(matrix[0][2] ) * d(matrix[2][1] )) ) lowercase = (d(matrix[0][0] ) * d(matrix[2][2] )) - ( d(matrix[0][2] ) * d(matrix[2][0] ) ) lowercase = -( (d(matrix[0][0] ) * d(matrix[2][1] )) - (d(matrix[0][1] ) * d(matrix[2][0] )) ) lowercase = (d(matrix[0][1] ) * d(matrix[1][2] )) - ( d(matrix[0][2] ) * d(matrix[1][1] ) ) lowercase = -( (d(matrix[0][0] ) * d(matrix[1][2] )) - (d(matrix[0][2] ) * d(matrix[1][0] )) ) lowercase = (d(matrix[0][0] ) * d(matrix[1][1] )) - ( d(matrix[0][1] ) * d(matrix[1][0] ) ) # Transpose the cofactor matrix (Adjoint matrix) lowercase = array(lowerCAmelCase__ ) for i in range(3 ): for j in range(3 ): lowercase = cofactor_matrix[j][i] # Inverse of the matrix using the formula (1/determinant) * adjoint matrix lowercase = array(lowerCAmelCase__ ) for i in range(3 ): for j in range(3 ): inverse_matrix[i][j] /= d(lowerCAmelCase__ ) # Calculate the inverse of the matrix return [[float(d(lowerCAmelCase__ ) ) or 0.0 for n in row] for row in inverse_matrix] raise ValueError('''Please provide a matrix of size 2x2 or 3x3.''' )
97
0
'''simple docstring''' from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase_ = logging.get_logger(__name__) UpperCamelCase_ = { "xlm-mlm-en-2048": "https://huggingface.co/xlm-mlm-en-2048/resolve/main/config.json", "xlm-mlm-ende-1024": "https://huggingface.co/xlm-mlm-ende-1024/resolve/main/config.json", "xlm-mlm-enfr-1024": "https://huggingface.co/xlm-mlm-enfr-1024/resolve/main/config.json", "xlm-mlm-enro-1024": "https://huggingface.co/xlm-mlm-enro-1024/resolve/main/config.json", "xlm-mlm-tlm-xnli15-1024": "https://huggingface.co/xlm-mlm-tlm-xnli15-1024/resolve/main/config.json", "xlm-mlm-xnli15-1024": "https://huggingface.co/xlm-mlm-xnli15-1024/resolve/main/config.json", "xlm-clm-enfr-1024": "https://huggingface.co/xlm-clm-enfr-1024/resolve/main/config.json", "xlm-clm-ende-1024": "https://huggingface.co/xlm-clm-ende-1024/resolve/main/config.json", "xlm-mlm-17-1280": "https://huggingface.co/xlm-mlm-17-1280/resolve/main/config.json", "xlm-mlm-100-1280": "https://huggingface.co/xlm-mlm-100-1280/resolve/main/config.json", } class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' A : Union[str, Any] = '''xlm''' A : Dict = { '''hidden_size''': '''emb_dim''', '''num_attention_heads''': '''n_heads''', '''num_hidden_layers''': '''n_layers''', '''n_words''': '''vocab_size''', # For backward compatibility } def __init__( self, A=30_145, A=2_048, A=12, A=16, A=0.1, A=0.1, A=True, A=False, A=False, A=False, A=1, A=True, A=512, A=2_048**-0.5, A=1E-12, A=0.02, A=0, A=1, A=2, A=3, A=5, A=True, A="first", A=True, A=None, A=True, A=0.1, A=5, A=5, A=0, A=0, A=2, A=0, **A, ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = vocab_size SCREAMING_SNAKE_CASE : int = emb_dim SCREAMING_SNAKE_CASE : Dict = n_layers SCREAMING_SNAKE_CASE : Tuple = n_heads SCREAMING_SNAKE_CASE : List[str] = dropout SCREAMING_SNAKE_CASE : Optional[Any] = attention_dropout SCREAMING_SNAKE_CASE : Dict = gelu_activation SCREAMING_SNAKE_CASE : Optional[int] = sinusoidal_embeddings SCREAMING_SNAKE_CASE : str = causal SCREAMING_SNAKE_CASE : Tuple = asm SCREAMING_SNAKE_CASE : List[str] = n_langs SCREAMING_SNAKE_CASE : Any = use_lang_emb SCREAMING_SNAKE_CASE : List[Any] = layer_norm_eps SCREAMING_SNAKE_CASE : Union[str, Any] = bos_index SCREAMING_SNAKE_CASE : Any = eos_index SCREAMING_SNAKE_CASE : int = pad_index SCREAMING_SNAKE_CASE : List[str] = unk_index SCREAMING_SNAKE_CASE : Tuple = mask_index SCREAMING_SNAKE_CASE : Tuple = is_encoder SCREAMING_SNAKE_CASE : int = max_position_embeddings SCREAMING_SNAKE_CASE : Dict = embed_init_std SCREAMING_SNAKE_CASE : Union[str, Any] = init_std SCREAMING_SNAKE_CASE : Optional[Any] = summary_type SCREAMING_SNAKE_CASE : Optional[int] = summary_use_proj SCREAMING_SNAKE_CASE : Optional[int] = summary_activation SCREAMING_SNAKE_CASE : Optional[Any] = summary_proj_to_labels SCREAMING_SNAKE_CASE : Any = summary_first_dropout SCREAMING_SNAKE_CASE : Optional[Any] = start_n_top SCREAMING_SNAKE_CASE : List[Any] = end_n_top SCREAMING_SNAKE_CASE : Any = mask_token_id SCREAMING_SNAKE_CASE : Tuple = lang_id if "n_words" in kwargs: SCREAMING_SNAKE_CASE : List[str] = kwargs['n_words'] super().__init__(pad_token_id=A, bos_token_id=A, **A ) class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' @property def UpperCamelCase_ ( self ): '''simple docstring''' if self.task == "multiple-choice": SCREAMING_SNAKE_CASE : Optional[Any] = {0: 'batch', 1: 'choice', 2: 'sequence'} else: SCREAMING_SNAKE_CASE : Tuple = {0: 'batch', 1: 'sequence'} return OrderedDict( [ ('input_ids', dynamic_axis), ('attention_mask', dynamic_axis), ('token_type_ids', dynamic_axis), ] )
251
'''simple docstring''' class _a : '''simple docstring''' def __init__( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = 0 SCREAMING_SNAKE_CASE : str = 0 SCREAMING_SNAKE_CASE : Optional[int] = {} def UpperCamelCase_ ( self, A ): '''simple docstring''' if vertex not in self.adjacency: SCREAMING_SNAKE_CASE : List[str] = {} self.num_vertices += 1 def UpperCamelCase_ ( self, A, A, A ): '''simple docstring''' self.add_vertex(A ) self.add_vertex(A ) if head == tail: return SCREAMING_SNAKE_CASE : List[Any] = weight SCREAMING_SNAKE_CASE : Optional[Any] = weight def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[Any] = self.get_edges() for edge in edges: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = edge edges.remove((tail, head, weight) ) for i in range(len(A ) ): SCREAMING_SNAKE_CASE : Optional[Any] = list(edges[i] ) edges.sort(key=lambda A : e[2] ) for i in range(len(A ) - 1 ): if edges[i][2] >= edges[i + 1][2]: SCREAMING_SNAKE_CASE : Dict = edges[i][2] + 1 for edge in edges: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[Any] = edge SCREAMING_SNAKE_CASE : Union[str, Any] = weight SCREAMING_SNAKE_CASE : Optional[Any] = weight def __str__( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = '' for tail in self.adjacency: for head in self.adjacency[tail]: SCREAMING_SNAKE_CASE : Union[str, Any] = self.adjacency[head][tail] string += F"{head} -> {tail} == {weight}\n" return string.rstrip('\n' ) def UpperCamelCase_ ( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Optional[int] = [] for tail in self.adjacency: for head in self.adjacency[tail]: output.append((tail, head, self.adjacency[head][tail]) ) return output def UpperCamelCase_ ( self ): '''simple docstring''' return self.adjacency.keys() @staticmethod def UpperCamelCase_ ( A=None, A=None ): '''simple docstring''' SCREAMING_SNAKE_CASE : List[str] = Graph() if vertices is None: SCREAMING_SNAKE_CASE : Optional[Any] = [] if edges is None: SCREAMING_SNAKE_CASE : List[Any] = [] for vertex in vertices: g.add_vertex(A ) for edge in edges: g.add_edge(*A ) return g class _a : '''simple docstring''' def __init__( self ): '''simple docstring''' SCREAMING_SNAKE_CASE : Union[str, Any] = {} SCREAMING_SNAKE_CASE : str = {} def __len__( self ): '''simple docstring''' return len(self.parent ) def UpperCamelCase_ ( self, A ): '''simple docstring''' if item in self.parent: return self.find(A ) SCREAMING_SNAKE_CASE : Dict = item SCREAMING_SNAKE_CASE : Tuple = 0 return item def UpperCamelCase_ ( self, A ): '''simple docstring''' if item not in self.parent: return self.make_set(A ) if item != self.parent[item]: SCREAMING_SNAKE_CASE : Tuple = self.find(self.parent[item] ) return self.parent[item] def UpperCamelCase_ ( self, A, A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Dict = self.find(A ) SCREAMING_SNAKE_CASE : Optional[Any] = self.find(A ) if roota == roota: return roota if self.rank[roota] > self.rank[roota]: SCREAMING_SNAKE_CASE : Tuple = roota return roota if self.rank[roota] < self.rank[roota]: SCREAMING_SNAKE_CASE : Optional[int] = roota return roota if self.rank[roota] == self.rank[roota]: self.rank[roota] += 1 SCREAMING_SNAKE_CASE : Tuple = roota return roota return None @staticmethod def UpperCamelCase_ ( A ): '''simple docstring''' SCREAMING_SNAKE_CASE : Any = graph.num_vertices SCREAMING_SNAKE_CASE : List[str] = Graph.UnionFind() SCREAMING_SNAKE_CASE : Dict = [] while num_components > 1: SCREAMING_SNAKE_CASE : List[Any] = {} for vertex in graph.get_vertices(): SCREAMING_SNAKE_CASE : str = -1 SCREAMING_SNAKE_CASE : Tuple = graph.get_edges() for edge in edges: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : List[Any] = edge edges.remove((tail, head, weight) ) for edge in edges: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : int = edge SCREAMING_SNAKE_CASE : Optional[int] = union_find.find(A ) SCREAMING_SNAKE_CASE : int = union_find.find(A ) if seta != seta: if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE : str = [head, tail, weight] if cheap_edge[seta] == -1 or cheap_edge[seta][2] > weight: SCREAMING_SNAKE_CASE : Optional[int] = [head, tail, weight] for vertex in cheap_edge: if cheap_edge[vertex] != -1: SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE : Optional[int] = cheap_edge[vertex] if union_find.find(A ) != union_find.find(A ): union_find.union(A, A ) mst_edges.append(cheap_edge[vertex] ) SCREAMING_SNAKE_CASE : Dict = num_components - 1 SCREAMING_SNAKE_CASE : Tuple = Graph.build(edges=A ) return mst
251
1
"""simple docstring""" import random import unittest from torch.utils.data import BatchSampler, DataLoader, IterableDataset from accelerate import Accelerator from accelerate.data_loader import ( BatchSamplerShard, DataLoaderDispatcher, DataLoaderShard, IterableDatasetShard, SkipBatchSampler, SkipDataLoader, skip_first_batches, ) class _lowerCAmelCase ( _UpperCAmelCase ): """simple docstring""" def __init__( self , __UpperCAmelCase=0.01 , __UpperCAmelCase=1_0_0_0 ): '''simple docstring''' lowerCAmelCase__ :List[Any] = p_stop lowerCAmelCase__ :List[Any] = max_length def __iter__( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = 0 lowerCAmelCase__ :Union[str, Any] = False while not stop and count < self.max_length: yield count count += 1 lowerCAmelCase__ :Optional[Any] = random.random() < self.p_stop class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=True ): '''simple docstring''' lowerCAmelCase__ :List[str] = [ BatchSamplerShard(lowercase_ , 2 , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) for i in range(2 ) ] lowerCAmelCase__ :Dict = [list(lowercase_ ) for batch_sampler_shard in batch_sampler_shards] if not split_batches: self.assertListEqual([len(lowercase_ ) for shard in batch_sampler_shards] , [len(lowercase_ ) for e in expected] ) self.assertListEqual(lowercase_ , lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :int = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) lowerCAmelCase__ :Any = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase_ , lowercase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [0, 1, 2]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) lowerCAmelCase__ :int = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Dict = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCAmelCase__ :List[Any] = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[Any] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 0, 1]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) lowerCAmelCase__ :Optional[int] = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Optional[int] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCAmelCase__ :Optional[Any] = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [1, 2, 3]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ :int = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Optional[Any] = [[[0, 1, 0]], [[1, 0, 1]]] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) lowerCAmelCase__ :Optional[int] = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [[], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [0, 1]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) lowerCAmelCase__ :Dict = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :List[Any] = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCAmelCase__ :str = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 0]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [1, 2]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) lowerCAmelCase__ :str = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :str = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ :int = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Tuple = [[[0, 1]], [[0, 1]]] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) lowerCAmelCase__ :List[str] = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [[], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1, 2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Tuple = BatchSampler(range(2_4 ) , batch_size=3 , drop_last=lowercase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is a round multiple of batch size but not total batch size. lowerCAmelCase__ :Optional[Any] = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[Any] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Tuple = BatchSampler(range(2_1 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size but has a multiple of # num_processes batch. lowerCAmelCase__ :Dict = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9, 2_0]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7], [2_1]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[Any] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size but and has not a multiple of # num_processes batch. lowerCAmelCase__ :Optional[int] = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :int = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4], [1_8, 1_9]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Dict = BatchSampler(range(2_0 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Optional[Any] = [ [[0, 1, 2], [6, 7, 8], [1_2, 1_3, 1_4]], [[3, 4, 5], [9, 1_0, 1_1], [1_5, 1_6, 1_7]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ :Tuple = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :List[str] = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Optional[int] = BatchSampler(range(2 ) , batch_size=3 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [[], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ , even_batches=lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Union[str, Any] = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9], [2_2, 2_3]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :int = BatchSampler(range(2_4 ) , batch_size=4 , drop_last=lowercase_ ) # Expected shouldn't change self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size. lowerCAmelCase__ :Dict = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Any = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0, 2_1]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Optional[int] = BatchSampler(range(2_2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :Optional[int] = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is not a round multiple of batch size or num_processes. lowerCAmelCase__ :Any = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :int = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7], [2_0]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :Union[str, Any] = BatchSampler(range(2_1 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :int = [ [[0, 1], [4, 5], [8, 9], [1_2, 1_3], [1_6, 1_7]], [[2, 3], [6, 7], [1_0, 1_1], [1_4, 1_5], [1_8, 1_9]], ] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) # Check the shards when the dataset is very small. lowerCAmelCase__ :Optional[Any] = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :int = [[[0, 1]], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) lowerCAmelCase__ :str = BatchSampler(range(2 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :str = [[], []] self.check_batch_sampler_shards(lowercase_ , lowercase_ , split_batches=lowercase_ , even_batches=lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[str] = [[0, 1, 2], [3, 4], [5, 6, 7, 8], [9, 1_0, 1_1], [1_2, 1_3]] lowerCAmelCase__ :str = [BatchSamplerShard(lowercase_ , 2 , lowercase_ , even_batches=lowercase_ ) for i in range(2 )] self.assertEqual(len(batch_sampler_shards[0] ) , 3 ) self.assertEqual(len(batch_sampler_shards[1] ) , 2 ) self.assertListEqual(list(batch_sampler_shards[0] ) , [[0, 1, 2], [5, 6, 7, 8], [1_2, 1_3]] ) self.assertListEqual(list(batch_sampler_shards[1] ) , [[3, 4], [9, 1_0, 1_1]] ) def snake_case ( self , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase=False , __UpperCAmelCase=2 , __UpperCAmelCase=False ): '''simple docstring''' random.seed(lowercase_ ) lowerCAmelCase__ :List[str] = list(lowercase_ ) lowerCAmelCase__ :Optional[Any] = [ IterableDatasetShard( lowercase_ , batch_size=lowercase_ , drop_last=lowercase_ , num_processes=lowercase_ , process_index=lowercase_ , split_batches=lowercase_ , ) for i in range(lowercase_ ) ] lowerCAmelCase__ :List[str] = [] for iterable_dataset_shard in iterable_dataset_shards: # Since our random iterable dataset will be... random... we need to use a seed to get reproducible results. random.seed(lowercase_ ) iterable_dataset_lists.append(list(lowercase_ ) ) lowerCAmelCase__ :Dict = batch_size // num_processes if split_batches else batch_size # All iterable dataset shard should have the same length, a round multiple of shard_batch_size lowerCAmelCase__ :str = iterable_dataset_lists[0] for l in iterable_dataset_lists[1:]: self.assertEqual(len(lowercase_ ) , len(lowercase_ ) ) self.assertTrue(len(lowercase_ ) % shard_batch_size == 0 ) lowerCAmelCase__ :int = [] for idx in range(0 , len(lowercase_ ) , lowercase_ ): for l in iterable_dataset_lists: observed += l[idx : idx + shard_batch_size] if not drop_last: while len(lowercase_ ) < len(lowercase_ ): reference += reference self.assertListEqual(lowercase_ , reference[: len(lowercase_ )] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Tuple = 4_2 lowerCAmelCase__ :Dict = RandomIterableDataset() self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) # Edge case with a very small dataset lowerCAmelCase__ :Tuple = RandomIterableDataset(max_length=2 ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) self.check_iterable_dataset_shards(lowercase_ , lowercase_ , batch_size=4 , drop_last=lowercase_ , split_batches=lowercase_ ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :List[Any] = BatchSampler(range(1_6 ) , batch_size=4 , drop_last=lowercase_ ) lowerCAmelCase__ :int = SkipBatchSampler(lowercase_ , 2 ) self.assertListEqual(list(lowercase_ ) , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = SkipDataLoader(list(range(1_6 ) ) , batch_size=4 , skip_batches=2 ) self.assertListEqual([t.tolist() for t in dataloader] , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :Optional[Any] = DataLoader(list(range(1_6 ) ) , batch_size=4 ) lowerCAmelCase__ :int = skip_first_batches(lowercase_ , num_batches=2 ) self.assertListEqual([t.tolist() for t in new_dataloader] , [[8, 9, 1_0, 1_1], [1_2, 1_3, 1_4, 1_5]] ) def snake_case ( self ): '''simple docstring''' lowerCAmelCase__ :int = DataLoaderShard(list(range(1_6 ) ) , batch_size=4 ) for idx, _ in enumerate(lowercase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) def snake_case ( self ): '''simple docstring''' Accelerator() lowerCAmelCase__ :Tuple = DataLoaderDispatcher(range(1_6 ) , batch_size=4 ) for idx, _ in enumerate(lowercase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 ) # Test it also works on the second iteration for idx, _ in enumerate(lowercase_ ): self.assertEqual(dataloader.end_of_dataloader , idx == 3 )
371
"""simple docstring""" def __A (_SCREAMING_SNAKE_CASE ) ->int: """simple docstring""" lowerCAmelCase__ :list[list[int]] = [[0 for _ in range(_SCREAMING_SNAKE_CASE )] for _ in range(m + 1 )] for i in range(m + 1 ): lowerCAmelCase__ :str = 1 for n in range(m + 1 ): for k in range(1 , _SCREAMING_SNAKE_CASE ): memo[n][k] += memo[n][k - 1] if n - k > 0: memo[n][k] += memo[n - k - 1][k] return memo[m][m - 1] if __name__ == "__main__": import sys if len(sys.argv) == 1: try: __A = int(input("""Enter a number: """).strip()) print(partition(n)) except ValueError: print("""Please enter a number.""") else: try: __A = int(sys.argv[1]) print(partition(n)) except ValueError: print("""Please pass a number.""")
254
0
'''simple docstring''' import pickle import shutil import tempfile import unittest from transformers import SPIECE_UNDERLINE, XLMRobertaTokenizer, XLMRobertaTokenizerFast from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin snake_case_ : List[Any] = get_tests_dir('fixtures/test_sentencepiece.model') @require_sentencepiece @require_tokenizers class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = XLMRobertaTokenizer lowercase__ = XLMRobertaTokenizerFast lowercase__ = True lowercase__ = True def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' super().setUp() # We have a SentencePiece fixture for testing _UpperCamelCase : Tuple = XLMRobertaTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) tokenizer.save_pretrained(self.tmpdirname ) def UpperCamelCase_ ( self : Dict ): '''simple docstring''' _UpperCamelCase : Optional[int] = '<pad>' _UpperCamelCase : str = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(lowerCamelCase__ ) ,lowerCamelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(lowerCamelCase__ ) ,lowerCamelCase__ ) def UpperCamelCase_ ( self : Optional[Any] ): '''simple docstring''' _UpperCamelCase : Any = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] ,'<s>' ) self.assertEqual(vocab_keys[1] ,'<pad>' ) self.assertEqual(vocab_keys[-1] ,'<mask>' ) self.assertEqual(len(lowerCamelCase__ ) ,1002 ) def UpperCamelCase_ ( self : str ): '''simple docstring''' self.assertEqual(self.get_tokenizer().vocab_size ,1002 ) def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' _UpperCamelCase : str = XLMRobertaTokenizer(lowerCamelCase__ ,keep_accents=lowerCamelCase__ ) _UpperCamelCase : Dict = tokenizer.tokenize('This is a test' ) self.assertListEqual(lowerCamelCase__ ,['▁This', '▁is', '▁a', '▁t', 'est'] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) ,[value + tokenizer.fairseq_offset for value in [285, 46, 10, 170, 382]] ,) _UpperCamelCase : Union[str, Any] = tokenizer.tokenize('I was born in 92000, and this is falsé.' ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '9', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', 'é', '.', ] ,) _UpperCamelCase : int = tokenizer.convert_tokens_to_ids(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ value + tokenizer.fairseq_offset for value in [8, 21, 84, 55, 24, 19, 7, 2, 602, 347, 347, 347, 3, 12, 66, 46, 72, 80, 6, 2, 4] # ^ unk: 2 + 1 = 3 unk: 2 + 1 = 3 ^ ] ,) _UpperCamelCase : Union[str, Any] = tokenizer.convert_ids_to_tokens(lowerCamelCase__ ) self.assertListEqual( lowerCamelCase__ ,[ SPIECE_UNDERLINE + 'I', SPIECE_UNDERLINE + 'was', SPIECE_UNDERLINE + 'b', 'or', 'n', SPIECE_UNDERLINE + 'in', SPIECE_UNDERLINE + '', '<unk>', '2', '0', '0', '0', ',', SPIECE_UNDERLINE + 'and', SPIECE_UNDERLINE + 'this', SPIECE_UNDERLINE + 'is', SPIECE_UNDERLINE + 'f', 'al', 's', '<unk>', '.', ] ,) def UpperCamelCase_ ( self : str ): '''simple docstring''' if not self.test_slow_tokenizer: # as we don't have a slow version, we can't compare the outputs between slow and fast versions return _UpperCamelCase : Tuple = (self.rust_tokenizer_class, 'hf-internal-testing/tiny-xlm-roberta', {}) for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): _UpperCamelCase : List[Any] = self.rust_tokenizer_class.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Tuple = self.tokenizer_class.from_pretrained(lowerCamelCase__ ,**lowerCamelCase__ ) _UpperCamelCase : Optional[int] = tempfile.mkdtemp() _UpperCamelCase : str = tokenizer_r.save_pretrained(lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files + the tokenizer.json file for the fast one self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) _UpperCamelCase : str = tuple(f for f in tokenizer_r_files if 'tokenizer.json' not in f ) self.assertSequenceEqual(lowerCamelCase__ ,lowerCamelCase__ ) # Checks everything loads correctly in the same way _UpperCamelCase : int = tokenizer_r.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : str = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ ,lowerCamelCase__ ) ) # self.assertEqual(getattr(tokenizer_rp, key), getattr(tokenizer_pp, key)) # self.assertEqual(getattr(tokenizer_rp, key + "_id"), getattr(tokenizer_pp, key + "_id")) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=True _UpperCamelCase : str = tempfile.mkdtemp() _UpperCamelCase : Any = tokenizer_r.save_pretrained(lowerCamelCase__ ,legacy_format=lowerCamelCase__ ) _UpperCamelCase : Optional[Any] = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it save with the same files self.assertSequenceEqual(lowerCamelCase__ ,lowerCamelCase__ ) # Checks everything loads correctly in the same way _UpperCamelCase : Tuple = tokenizer_r.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : str = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ ,lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) # Save tokenizer rust, legacy_format=False _UpperCamelCase : Tuple = tempfile.mkdtemp() _UpperCamelCase : Dict = tokenizer_r.save_pretrained(lowerCamelCase__ ,legacy_format=lowerCamelCase__ ) _UpperCamelCase : int = tokenizer_p.save_pretrained(lowerCamelCase__ ) # Checks it saved the tokenizer.json file self.assertTrue(any('tokenizer.json' in f for f in tokenizer_r_files ) ) # Checks everything loads correctly in the same way _UpperCamelCase : List[Any] = tokenizer_r.from_pretrained(lowerCamelCase__ ) _UpperCamelCase : Any = tokenizer_p.from_pretrained(lowerCamelCase__ ) # Check special tokens are set accordingly on Rust and Python for key in tokenizer_pp.special_tokens_map: self.assertTrue(hasattr(lowerCamelCase__ ,lowerCamelCase__ ) ) shutil.rmtree(lowerCamelCase__ ) @cached_property def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' return XLMRobertaTokenizer.from_pretrained('xlm-roberta-base' ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' with tempfile.NamedTemporaryFile() as f: shutil.copyfile(lowerCamelCase__ ,f.name ) _UpperCamelCase : Optional[Any] = XLMRobertaTokenizer(f.name ,keep_accents=lowerCamelCase__ ) _UpperCamelCase : str = pickle.dumps(lowerCamelCase__ ) pickle.loads(lowerCamelCase__ ) def UpperCamelCase_ ( self : Tuple ): '''simple docstring''' if not self.test_rust_tokenizer: return _UpperCamelCase : int = self.get_tokenizer() _UpperCamelCase : List[str] = self.get_rust_tokenizer() _UpperCamelCase : Any = 'I was born in 92000, and this is falsé.' _UpperCamelCase : str = tokenizer.tokenize(lowerCamelCase__ ) _UpperCamelCase : Any = rust_tokenizer.tokenize(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) _UpperCamelCase : Tuple = rust_tokenizer.encode(lowerCamelCase__ ,add_special_tokens=lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) _UpperCamelCase : Tuple = self.get_rust_tokenizer() _UpperCamelCase : Union[str, Any] = tokenizer.encode(lowerCamelCase__ ) _UpperCamelCase : Optional[int] = rust_tokenizer.encode(lowerCamelCase__ ) self.assertListEqual(lowerCamelCase__ ,lowerCamelCase__ ) @slow def UpperCamelCase_ ( self : Any ): '''simple docstring''' _UpperCamelCase : Union[str, Any] = 'Hello World!' _UpperCamelCase : str = [0, 35378, 6661, 38, 2] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : Optional[Any] = ( 'This is a very long text with a lot of weird characters, such as: . , ~ ? ( ) " [ ] ! : - . Also we will' ' add words that should not exsist and be tokenized to <unk>, such as saoneuhaoesuth' ) _UpperCamelCase : str = [ 0, 3293, 83, 10, 4552, 4989, 7986, 678, 10, 5915, 111, 179459, 124850, 4, 6044, 237, 12, 6, 5, 6, 4, 6780, 705, 15, 1388, 44, 378, 10114, 711, 152, 20, 6, 5, 22376, 642, 1221, 15190, 34153, 450, 5608, 959, 1119, 57702, 136, 186, 47, 1098, 29367, 47, # 4426, # What fairseq tokenizes from "<unk>": "_<" # 3678, # What fairseq tokenizes from "<unk>": "unk" # 2740, # What fairseq tokenizes from "<unk>": ">" 3, # What we tokenize from "<unk>": "<unk>" 6, # Residue from the tokenization: an extra sentencepiece underline 4, 6044, 237, 6284, 50901, 528, 31, 90, 34, 927, 2, ] # xlmr = torch.hub.load('pytorch/fairseq', 'xlmr.base') # xlmr.large has same tokenizer # xlmr.eval() # xlmr.encode(symbols) self.assertListEqual(lowerCamelCase__ ,self.big_tokenizer.encode(lowerCamelCase__ ) ) @slow def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' # fmt: off _UpperCamelCase : Union[str, Any] = {'input_ids': [[0, 11062, 82772, 7, 15, 82772, 538, 51529, 237, 17198, 1290, 206, 9, 215175, 1314, 136, 17198, 1290, 206, 9, 56359, 42, 122009, 9, 16466, 16, 87344, 4537, 9, 4717, 78381, 6, 159958, 7, 15, 24480, 618, 4, 527, 22693, 5428, 4, 2777, 24480, 9874, 4, 43523, 594, 4, 803, 18392, 33189, 18, 4, 43523, 24447, 12399, 100, 24955, 83658, 9626, 144057, 15, 839, 22335, 16, 136, 24955, 83658, 83479, 15, 39102, 724, 16, 678, 645, 2789, 1328, 4589, 42, 122009, 115774, 23, 805, 1328, 46876, 7, 136, 53894, 1940, 42227, 41159, 17721, 823, 425, 4, 27512, 98722, 206, 136, 5531, 4970, 919, 17336, 5, 2], [0, 20080, 618, 83, 82775, 47, 479, 9, 1517, 73, 53894, 333, 80581, 110117, 18811, 5256, 1295, 51, 152526, 297, 7986, 390, 124416, 538, 35431, 214, 98, 15044, 25737, 136, 7108, 43701, 23, 756, 135355, 7, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 581, 63773, 119455, 6, 147797, 88203, 7, 645, 70, 21, 3285, 10269, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], 'attention_mask': [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=lowerCamelCase__ ,model_name='xlm-roberta-base' ,revision='d9d8a8ea5eb94b1c6654ae9249df7793cd2933d3' ,)
83
'''simple docstring''' from typing import Callable, Optional, Union from ...configuration_utils import PretrainedConfig from ...utils import logging snake_case_ : Optional[Any] = logging.get_logger(__name__) snake_case_ : int = { 'microsoft/xprophetnet-large-wiki100-cased': ( 'https://huggingface.co/microsoft/xprophetnet-large-wiki100-cased/resolve/main/config.json' ), } class lowercase__ ( lowercase ): lowercase__ = """xlm-prophetnet""" lowercase__ = ["""past_key_values"""] lowercase__ = { """num_attention_heads""": """num_encoder_attention_heads""", } def __init__( self : Optional[int] ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[Union[str, Callable]] = "gelu" ,lowerCamelCase__ : Optional[int] = 30522 ,lowerCamelCase__ : Optional[int] = 1024 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[int] = 4096 ,lowerCamelCase__ : Optional[int] = 12 ,lowerCamelCase__ : Optional[int] = 16 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[float] = 0.1 ,lowerCamelCase__ : Optional[int] = 512 ,lowerCamelCase__ : Optional[float] = 0.0_2 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 2 ,lowerCamelCase__ : Optional[int] = 32 ,lowerCamelCase__ : Optional[int] = 128 ,lowerCamelCase__ : Optional[bool] = False ,lowerCamelCase__ : Optional[float] = 0.0 ,lowerCamelCase__ : Optional[bool] = True ,lowerCamelCase__ : Optional[int] = 0 ,lowerCamelCase__ : Optional[int] = 1 ,lowerCamelCase__ : Optional[int] = 2 ,**lowerCamelCase__ : Union[str, Any] ,): '''simple docstring''' _UpperCamelCase : List[Any] = vocab_size _UpperCamelCase : Union[str, Any] = hidden_size _UpperCamelCase : str = encoder_ffn_dim _UpperCamelCase : List[Any] = num_encoder_layers _UpperCamelCase : Tuple = num_encoder_attention_heads _UpperCamelCase : Optional[int] = decoder_ffn_dim _UpperCamelCase : List[Any] = num_decoder_layers _UpperCamelCase : List[Any] = num_decoder_attention_heads _UpperCamelCase : Optional[Any] = max_position_embeddings _UpperCamelCase : str = init_std # Normal(0, this parameter) _UpperCamelCase : List[str] = activation_function # parameters for xlmprophetnet _UpperCamelCase : Tuple = ngram _UpperCamelCase : Optional[Any] = num_buckets _UpperCamelCase : Tuple = relative_max_distance _UpperCamelCase : str = disable_ngram_loss _UpperCamelCase : str = eps # 3 Types of Dropout _UpperCamelCase : Union[str, Any] = attention_dropout _UpperCamelCase : str = activation_dropout _UpperCamelCase : List[str] = dropout _UpperCamelCase : Tuple = use_cache super().__init__( pad_token_id=lowerCamelCase__ ,bos_token_id=lowerCamelCase__ ,eos_token_id=lowerCamelCase__ ,is_encoder_decoder=lowerCamelCase__ ,add_cross_attention=lowerCamelCase__ ,decoder_start_token_id=lowerCamelCase__ ,**lowerCamelCase__ ,) @property def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' return self.num_encoder_layers + self.num_decoder_layers @num_hidden_layers.setter def UpperCamelCase_ ( self : str ,lowerCamelCase__ : Union[str, Any] ): '''simple docstring''' raise NotImplementedError( 'This model does not support the setting of `num_hidden_layers`. Please set `num_encoder_layers` and' ' `num_decoder_layers`.' )
83
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices lowercase__ = logging.get_logger(__name__) lowercase__ = { """microsoft/resnet-50""": """https://huggingface.co/microsoft/resnet-50/blob/main/config.json""", } class __lowerCamelCase ( A__ , A__ ): '''simple docstring''' a_ : Any = """resnet""" a_ : int = ["""basic""", """bottleneck"""] def __init__( self : Union[str, Any] , a_ : Optional[int]=3 , a_ : List[str]=64 , a_ : Optional[int]=[2_56, 5_12, 10_24, 20_48] , a_ : List[Any]=[3, 4, 6, 3] , a_ : Dict="bottleneck" , a_ : Optional[int]="relu" , a_ : Dict=False , a_ : List[str]=None , a_ : Tuple=None , **a_ : int , ): super().__init__(**a_ ) if layer_type not in self.layer_types: raise ValueError(f'''layer_type={layer_type} is not one of {",".join(self.layer_types )}''' ) lowerCAmelCase_ : Union[str, Any] = num_channels lowerCAmelCase_ : str = embedding_size lowerCAmelCase_ : int = hidden_sizes lowerCAmelCase_ : Dict = depths lowerCAmelCase_ : Union[str, Any] = layer_type lowerCAmelCase_ : List[Any] = hidden_act lowerCAmelCase_ : Union[str, Any] = downsample_in_first_stage lowerCAmelCase_ : Dict = ["stem"] + [f'''stage{idx}''' for idx in range(1 , len(a_ ) + 1 )] lowerCAmelCase_ : Optional[int] = get_aligned_output_features_output_indices( out_features=a_ , out_indices=a_ , stage_names=self.stage_names ) class __lowerCamelCase ( A__ ): '''simple docstring''' a_ : Optional[int] = version.parse("""1.11""" ) @property def lowerCamelCase ( self : Any ): return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowerCamelCase ( self : Optional[Any] ): return 1e-3
366
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class __lowerCamelCase ( unittest.TestCase ): '''simple docstring''' def __init__( self : List[str] , a_ : List[str] , a_ : Tuple=7 , a_ : Any=3 , a_ : Union[str, Any]=18 , a_ : List[str]=30 , a_ : List[str]=4_00 , a_ : str=True , a_ : Tuple=None , a_ : str=True , a_ : Optional[int]=None , ): lowerCAmelCase_ : Any = size if size is not None else {"shortest_edge": 20} lowerCAmelCase_ : Any = crop_size if crop_size is not None else {"height": 18, "width": 18} lowerCAmelCase_ : int = parent lowerCAmelCase_ : Dict = batch_size lowerCAmelCase_ : Any = num_channels lowerCAmelCase_ : str = image_size lowerCAmelCase_ : int = min_resolution lowerCAmelCase_ : Tuple = max_resolution lowerCAmelCase_ : str = do_resize lowerCAmelCase_ : List[Any] = size lowerCAmelCase_ : Any = do_center_crop lowerCAmelCase_ : Tuple = crop_size def lowerCamelCase ( self : List[str] ): return { "do_resize": self.do_resize, "size": self.size, "do_center_crop": self.do_center_crop, "crop_size": self.crop_size, } @require_torch @require_vision class __lowerCamelCase ( A__ , unittest.TestCase ): '''simple docstring''' a_ : Optional[Any] = MobileNetVaImageProcessor if is_vision_available() else None def lowerCamelCase ( self : Optional[int] ): lowerCAmelCase_ : int = MobileNetVaImageProcessingTester(self ) @property def lowerCamelCase ( self : Optional[Any] ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(a_ , "do_resize" ) ) self.assertTrue(hasattr(a_ , "size" ) ) self.assertTrue(hasattr(a_ , "do_center_crop" ) ) self.assertTrue(hasattr(a_ , "crop_size" ) ) def lowerCamelCase ( self : Tuple ): lowerCAmelCase_ : Optional[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"shortest_edge": 20} ) self.assertEqual(image_processor.crop_size , {"height": 18, "width": 18} ) lowerCAmelCase_ : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"shortest_edge": 42} ) self.assertEqual(image_processor.crop_size , {"height": 84, "width": 84} ) def lowerCamelCase ( self : Tuple ): pass def lowerCamelCase ( self : Any ): # Initialize image_processing lowerCAmelCase_ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL images lowerCAmelCase_ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=a_ ) for image in image_inputs: self.assertIsInstance(a_ , Image.Image ) # Test not batched input lowerCAmelCase_ : str = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase_ : List[str] = image_processing(a_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowerCamelCase ( self : str ): # Initialize image_processing lowerCAmelCase_ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase_ : List[Any] = prepare_image_inputs(self.image_processor_tester , equal_resolution=a_ , numpify=a_ ) for image in image_inputs: self.assertIsInstance(a_ , np.ndarray ) # Test not batched input lowerCAmelCase_ : Tuple = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase_ : Dict = image_processing(a_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) def lowerCamelCase ( self : Union[str, Any] ): # Initialize image_processing lowerCAmelCase_ : str = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase_ : int = prepare_image_inputs(self.image_processor_tester , equal_resolution=a_ , torchify=a_ ) for image in image_inputs: self.assertIsInstance(a_ , torch.Tensor ) # Test not batched input lowerCAmelCase_ : int = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , ) # Test batched lowerCAmelCase_ : str = image_processing(a_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["height"], self.image_processor_tester.crop_size["width"], ) , )
161
0
'''simple docstring''' def _UpperCamelCase ( __A ) -> int: '''simple docstring''' if not numbers: return 0 if not isinstance(__A , (list, tuple) ) or not all( isinstance(__A , __A ) for number in numbers ): raise ValueError("numbers must be an iterable of integers" ) UpperCamelCase__ = UpperCamelCase__ = UpperCamelCase__ = numbers[0] for i in range(1 , len(__A ) ): # update the maximum and minimum subarray products UpperCamelCase__ = numbers[i] if number < 0: UpperCamelCase__ , UpperCamelCase__ = min_till_now, max_till_now UpperCamelCase__ = max(__A , max_till_now * number ) UpperCamelCase__ = min(__A , min_till_now * number ) # update the maximum product found till now UpperCamelCase__ = max(__A , __A ) return max_prod
80
"""simple docstring""" import math class a : """simple docstring""" def __init__( self: List[Any] , UpperCamelCase: List[str]=0 ): # a graph with Node 0,1,...,N-1 """simple docstring""" A__ = n A__ = [ [math.inf for j in range(0 , UpperCamelCase )] for i in range(0 , UpperCamelCase ) ] # adjacency matrix for weight A__ = [ [math.inf for j in range(0 , UpperCamelCase )] for i in range(0 , UpperCamelCase ) ] # dp[i][j] stores minimum distance from i to j def UpperCamelCase ( self: Union[str, Any] , UpperCamelCase: Optional[int] , UpperCamelCase: Union[str, Any] , UpperCamelCase: Tuple ): """simple docstring""" A__ = w def UpperCamelCase ( self: int ): """simple docstring""" for k in range(0 , self.n ): for i in range(0 , self.n ): for j in range(0 , self.n ): A__ = min(self.dp[i][j] , self.dp[i][k] + self.dp[k][j] ) def UpperCamelCase ( self: int , UpperCamelCase: List[str] , UpperCamelCase: Dict ): """simple docstring""" return self.dp[u][v] if __name__ == "__main__": SCREAMING_SNAKE_CASE_ : List[Any] = Graph(5) graph.add_edge(0, 2, 9) graph.add_edge(0, 4, 1_0) graph.add_edge(1, 3, 5) graph.add_edge(2, 3, 7) graph.add_edge(3, 0, 1_0) graph.add_edge(3, 1, 2) graph.add_edge(3, 2, 1) graph.add_edge(3, 4, 6) graph.add_edge(4, 1, 3) graph.add_edge(4, 2, 4) graph.add_edge(4, 3, 9) graph.floyd_warshall() graph.show_min(1, 4) graph.show_min(0, 3)
335
0
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_video_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import VivitImageProcessor class SCREAMING_SNAKE_CASE ( unittest.TestCase ): """simple docstring""" def __init__( self : Optional[Any] ,lowercase_ : List[Any] ,lowercase_ : List[Any]=7 ,lowercase_ : List[str]=3 ,lowercase_ : Union[str, Any]=1_0 ,lowercase_ : Tuple=1_8 ,lowercase_ : Tuple=3_0 ,lowercase_ : Optional[Any]=4_0_0 ,lowercase_ : int=True ,lowercase_ : List[str]=None ,lowercase_ : List[str]=True ,lowercase_ : List[Any]=[0.5, 0.5, 0.5] ,lowercase_ : Dict=[0.5, 0.5, 0.5] ,lowercase_ : List[str]=None ,): lowerCAmelCase__ : Dict = size if size is not None else {'''shortest_edge''': 1_8} lowerCAmelCase__ : List[str] = crop_size if crop_size is not None else {'''height''': 1_8, '''width''': 1_8} lowerCAmelCase__ : Dict = parent lowerCAmelCase__ : Tuple = batch_size lowerCAmelCase__ : Union[str, Any] = num_channels lowerCAmelCase__ : Optional[Any] = num_frames lowerCAmelCase__ : List[str] = image_size lowerCAmelCase__ : Dict = min_resolution lowerCAmelCase__ : Tuple = max_resolution lowerCAmelCase__ : List[str] = do_resize lowerCAmelCase__ : Tuple = size lowerCAmelCase__ : List[Any] = do_normalize lowerCAmelCase__ : Tuple = image_mean lowerCAmelCase__ : str = image_std lowerCAmelCase__ : Any = crop_size def __lowerCAmelCase ( self : int ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class SCREAMING_SNAKE_CASE ( a_ , unittest.TestCase ): """simple docstring""" lowercase__ = VivitImageProcessor if is_vision_available() else None def __lowerCAmelCase ( self : Optional[Any] ): lowerCAmelCase__ : Dict = VivitImageProcessingTester(self ) @property def __lowerCAmelCase ( self : int ): return self.image_processor_tester.prepare_image_processor_dict() def __lowerCAmelCase ( self : List[str] ): lowerCAmelCase__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_ ,'''image_mean''' ) ) self.assertTrue(hasattr(lowercase_ ,'''image_std''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_normalize''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_resize''' ) ) self.assertTrue(hasattr(lowercase_ ,'''do_center_crop''' ) ) self.assertTrue(hasattr(lowercase_ ,'''size''' ) ) def __lowerCAmelCase ( self : Dict ): lowerCAmelCase__ : List[str] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 1_8} ) self.assertEqual(image_processor.crop_size ,{'''height''': 1_8, '''width''': 1_8} ) lowerCAmelCase__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ,size=4_2 ,crop_size=8_4 ) self.assertEqual(image_processor.size ,{'''shortest_edge''': 4_2} ) self.assertEqual(image_processor.crop_size ,{'''height''': 8_4, '''width''': 8_4} ) def __lowerCAmelCase ( self : Dict ): # Initialize image_processing lowerCAmelCase__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random PIL videos lowerCAmelCase__ : str = prepare_video_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ) for video in video_inputs: self.assertIsInstance(lowercase_ ,lowercase_ ) self.assertIsInstance(video[0] ,Image.Image ) # Test not batched input lowerCAmelCase__ : int = image_processing(video_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched lowerCAmelCase__ : Tuple = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def __lowerCAmelCase ( self : Dict ): # Initialize image_processing lowerCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors lowerCAmelCase__ : Union[str, Any] = prepare_video_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ,numpify=lowercase_ ) for video in video_inputs: self.assertIsInstance(lowercase_ ,lowercase_ ) self.assertIsInstance(video[0] ,np.ndarray ) # Test not batched input lowerCAmelCase__ : str = image_processing(video_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched lowerCAmelCase__ : Any = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) def __lowerCAmelCase ( self : Optional[Any] ): # Initialize image_processing lowerCAmelCase__ : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors lowerCAmelCase__ : Union[str, Any] = prepare_video_inputs(self.image_processor_tester ,equal_resolution=lowercase_ ,torchify=lowercase_ ) for video in video_inputs: self.assertIsInstance(lowercase_ ,lowercase_ ) self.assertIsInstance(video[0] ,torch.Tensor ) # Test not batched input lowerCAmelCase__ : Optional[Any] = image_processing(video_inputs[0] ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( 1, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,) # Test batched lowerCAmelCase__ : int = image_processing(lowercase_ ,return_tensors='''pt''' ).pixel_values self.assertEqual( encoded_videos.shape ,( self.image_processor_tester.batch_size, self.image_processor_tester.num_frames, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size['''height'''], self.image_processor_tester.crop_size['''width'''], ) ,)
74
"""simple docstring""" def __SCREAMING_SNAKE_CASE ( A_ ): lowerCAmelCase__ : int = 0 for ch in input_str: lowerCAmelCase__ : Any = ord(A_ ) lowerCAmelCase__ : Any = pow(2 , A_ ) # If we already turned on bit for current character's unicode if bitmap >> ch_unicode & 1 == 1: return False bitmap |= ch_bit_index_on return True if __name__ == "__main__": import doctest doctest.testmod()
74
1
from typing import Any import numpy as np def __SCREAMING_SNAKE_CASE ( snake_case_ ): '''simple docstring''' return np.array_equal(snake_case_ , matrix.conjugate().T ) def __SCREAMING_SNAKE_CASE ( snake_case_ , snake_case_ ): '''simple docstring''' _UpperCAmelCase = v.conjugate().T _UpperCAmelCase = v_star.dot(snake_case_ ) assert isinstance(snake_case_ , np.ndarray ) return (v_star_dot.dot(snake_case_ )) / (v_star.dot(snake_case_ )) def __SCREAMING_SNAKE_CASE ( ): '''simple docstring''' _UpperCAmelCase = np.array([[2, 2 + 1j, 4], [2 - 1j, 3, 1j], [4, -1j, 1]] ) _UpperCAmelCase = np.array([[1], [2], [3]] ) assert is_hermitian(snake_case_ ), f"""{a} is not hermitian.""" print(rayleigh_quotient(snake_case_ , snake_case_ ) ) _UpperCAmelCase = np.array([[1, 2, 4], [2, 3, -1], [4, -1, 1]] ) assert is_hermitian(snake_case_ ), f"""{a} is not hermitian.""" assert rayleigh_quotient(snake_case_ , snake_case_ ) == float(3 ) if __name__ == "__main__": import doctest doctest.testmod() tests()
133
lowercase_ : Optional[int] = [sum(int(c, 10) ** 2 for c in i.__str__()) for i in range(10_00_00)] def __SCREAMING_SNAKE_CASE ( snake_case_ ): '''simple docstring''' _UpperCAmelCase = 0 while number: # Increased Speed Slightly by checking every 5 digits together. sum_of_digits_squared += DIGITS_SQUARED[number % 10_0000] number //= 10_0000 return sum_of_digits_squared # There are 2 Chains made, # One ends with 89 with the chain member 58 being the one which when declared first, # there will be the least number of iterations for all the members to be checked. # The other one ends with 1 and has only one element 1. # So 58 and 1 are chosen to be declared at the starting. # Changed dictionary to an array to quicken the solution lowercase_ : list[bool | None] = [None] * 10_00_00_00 lowercase_ : Optional[int] = True lowercase_ : str = False def __SCREAMING_SNAKE_CASE ( snake_case_ ): '''simple docstring''' if CHAINS[number - 1] is not None: return CHAINS[number - 1] # type: ignore _UpperCAmelCase = chain(next_number(snake_case_ ) ) _UpperCAmelCase = number_chain while number < 1000_0000: _UpperCAmelCase = number_chain number *= 10 return number_chain def __SCREAMING_SNAKE_CASE ( snake_case_ = 1000_0000 ): '''simple docstring''' for i in range(1 , snake_case_ ): if CHAINS[i] is None: chain(i + 1 ) return CHAINS[:number].count(snake_case_ ) if __name__ == "__main__": import doctest doctest.testmod() print(f"""{solution() = }""")
133
1
'''simple docstring''' # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool __snake_case = { '''Acehnese Arabic''': '''ace_Arab''', '''Acehnese Latin''': '''ace_Latn''', '''Mesopotamian Arabic''': '''acm_Arab''', '''Ta\'izzi-Adeni Arabic''': '''acq_Arab''', '''Tunisian Arabic''': '''aeb_Arab''', '''Afrikaans''': '''afr_Latn''', '''South Levantine Arabic''': '''ajp_Arab''', '''Akan''': '''aka_Latn''', '''Amharic''': '''amh_Ethi''', '''North Levantine Arabic''': '''apc_Arab''', '''Modern Standard Arabic''': '''arb_Arab''', '''Modern Standard Arabic Romanized''': '''arb_Latn''', '''Najdi Arabic''': '''ars_Arab''', '''Moroccan Arabic''': '''ary_Arab''', '''Egyptian Arabic''': '''arz_Arab''', '''Assamese''': '''asm_Beng''', '''Asturian''': '''ast_Latn''', '''Awadhi''': '''awa_Deva''', '''Central Aymara''': '''ayr_Latn''', '''South Azerbaijani''': '''azb_Arab''', '''North Azerbaijani''': '''azj_Latn''', '''Bashkir''': '''bak_Cyrl''', '''Bambara''': '''bam_Latn''', '''Balinese''': '''ban_Latn''', '''Belarusian''': '''bel_Cyrl''', '''Bemba''': '''bem_Latn''', '''Bengali''': '''ben_Beng''', '''Bhojpuri''': '''bho_Deva''', '''Banjar Arabic''': '''bjn_Arab''', '''Banjar Latin''': '''bjn_Latn''', '''Standard Tibetan''': '''bod_Tibt''', '''Bosnian''': '''bos_Latn''', '''Buginese''': '''bug_Latn''', '''Bulgarian''': '''bul_Cyrl''', '''Catalan''': '''cat_Latn''', '''Cebuano''': '''ceb_Latn''', '''Czech''': '''ces_Latn''', '''Chokwe''': '''cjk_Latn''', '''Central Kurdish''': '''ckb_Arab''', '''Crimean Tatar''': '''crh_Latn''', '''Welsh''': '''cym_Latn''', '''Danish''': '''dan_Latn''', '''German''': '''deu_Latn''', '''Southwestern Dinka''': '''dik_Latn''', '''Dyula''': '''dyu_Latn''', '''Dzongkha''': '''dzo_Tibt''', '''Greek''': '''ell_Grek''', '''English''': '''eng_Latn''', '''Esperanto''': '''epo_Latn''', '''Estonian''': '''est_Latn''', '''Basque''': '''eus_Latn''', '''Ewe''': '''ewe_Latn''', '''Faroese''': '''fao_Latn''', '''Fijian''': '''fij_Latn''', '''Finnish''': '''fin_Latn''', '''Fon''': '''fon_Latn''', '''French''': '''fra_Latn''', '''Friulian''': '''fur_Latn''', '''Nigerian Fulfulde''': '''fuv_Latn''', '''Scottish Gaelic''': '''gla_Latn''', '''Irish''': '''gle_Latn''', '''Galician''': '''glg_Latn''', '''Guarani''': '''grn_Latn''', '''Gujarati''': '''guj_Gujr''', '''Haitian Creole''': '''hat_Latn''', '''Hausa''': '''hau_Latn''', '''Hebrew''': '''heb_Hebr''', '''Hindi''': '''hin_Deva''', '''Chhattisgarhi''': '''hne_Deva''', '''Croatian''': '''hrv_Latn''', '''Hungarian''': '''hun_Latn''', '''Armenian''': '''hye_Armn''', '''Igbo''': '''ibo_Latn''', '''Ilocano''': '''ilo_Latn''', '''Indonesian''': '''ind_Latn''', '''Icelandic''': '''isl_Latn''', '''Italian''': '''ita_Latn''', '''Javanese''': '''jav_Latn''', '''Japanese''': '''jpn_Jpan''', '''Kabyle''': '''kab_Latn''', '''Jingpho''': '''kac_Latn''', '''Kamba''': '''kam_Latn''', '''Kannada''': '''kan_Knda''', '''Kashmiri Arabic''': '''kas_Arab''', '''Kashmiri Devanagari''': '''kas_Deva''', '''Georgian''': '''kat_Geor''', '''Central Kanuri Arabic''': '''knc_Arab''', '''Central Kanuri Latin''': '''knc_Latn''', '''Kazakh''': '''kaz_Cyrl''', '''Kabiyè''': '''kbp_Latn''', '''Kabuverdianu''': '''kea_Latn''', '''Khmer''': '''khm_Khmr''', '''Kikuyu''': '''kik_Latn''', '''Kinyarwanda''': '''kin_Latn''', '''Kyrgyz''': '''kir_Cyrl''', '''Kimbundu''': '''kmb_Latn''', '''Northern Kurdish''': '''kmr_Latn''', '''Kikongo''': '''kon_Latn''', '''Korean''': '''kor_Hang''', '''Lao''': '''lao_Laoo''', '''Ligurian''': '''lij_Latn''', '''Limburgish''': '''lim_Latn''', '''Lingala''': '''lin_Latn''', '''Lithuanian''': '''lit_Latn''', '''Lombard''': '''lmo_Latn''', '''Latgalian''': '''ltg_Latn''', '''Luxembourgish''': '''ltz_Latn''', '''Luba-Kasai''': '''lua_Latn''', '''Ganda''': '''lug_Latn''', '''Luo''': '''luo_Latn''', '''Mizo''': '''lus_Latn''', '''Standard Latvian''': '''lvs_Latn''', '''Magahi''': '''mag_Deva''', '''Maithili''': '''mai_Deva''', '''Malayalam''': '''mal_Mlym''', '''Marathi''': '''mar_Deva''', '''Minangkabau Arabic ''': '''min_Arab''', '''Minangkabau Latin''': '''min_Latn''', '''Macedonian''': '''mkd_Cyrl''', '''Plateau Malagasy''': '''plt_Latn''', '''Maltese''': '''mlt_Latn''', '''Meitei Bengali''': '''mni_Beng''', '''Halh Mongolian''': '''khk_Cyrl''', '''Mossi''': '''mos_Latn''', '''Maori''': '''mri_Latn''', '''Burmese''': '''mya_Mymr''', '''Dutch''': '''nld_Latn''', '''Norwegian Nynorsk''': '''nno_Latn''', '''Norwegian Bokmål''': '''nob_Latn''', '''Nepali''': '''npi_Deva''', '''Northern Sotho''': '''nso_Latn''', '''Nuer''': '''nus_Latn''', '''Nyanja''': '''nya_Latn''', '''Occitan''': '''oci_Latn''', '''West Central Oromo''': '''gaz_Latn''', '''Odia''': '''ory_Orya''', '''Pangasinan''': '''pag_Latn''', '''Eastern Panjabi''': '''pan_Guru''', '''Papiamento''': '''pap_Latn''', '''Western Persian''': '''pes_Arab''', '''Polish''': '''pol_Latn''', '''Portuguese''': '''por_Latn''', '''Dari''': '''prs_Arab''', '''Southern Pashto''': '''pbt_Arab''', '''Ayacucho Quechua''': '''quy_Latn''', '''Romanian''': '''ron_Latn''', '''Rundi''': '''run_Latn''', '''Russian''': '''rus_Cyrl''', '''Sango''': '''sag_Latn''', '''Sanskrit''': '''san_Deva''', '''Santali''': '''sat_Olck''', '''Sicilian''': '''scn_Latn''', '''Shan''': '''shn_Mymr''', '''Sinhala''': '''sin_Sinh''', '''Slovak''': '''slk_Latn''', '''Slovenian''': '''slv_Latn''', '''Samoan''': '''smo_Latn''', '''Shona''': '''sna_Latn''', '''Sindhi''': '''snd_Arab''', '''Somali''': '''som_Latn''', '''Southern Sotho''': '''sot_Latn''', '''Spanish''': '''spa_Latn''', '''Tosk Albanian''': '''als_Latn''', '''Sardinian''': '''srd_Latn''', '''Serbian''': '''srp_Cyrl''', '''Swati''': '''ssw_Latn''', '''Sundanese''': '''sun_Latn''', '''Swedish''': '''swe_Latn''', '''Swahili''': '''swh_Latn''', '''Silesian''': '''szl_Latn''', '''Tamil''': '''tam_Taml''', '''Tatar''': '''tat_Cyrl''', '''Telugu''': '''tel_Telu''', '''Tajik''': '''tgk_Cyrl''', '''Tagalog''': '''tgl_Latn''', '''Thai''': '''tha_Thai''', '''Tigrinya''': '''tir_Ethi''', '''Tamasheq Latin''': '''taq_Latn''', '''Tamasheq Tifinagh''': '''taq_Tfng''', '''Tok Pisin''': '''tpi_Latn''', '''Tswana''': '''tsn_Latn''', '''Tsonga''': '''tso_Latn''', '''Turkmen''': '''tuk_Latn''', '''Tumbuka''': '''tum_Latn''', '''Turkish''': '''tur_Latn''', '''Twi''': '''twi_Latn''', '''Central Atlas Tamazight''': '''tzm_Tfng''', '''Uyghur''': '''uig_Arab''', '''Ukrainian''': '''ukr_Cyrl''', '''Umbundu''': '''umb_Latn''', '''Urdu''': '''urd_Arab''', '''Northern Uzbek''': '''uzn_Latn''', '''Venetian''': '''vec_Latn''', '''Vietnamese''': '''vie_Latn''', '''Waray''': '''war_Latn''', '''Wolof''': '''wol_Latn''', '''Xhosa''': '''xho_Latn''', '''Eastern Yiddish''': '''ydd_Hebr''', '''Yoruba''': '''yor_Latn''', '''Yue Chinese''': '''yue_Hant''', '''Chinese Simplified''': '''zho_Hans''', '''Chinese Traditional''': '''zho_Hant''', '''Standard Malay''': '''zsm_Latn''', '''Zulu''': '''zul_Latn''', } class lowercase ( A__ ): """simple docstring""" _a = 'facebook/nllb-200-distilled-600M' _a = ( 'This is a tool that translates text from a language to another. It takes three inputs: `text`, which should ' 'be the text to translate, `src_lang`, which should be the language of the text to translate and `tgt_lang`, ' 'which should be the language for the desired ouput language. Both `src_lang` and `tgt_lang` are written in ' 'plain English, such as \'Romanian\', or \'Albanian\'. It returns the text translated in `tgt_lang`.' ) _a = 'translator' _a = AutoTokenizer _a = AutoModelForSeqaSeqLM _a = LANGUAGE_CODES _a = ['text', 'text', 'text'] _a = ['text'] def lowerCAmelCase__ ( self , UpperCamelCase_ , UpperCamelCase_ , UpperCamelCase_ ): '''simple docstring''' if src_lang not in self.lang_to_code: raise ValueError(F'''{src_lang} is not a supported language.''' ) if tgt_lang not in self.lang_to_code: raise ValueError(F'''{tgt_lang} is not a supported language.''' ) UpperCamelCase__ :Dict = self.lang_to_code[src_lang] UpperCamelCase__ :Dict = self.lang_to_code[tgt_lang] return self.pre_processor._build_translation_inputs( UpperCamelCase_ , return_tensors='''pt''' , src_lang=UpperCamelCase_ , tgt_lang=UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return self.model.generate(**UpperCamelCase_ ) def lowerCAmelCase__ ( self , UpperCamelCase_ ): '''simple docstring''' return self.post_processor.decode(outputs[0].tolist() , skip_special_tokens=UpperCamelCase_ )
219
'''simple docstring''' from __future__ import annotations import math def a ( __a ) -> bool: '''simple docstring''' if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(__a ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def a ( __a ) -> list[int]: '''simple docstring''' UpperCamelCase__ :List[Any] = str(__a ) UpperCamelCase__ :Dict = [n] for i in range(1 , len(__a ) ): list_nums.append(int(str_num[i:] ) ) list_nums.append(int(str_num[:-i] ) ) return list_nums def a ( __a ) -> bool: '''simple docstring''' if len(str(__a ) ) > 3: if not is_prime(int(str(__a )[-3:] ) ) or not is_prime(int(str(__a )[:3] ) ): return False return True def a ( __a = 11 ) -> list[int]: '''simple docstring''' UpperCamelCase__ :list[int] = [] UpperCamelCase__ :int = 13 while len(__a ) != count: if validate(__a ): UpperCamelCase__ :Optional[int] = list_truncated_nums(__a ) if all(is_prime(__a ) for i in list_nums ): list_truncated_primes.append(__a ) num += 2 return list_truncated_primes def a ( ) -> int: '''simple docstring''' return sum(compute_truncated_primes(11 ) ) if __name__ == "__main__": print(F"""{sum(compute_truncated_primes(11)) = }""")
219
1
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets lowerCAmelCase__ = datasets.logging.get_logger(__name__) lowerCAmelCase__ = '''\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = "{COMET}: A Neural Framework for {MT} Evaluation", author = "Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.213", pages = "2685--2702", } ''' lowerCAmelCase__ = '''\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. ''' lowerCAmelCase__ = ''' COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric(\'comet\') >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use >>> source = ["Dem Feuer konnte Einhalt geboten werden", "Schulen und Kindergärten wurden eröffnet."] >>> hypothesis = ["The fire could be stopped", "Schools and kindergartens were open"] >>> reference = ["They were able to control the fire.", "Schools and kindergartens opened"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results["scores"]]) [0.19, 0.92] ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCamelCase ( datasets.Metric ): def snake_case_ (self ) -> Optional[Any]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage="https://unbabel.github.io/COMET/html/index.html" , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "sources": datasets.Value("string" , id="sequence" ), "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/Unbabel/COMET"] , reference_urls=[ "https://github.com/Unbabel/COMET", "https://www.aclweb.org/anthology/2020.emnlp-main.213/", "http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6", ] , ) def snake_case_ (self , __a ) -> str: if self.config_name == "default": UpperCamelCase = comet.load_from_checkpoint(comet.download_model("wmt20-comet-da" ) ) else: UpperCamelCase = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def snake_case_ (self , __a , __a , __a , __a=None , __a=False ) -> List[Any]: if gpus is None: UpperCamelCase = 1 if torch.cuda.is_available() else 0 UpperCamelCase = {"src": sources, "mt": predictions, "ref": references} UpperCamelCase = [dict(zip(__a , __a ) ) for t in zip(*data.values() )] UpperCamelCase , UpperCamelCase = self.scorer.predict(__a , gpus=__a , progress_bar=__a ) return {"mean_score": mean_score, "scores": scores}
153
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets lowerCAmelCase__ = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' lowerCAmelCase__ = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' lowerCAmelCase__ = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class _lowerCamelCase ( datasets.Metric ): def snake_case_ (self ) -> Optional[int]: return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("string" , id="sequence" ), "references": datasets.Value("string" , id="sequence" ), } ) , codebase_urls=["https://github.com/google-research/google-research/tree/master/rouge"] , reference_urls=[ "https://en.wikipedia.org/wiki/ROUGE_(metric)", "https://github.com/google-research/google-research/tree/master/rouge", ] , ) def snake_case_ (self , __a , __a , __a=None , __a=True , __a=False ) -> Union[str, Any]: if rouge_types is None: UpperCamelCase = ["rouge1", "rouge2", "rougeL", "rougeLsum"] UpperCamelCase = rouge_scorer.RougeScorer(rouge_types=__a , use_stemmer=__a ) if use_aggregator: UpperCamelCase = scoring.BootstrapAggregator() else: UpperCamelCase = [] for ref, pred in zip(__a , __a ): UpperCamelCase = scorer.score(__a , __a ) if use_aggregator: aggregator.add_scores(__a ) else: scores.append(__a ) if use_aggregator: UpperCamelCase = aggregator.aggregate() else: UpperCamelCase = {} for key in scores[0]: UpperCamelCase = [score[key] for score in scores] return result
153
1
"""simple docstring""" import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class UpperCAmelCase_ ( A_, A_, unittest.TestCase ): lowercase__ = IFInpaintingSuperResolutionPipeline lowercase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} lowercase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''} ) lowercase__ = PipelineTesterMixin.required_optional_params - {'''latents'''} def __magic_name__ ( self : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' return self._get_superresolution_dummy_components() def __magic_name__ ( self : Tuple , snake_case_ : List[str] , snake_case_ : Any=0 ) -> Optional[int]: '''simple docstring''' if str(snake_case_ ).startswith("mps" ): A__ = torch.manual_seed(snake_case_ ) else: A__ = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) A__ = floats_tensor((1, 3, 16, 16) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A__ = floats_tensor((1, 3, 32, 32) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) A__ = { "prompt": "A painting of a squirrel eating a burger", "image": image, "original_image": original_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def __magic_name__ ( self : Tuple ) -> Dict: '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1e-3 ) def __magic_name__ ( self : str ) -> List[Any]: '''simple docstring''' self._test_save_load_optional_components() @unittest.skipIf(torch_device != "cuda" , reason="float16 requires CUDA" ) def __magic_name__ ( self : Optional[int] ) -> Optional[Any]: '''simple docstring''' super().test_save_load_floataa(expected_max_diff=1e-1 ) def __magic_name__ ( self : str ) -> str: '''simple docstring''' self._test_attention_slicing_forward_pass(expected_max_diff=1e-2 ) def __magic_name__ ( self : str ) -> Dict: '''simple docstring''' self._test_save_load_local() def __magic_name__ ( self : str ) -> List[str]: '''simple docstring''' self._test_inference_batch_single_identical( expected_max_diff=1e-2 , )
230
"""simple docstring""" import argparse import os import transformers from .convert_slow_tokenizer import SLOW_TO_FAST_CONVERTERS from .utils import logging logging.set_verbosity_info() SCREAMING_SNAKE_CASE = logging.get_logger(__name__) SCREAMING_SNAKE_CASE = {name: getattr(transformers, name + "Fast") for name in SLOW_TO_FAST_CONVERTERS} def _SCREAMING_SNAKE_CASE ( lowercase_ , lowercase_ , lowercase_ , lowercase_ ) -> Any: if tokenizer_name is not None and tokenizer_name not in TOKENIZER_CLASSES: raise ValueError(f"""Unrecognized tokenizer name, should be one of {list(TOKENIZER_CLASSES.keys() )}.""" ) if tokenizer_name is None: A__ = TOKENIZER_CLASSES else: A__ = {tokenizer_name: getattr(lowercase_ , tokenizer_name + "Fast" )} logger.info(f"""Loading tokenizer classes: {tokenizer_names}""" ) for tokenizer_name in tokenizer_names: A__ = TOKENIZER_CLASSES[tokenizer_name] A__ = True if checkpoint_name is None: A__ = list(tokenizer_class.max_model_input_sizes.keys() ) else: A__ = [checkpoint_name] logger.info(f"""For tokenizer {tokenizer_class.__class__.__name__} loading checkpoints: {checkpoint_names}""" ) for checkpoint in checkpoint_names: logger.info(f"""Loading {tokenizer_class.__class__.__name__} {checkpoint}""" ) # Load tokenizer A__ = tokenizer_class.from_pretrained(lowercase_ , force_download=lowercase_ ) # Save fast tokenizer logger.info(f"""Save fast tokenizer to {dump_path} with prefix {checkpoint} add_prefix {add_prefix}""" ) # For organization names we create sub-directories if "/" in checkpoint: A__, A__ = checkpoint.split("/" ) A__ = os.path.join(lowercase_ , lowercase_ ) elif add_prefix: A__ = checkpoint A__ = dump_path else: A__ = None A__ = dump_path logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) if checkpoint in list(tokenizer.pretrained_vocab_files_map.values() )[0]: A__ = list(tokenizer.pretrained_vocab_files_map.values() )[0][checkpoint] A__ = file_path.split(lowercase_ )[-1][0] if next_char == "/": A__ = os.path.join(lowercase_ , lowercase_ ) A__ = None logger.info(f"""=> {dump_path_full} with prefix {checkpoint_prefix_name}, add_prefix {add_prefix}""" ) A__ = tokenizer.save_pretrained( lowercase_ , legacy_format=lowercase_ , filename_prefix=lowercase_ ) logger.info(f"""=> File names {file_names}""" ) for file_name in file_names: if not file_name.endswith("tokenizer.json" ): os.remove(lowercase_ ) logger.info(f"""=> removing {file_name}""" ) if __name__ == "__main__": SCREAMING_SNAKE_CASE = argparse.ArgumentParser() # Required parameters parser.add_argument( "--dump_path", default=None, type=str, required=True, help="Path to output generated fast tokenizer files." ) parser.add_argument( "--tokenizer_name", default=None, type=str, help=( f'Optional tokenizer type selected in the list of {list(TOKENIZER_CLASSES.keys())}. If not given, will ' "download and convert all the checkpoints from AWS." ), ) parser.add_argument( "--checkpoint_name", default=None, type=str, help="Optional checkpoint name. If not given, will download and convert the canonical checkpoints from AWS.", ) parser.add_argument( "--force_download", action="store_true", help="Re-download checkpoints.", ) SCREAMING_SNAKE_CASE = parser.parse_args() convert_slow_checkpoint_to_fast(args.tokenizer_name, args.checkpoint_name, args.dump_path, args.force_download)
230
1
class __A: def __init__( self , _snake_case , _snake_case ) -> Optional[int]: '''simple docstring''' __a = name __a = val def __str__( self ) -> List[Any]: '''simple docstring''' return F"""{self.__class__.__name__}({self.name}, {self.val})""" def __lt__( self , _snake_case ) -> Optional[Any]: '''simple docstring''' return self.val < other.val class __A: def __init__( self , _snake_case ) -> List[str]: '''simple docstring''' __a = {} __a = {} __a = self.build_heap(_snake_case ) def __getitem__( self , _snake_case ) -> Any: '''simple docstring''' return self.get_value(_snake_case ) def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Tuple: '''simple docstring''' return (idx - 1) // 2 def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> List[str]: '''simple docstring''' return idx * 2 + 1 def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Tuple: '''simple docstring''' return idx * 2 + 2 def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' return self.heap_dict[key] def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Dict: '''simple docstring''' __a = len(_snake_case ) - 1 __a = self.get_parent_idx(_snake_case ) for idx, i in enumerate(_snake_case ): __a = idx __a = i.val for i in range(_snake_case , -1 , -1 ): self.sift_down(_snake_case , _snake_case ) return array def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> List[Any]: '''simple docstring''' while True: __a = self.get_left_child_idx(_snake_case ) # noqa: E741 __a = self.get_right_child_idx(_snake_case ) __a = idx if l < len(_snake_case ) and array[l] < array[idx]: __a = l if r < len(_snake_case ) and array[r] < array[smallest]: __a = r if smallest != idx: __a , __a = array[smallest], array[idx] ( ( __a ) , ( __a ) , ) = ( self.idx_of_element[array[smallest]], self.idx_of_element[array[idx]], ) __a = smallest else: break def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> Union[str, Any]: '''simple docstring''' __a = self.get_parent_idx(_snake_case ) while p >= 0 and self.heap[p] > self.heap[idx]: __a , __a = self.heap[idx], self.heap[p] __a , __a = ( self.idx_of_element[self.heap[idx]], self.idx_of_element[self.heap[p]], ) __a = p __a = self.get_parent_idx(_snake_case ) def SCREAMING_SNAKE_CASE_ ( self ) -> Optional[Any]: '''simple docstring''' return self.heap[0] def SCREAMING_SNAKE_CASE_ ( self ) -> List[Any]: '''simple docstring''' __a , __a = self.heap[-1], self.heap[0] __a , __a = ( self.idx_of_element[self.heap[-1]], self.idx_of_element[self.heap[0]], ) __a = self.heap.pop() del self.idx_of_element[x] self.sift_down(0 , self.heap ) return x def SCREAMING_SNAKE_CASE_ ( self , _snake_case ) -> str: '''simple docstring''' self.heap.append(_snake_case ) __a = len(self.heap ) - 1 __a = node.val self.sift_up(len(self.heap ) - 1 ) def SCREAMING_SNAKE_CASE_ ( self ) -> int: '''simple docstring''' return len(self.heap ) == 0 def SCREAMING_SNAKE_CASE_ ( self , _snake_case , _snake_case ) -> Optional[Any]: '''simple docstring''' assert ( self.heap[self.idx_of_element[node]].val > new_value ), "newValue must be less that current value" __a = new_value __a = new_value self.sift_up(self.idx_of_element[node] ) A : Union[str, Any] = Node('R', -1) A : Union[str, Any] = Node('B', 6) A : List[Any] = Node('A', 3) A : List[Any] = Node('X', 1) A : Union[str, Any] = Node('E', 4) # Use one of these two ways to generate Min-Heap # Generating Min-Heap from array A : Union[str, Any] = MinHeap([r, b, a, x, e]) # Generating Min-Heap by Insert method # myMinHeap.insert(a) # myMinHeap.insert(b) # myMinHeap.insert(x) # myMinHeap.insert(r) # myMinHeap.insert(e) # Before print('Min Heap - before decrease key') for i in my_min_heap.heap: print(i) print('Min Heap - After decrease key of node [B -> -17]') my_min_heap.decrease_key(b, -1_7) # After for i in my_min_heap.heap: print(i) if __name__ == "__main__": import doctest doctest.testmod()
6
def __lowerCAmelCase ( a__ , a__ ) -> float: def get_matched_characters(a__ , a__ ) -> str: __a = [] __a = min(len(_stra ) , len(_stra ) ) // 2 for i, l in enumerate(_stra ): __a = int(max(0 , i - limit ) ) __a = int(min(i + limit + 1 , len(_stra ) ) ) if l in _stra[left:right]: matched.append(a__ ) __a = F"""{_stra[0:_stra.index(a__ )]} {_stra[_stra.index(a__ ) + 1:]}""" return "".join(a__ ) # matching characters __a = get_matched_characters(a__ , a__ ) __a = get_matched_characters(a__ , a__ ) __a = len(a__ ) # transposition __a = ( len([(ca, ca) for ca, ca in zip(a__ , a__ ) if ca != ca] ) // 2 ) if not match_count: __a = 0.0 else: __a = ( 1 / 3 * ( match_count / len(a__ ) + match_count / len(a__ ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters __a = 0 for ca, ca in zip(stra[:4] , stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
6
1
"""simple docstring""" import copy from collections import OrderedDict from typing import Dict, Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING _a : Optional[int] = logging.get_logger(__name__) _a : Dict = { 'facebook/detr-resnet-50': 'https://huggingface.co/facebook/detr-resnet-50/resolve/main/config.json', # See all DETR models at https://huggingface.co/models?filter=detr } class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Optional[Any] = "detr" _UpperCamelCase : Union[str, Any] = ["past_key_values"] _UpperCamelCase : Any = { "hidden_size": "d_model", "num_attention_heads": "encoder_attention_heads", } def __init__( self , a__=True , a__=None , a__=3 , a__=100 , a__=6 , a__=2048 , a__=8 , a__=6 , a__=2048 , a__=8 , a__=0.0 , a__=0.0 , a__=True , a__="relu" , a__=256 , a__=0.1 , a__=0.0 , a__=0.0 , a__=0.0_2 , a__=1.0 , a__=False , a__="sine" , a__="resnet50" , a__=True , a__=False , a__=1 , a__=5 , a__=2 , a__=1 , a__=1 , a__=5 , a__=2 , a__=0.1 , **a__ , ): if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) _lowerCAmelCase : Dict = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(a__ , a__ ): _lowerCAmelCase : str = backbone_config.get("""model_type""" ) _lowerCAmelCase : Any = CONFIG_MAPPING[backbone_model_type] _lowerCAmelCase : Union[str, Any] = config_class.from_dict(a__ ) # set timm attributes to None _lowerCAmelCase : Dict = None, None, None _lowerCAmelCase : Optional[int] = use_timm_backbone _lowerCAmelCase : Union[str, Any] = backbone_config _lowerCAmelCase : Tuple = num_channels _lowerCAmelCase : List[str] = num_queries _lowerCAmelCase : str = d_model _lowerCAmelCase : int = encoder_ffn_dim _lowerCAmelCase : Union[str, Any] = encoder_layers _lowerCAmelCase : List[Any] = encoder_attention_heads _lowerCAmelCase : Optional[Any] = decoder_ffn_dim _lowerCAmelCase : Tuple = decoder_layers _lowerCAmelCase : Dict = decoder_attention_heads _lowerCAmelCase : Optional[Any] = dropout _lowerCAmelCase : str = attention_dropout _lowerCAmelCase : Optional[int] = activation_dropout _lowerCAmelCase : Dict = activation_function _lowerCAmelCase : Any = init_std _lowerCAmelCase : Dict = init_xavier_std _lowerCAmelCase : int = encoder_layerdrop _lowerCAmelCase : int = decoder_layerdrop _lowerCAmelCase : Tuple = encoder_layers _lowerCAmelCase : List[Any] = auxiliary_loss _lowerCAmelCase : Optional[int] = position_embedding_type _lowerCAmelCase : List[str] = backbone _lowerCAmelCase : str = use_pretrained_backbone _lowerCAmelCase : Any = dilation # Hungarian matcher _lowerCAmelCase : List[str] = class_cost _lowerCAmelCase : Union[str, Any] = bbox_cost _lowerCAmelCase : List[str] = giou_cost # Loss coefficients _lowerCAmelCase : List[Any] = mask_loss_coefficient _lowerCAmelCase : Optional[int] = dice_loss_coefficient _lowerCAmelCase : int = bbox_loss_coefficient _lowerCAmelCase : Dict = giou_loss_coefficient _lowerCAmelCase : str = eos_coefficient super().__init__(is_encoder_decoder=a__ , **a__ ) @property def __A ( self ): return self.encoder_attention_heads @property def __A ( self ): return self.d_model @classmethod def __A ( cls , a__ , **a__ ): return cls(backbone_config=a__ , **a__ ) def __A ( self ): _lowerCAmelCase : Union[str, Any] = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: _lowerCAmelCase : Dict = self.backbone_config.to_dict() _lowerCAmelCase : Any = self.__class__.model_type return output class __A ( SCREAMING_SNAKE_CASE_ ): _UpperCamelCase : Optional[int] = version.parse("1.11" ) @property def __A ( self ): return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def __A ( self ): return 1e-5 @property def __A ( self ): return 12
367
"""simple docstring""" import math def SCREAMING_SNAKE_CASE ( ) -> None: _lowerCAmelCase : Any = input("""Enter message: """ ) _lowerCAmelCase : List[Any] = int(input(f"Enter key [2-{len(_lowerCamelCase ) - 1}]: " ) ) _lowerCAmelCase : Optional[Any] = input("""Encryption/Decryption [e/d]: """ ) if mode.lower().startswith("""e""" ): _lowerCAmelCase : Tuple = encrypt_message(_lowerCamelCase ,_lowerCamelCase ) elif mode.lower().startswith("""d""" ): _lowerCAmelCase : Dict = decrypt_message(_lowerCamelCase ,_lowerCamelCase ) # Append pipe symbol (vertical bar) to identify spaces at the end. print(f"Output:\n{text + '|'}" ) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : int ,_lowerCamelCase : str ) -> str: _lowerCAmelCase : Dict = [""""""] * key for col in range(_lowerCamelCase ): _lowerCAmelCase : List[str] = col while pointer < len(_lowerCamelCase ): cipher_text[col] += message[pointer] pointer += key return "".join(_lowerCamelCase ) def SCREAMING_SNAKE_CASE ( _lowerCamelCase : int ,_lowerCamelCase : str ) -> str: _lowerCAmelCase : str = math.ceil(len(_lowerCamelCase ) / key ) _lowerCAmelCase : Union[str, Any] = key _lowerCAmelCase : Any = (num_cols * num_rows) - len(_lowerCamelCase ) _lowerCAmelCase : Dict = [""""""] * num_cols _lowerCAmelCase : Optional[Any] = 0 _lowerCAmelCase : Dict = 0 for symbol in message: plain_text[col] += symbol col += 1 if ( (col == num_cols) or (col == num_cols - 1) and (row >= num_rows - num_shaded_boxes) ): _lowerCAmelCase : str = 0 row += 1 return "".join(_lowerCamelCase ) if __name__ == "__main__": import doctest doctest.testmod() main()
126
0
'''simple docstring''' import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class __A ( UpperCamelCase__ ): a__ : Union[str, Any] = ["""image_processor""", """tokenizer"""] a__ : List[str] = """ViltImageProcessor""" a__ : List[str] = ("""BertTokenizer""", """BertTokenizerFast""") def __init__(self : Tuple , __a : int=None , __a : Union[str, Any]=None , **__a : Dict ): UpperCAmelCase_ = None if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , __a , ) UpperCAmelCase_ = kwargs.pop("feature_extractor" ) UpperCAmelCase_ = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(__a , __a ) UpperCAmelCase_ = self.image_processor def __call__(self : List[str] , __a : List[Any] , __a : Union[TextInput, PreTokenizedInput, List[TextInput], List[PreTokenizedInput]] = None , __a : bool = True , __a : Union[bool, str, PaddingStrategy] = False , __a : Union[bool, str, TruncationStrategy] = None , __a : Optional[int] = None , __a : int = 0 , __a : Optional[int] = None , __a : Optional[bool] = None , __a : Optional[bool] = None , __a : bool = False , __a : bool = False , __a : bool = False , __a : bool = False , __a : bool = True , __a : Optional[Union[str, TensorType]] = None , **__a : List[str] , ): UpperCAmelCase_ = self.tokenizer( text=__a , add_special_tokens=__a , padding=__a , truncation=__a , max_length=__a , stride=__a , pad_to_multiple_of=__a , return_token_type_ids=__a , return_attention_mask=__a , return_overflowing_tokens=__a , return_special_tokens_mask=__a , return_offsets_mapping=__a , return_length=__a , verbose=__a , return_tensors=__a , **__a , ) # add pixel_values + pixel_mask UpperCAmelCase_ = self.image_processor(__a , return_tensors=__a ) encoding.update(__a ) return encoding def _lowercase (self : Optional[Any] , *__a : Union[str, Any] , **__a : List[Any] ): return self.tokenizer.batch_decode(*__a , **__a ) def _lowercase (self : Any , *__a : Tuple , **__a : str ): return self.tokenizer.decode(*__a , **__a ) @property def _lowercase (self : Union[str, Any] ): UpperCAmelCase_ = self.tokenizer.model_input_names UpperCAmelCase_ = self.image_processor.model_input_names return list(dict.fromkeys(tokenizer_input_names + image_processor_input_names ) ) @property def _lowercase (self : List[Any] ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , __a , ) return self.image_processor_class @property def _lowercase (self : Any ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , __a , ) return self.image_processor
1
'''simple docstring''' from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE_: Dict =logging.get_logger(__name__) SCREAMING_SNAKE_CASE_: Tuple ={} class __A ( UpperCamelCase__ ): a__ : int = """llama""" a__ : Any = ["""past_key_values"""] def __init__(self : List[str] , __a : List[str]=32000 , __a : Tuple=4096 , __a : List[Any]=11008 , __a : Dict=32 , __a : Tuple=32 , __a : Any=None , __a : Any="silu" , __a : List[Any]=2048 , __a : List[Any]=0.02 , __a : str=1E-6 , __a : Optional[Any]=True , __a : Union[str, Any]=0 , __a : Any=1 , __a : Dict=2 , __a : Dict=1 , __a : str=False , __a : str=None , **__a : Optional[Any] , ): UpperCAmelCase_ = vocab_size UpperCAmelCase_ = max_position_embeddings UpperCAmelCase_ = hidden_size UpperCAmelCase_ = intermediate_size UpperCAmelCase_ = num_hidden_layers UpperCAmelCase_ = num_attention_heads # for backward compatibility if num_key_value_heads is None: UpperCAmelCase_ = num_attention_heads UpperCAmelCase_ = num_key_value_heads UpperCAmelCase_ = hidden_act UpperCAmelCase_ = initializer_range UpperCAmelCase_ = rms_norm_eps UpperCAmelCase_ = pretraining_tp UpperCAmelCase_ = use_cache UpperCAmelCase_ = rope_scaling self._rope_scaling_validation() super().__init__( pad_token_id=__a , bos_token_id=__a , eos_token_id=__a , tie_word_embeddings=__a , **__a , ) def _lowercase (self : List[str] ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling , __a ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " f"""got {self.rope_scaling}""" ) UpperCAmelCase_ = self.rope_scaling.get("type" , __a ) UpperCAmelCase_ = self.rope_scaling.get("factor" , __a ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( f"""`rope_scaling`'s name field must be one of ['linear', 'dynamic'], got {rope_scaling_type}""" ) if rope_scaling_factor is None or not isinstance(__a , __a ) or rope_scaling_factor <= 1.0: raise ValueError(f"""`rope_scaling`'s factor field must be an float > 1, got {rope_scaling_factor}""" )
1
1
'''simple docstring''' from . import ( albert, align, altclip, audio_spectrogram_transformer, auto, autoformer, bark, bart, barthez, bartpho, beit, bert, bert_generation, bert_japanese, bertweet, big_bird, bigbird_pegasus, biogpt, bit, blenderbot, blenderbot_small, blip, blip_a, bloom, bridgetower, byta, camembert, canine, chinese_clip, clap, clip, clipseg, codegen, conditional_detr, convbert, convnext, convnextva, cpm, cpmant, ctrl, cvt, dataavec, deberta, deberta_va, decision_transformer, deformable_detr, deit, deprecated, deta, detr, dialogpt, dinat, distilbert, dit, donut, dpr, dpt, efficientformer, efficientnet, electra, encodec, encoder_decoder, ernie, ernie_m, esm, falcon, flaubert, flava, fnet, focalnet, fsmt, funnel, git, glpn, gpta, gpt_bigcode, gpt_neo, gpt_neox, gpt_neox_japanese, gpt_swa, gptj, gptsan_japanese, graphormer, groupvit, herbert, hubert, ibert, imagegpt, informer, instructblip, jukebox, layoutlm, layoutlmva, layoutlmva, layoutxlm, led, levit, lilt, llama, longformer, longta, luke, lxmert, mam_aaa, marian, markuplm, maskaformer, maskformer, mbart, mbartaa, mega, megatron_bert, megatron_gpta, mgp_str, mluke, mobilebert, mobilenet_va, mobilenet_va, mobilevit, mobilevitva, mpnet, mra, mta, musicgen, mvp, nat, nezha, nllb, nllb_moe, nystromformer, oneformer, open_llama, openai, opt, owlvit, pegasus, pegasus_x, perceiver, phobert, pixastruct, plbart, poolformer, prophetnet, qdqbert, rag, realm, reformer, regnet, rembert, resnet, roberta, roberta_prelayernorm, roc_bert, roformer, rwkv, sam, segformer, sew, sew_d, speech_encoder_decoder, speech_to_text, speech_to_text_a, speechta, splinter, squeezebert, swiftformer, swin, swinasr, swinva, switch_transformers, ta, table_transformer, tapas, time_series_transformer, timesformer, timm_backbone, transfo_xl, trocr, tvlt, umta, unispeech, unispeech_sat, upernet, videomae, vilt, vision_encoder_decoder, vision_text_dual_encoder, visual_bert, vit, vit_hybrid, vit_mae, vit_msn, vivit, wavaveca, wavaveca_conformer, wavaveca_phoneme, wavaveca_with_lm, wavlm, whisper, x_clip, xglm, xlm, xlm_prophetnet, xlm_roberta, xlm_roberta_xl, xlnet, xmod, yolos, yoso, )
365
'''simple docstring''' from random import randint from tempfile import TemporaryFile import numpy as np def lowerCAmelCase_ ( snake_case_ : int , snake_case_ : Any , snake_case_ : int ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase_ = 0 if start < end: UpperCAmelCase_ = randint(snake_case_ , snake_case_ ) UpperCAmelCase_ = a[end] UpperCAmelCase_ = a[pivot] UpperCAmelCase_ = temp UpperCAmelCase_ , UpperCAmelCase_ = _in_place_partition(snake_case_ , snake_case_ , snake_case_ ) count += _in_place_quick_sort(snake_case_ , snake_case_ , p - 1 ) count += _in_place_quick_sort(snake_case_ , p + 1 , snake_case_ ) return count def lowerCAmelCase_ ( snake_case_ : List[Any] , snake_case_ : Dict , snake_case_ : List[str] ) -> Optional[int]: '''simple docstring''' UpperCAmelCase_ = 0 UpperCAmelCase_ = randint(snake_case_ , snake_case_ ) UpperCAmelCase_ = a[end] UpperCAmelCase_ = a[pivot] UpperCAmelCase_ = temp UpperCAmelCase_ = start - 1 for index in range(snake_case_ , snake_case_ ): count += 1 if a[index] < a[end]: # check if current val is less than pivot value UpperCAmelCase_ = new_pivot_index + 1 UpperCAmelCase_ = a[new_pivot_index] UpperCAmelCase_ = a[index] UpperCAmelCase_ = temp UpperCAmelCase_ = a[new_pivot_index + 1] UpperCAmelCase_ = a[end] UpperCAmelCase_ = temp return new_pivot_index + 1, count SCREAMING_SNAKE_CASE_: List[str] =TemporaryFile() SCREAMING_SNAKE_CASE_: int =1_00 # 1000 elements are to be sorted SCREAMING_SNAKE_CASE_, SCREAMING_SNAKE_CASE_: str =0, 1 # mean and standard deviation SCREAMING_SNAKE_CASE_: List[str] =np.random.normal(mu, sigma, p) np.save(outfile, X) print('The array is') print(X) outfile.seek(0) # using the same array SCREAMING_SNAKE_CASE_: str =np.load(outfile) SCREAMING_SNAKE_CASE_: List[Any] =len(M) - 1 SCREAMING_SNAKE_CASE_: Dict =_in_place_quick_sort(M, 0, r) print( 'No of Comparisons for 100 elements selected from a standard normal distribution' 'is :' ) print(z)
106
0
import os import re from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = {'vocab_file': 'spiece.model'} UpperCAmelCase_ = { 'vocab_file': { 'google/bigbird-roberta-base': 'https://huggingface.co/google/bigbird-roberta-base/resolve/main/spiece.model', 'google/bigbird-roberta-large': ( 'https://huggingface.co/google/bigbird-roberta-large/resolve/main/spiece.model' ), 'google/bigbird-base-trivia-itc': ( 'https://huggingface.co/google/bigbird-base-trivia-itc/resolve/main/spiece.model' ), } } UpperCAmelCase_ = { 'google/bigbird-roberta-base': 4096, 'google/bigbird-roberta-large': 4096, 'google/bigbird-base-trivia-itc': 4096, } class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : Optional[Any] = VOCAB_FILES_NAMES a : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP a : Optional[int] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES a : int = ["input_ids", "attention_mask"] a : List[int] = [] def __init__( self, __magic_name__, __magic_name__="<unk>", __magic_name__="<s>", __magic_name__="</s>", __magic_name__="<pad>", __magic_name__="[SEP]", __magic_name__="[MASK]", __magic_name__="[CLS]", __magic_name__ = None, **__magic_name__, ) -> None: """simple docstring""" UpperCamelCase__ : int = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else bos_token UpperCamelCase__ : Tuple = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else eos_token UpperCamelCase__ : Tuple = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else unk_token UpperCamelCase__ : str = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else pad_token UpperCamelCase__ : List[Any] = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else cls_token UpperCamelCase__ : Tuple = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else sep_token # Mask token behave like a normal word, i.e. include the space before it UpperCamelCase__ : List[Any] = AddedToken(__magic_name__, lstrip=__magic_name__, rstrip=__magic_name__ ) if isinstance(__magic_name__, __magic_name__ ) else mask_token UpperCamelCase__ : int = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__magic_name__, eos_token=__magic_name__, unk_token=__magic_name__, pad_token=__magic_name__, sep_token=__magic_name__, mask_token=__magic_name__, cls_token=__magic_name__, sp_model_kwargs=self.sp_model_kwargs, **__magic_name__, ) UpperCamelCase__ : List[Any] = vocab_file UpperCamelCase__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__magic_name__ ) @property def UpperCamelCase__ ( self ) -> str: """simple docstring""" return self.sp_model.get_piece_size() def UpperCamelCase__ ( self ) -> Dict: """simple docstring""" UpperCamelCase__ : Union[str, Any] = {self.convert_ids_to_tokens(__magic_name__ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self ) -> Union[str, Any]: """simple docstring""" UpperCamelCase__ : str = self.__dict__.copy() UpperCamelCase__ : Dict = None return state def __setstate__( self, __magic_name__ ) -> List[str]: """simple docstring""" UpperCamelCase__ : str = d # for backward compatibility if not hasattr(self, '''sp_model_kwargs''' ): UpperCamelCase__ : Optional[int] = {} UpperCamelCase__ : Optional[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def UpperCamelCase__ ( self, __magic_name__ ) -> List[str]: """simple docstring""" return self.sp_model.encode(__magic_name__, out_type=__magic_name__ ) def UpperCamelCase__ ( self, __magic_name__ ) -> Optional[int]: """simple docstring""" return self.sp_model.piece_to_id(__magic_name__ ) def UpperCamelCase__ ( self, __magic_name__ ) -> str: """simple docstring""" UpperCamelCase__ : Dict = self.sp_model.IdToPiece(__magic_name__ ) return token def UpperCamelCase__ ( self, __magic_name__ ) -> List[str]: """simple docstring""" UpperCamelCase__ : Optional[int] = [] UpperCamelCase__ : Any = '''''' UpperCamelCase__ : int = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(__magic_name__ ) + token UpperCamelCase__ : Tuple = True UpperCamelCase__ : Union[str, Any] = [] else: current_sub_tokens.append(__magic_name__ ) UpperCamelCase__ : List[str] = False out_string += self.sp_model.decode(__magic_name__ ) return out_string.strip() def UpperCamelCase__ ( self, __magic_name__, __magic_name__ = False, __magic_name__ = None, __magic_name__ = True, **__magic_name__, ) -> str: """simple docstring""" UpperCamelCase__ : Any = kwargs.pop('''use_source_tokenizer''', __magic_name__ ) UpperCamelCase__ : str = self.convert_ids_to_tokens(__magic_name__, skip_special_tokens=__magic_name__ ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 UpperCamelCase__ : Optional[int] = [] UpperCamelCase__ : List[str] = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__magic_name__ ) ) UpperCamelCase__ : Dict = [] sub_texts.append(__magic_name__ ) else: current_sub_text.append(__magic_name__ ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(__magic_name__ ) ) # Mimic the behavior of the Rust tokenizer: # No space before [MASK] and [SEP] if spaces_between_special_tokens: UpperCamelCase__ : Optional[Any] = re.sub(R''' (\[(MASK|SEP)\])''', R'''\1''', ''' '''.join(__magic_name__ ) ) else: UpperCamelCase__ : Tuple = ''''''.join(__magic_name__ ) UpperCamelCase__ : Dict = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: UpperCamelCase__ : str = self.clean_up_tokenization(__magic_name__ ) return clean_text else: return text def UpperCamelCase__ ( self, __magic_name__, __magic_name__ = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__magic_name__ ): logger.error(f"Vocabulary path ({save_directory}) should be a directory" ) return UpperCamelCase__ : Dict = os.path.join( __magic_name__, (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__magic_name__ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file, __magic_name__ ) elif not os.path.isfile(self.vocab_file ): with open(__magic_name__, '''wb''' ) as fi: UpperCamelCase__ : Union[str, Any] = self.sp_model.serialized_model_proto() fi.write(__magic_name__ ) return (out_vocab_file,) def UpperCamelCase__ ( self, __magic_name__, __magic_name__ = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCamelCase__ : Dict = [self.cls_token_id] UpperCamelCase__ : Tuple = [self.sep_token_id] return cls + token_ids_a + sep + token_ids_a + sep def UpperCamelCase__ ( self, __magic_name__, __magic_name__ = None, __magic_name__ = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__magic_name__, token_ids_a=__magic_name__, already_has_special_tokens=__magic_name__ ) if token_ids_a is None: return [1] + ([0] * len(__magic_name__ )) + [1] return [1] + ([0] * len(__magic_name__ )) + [1] + ([0] * len(__magic_name__ )) + [1] def UpperCamelCase__ ( self, __magic_name__, __magic_name__ = None ) -> List[int]: """simple docstring""" UpperCamelCase__ : Optional[int] = [self.sep_token_id] UpperCamelCase__ : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1]
201
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCAmelCase_ = logging.get_logger(__name__) UpperCAmelCase_ = { 'google/mobilenet_v2_1.4_224': 'https://huggingface.co/google/mobilenet_v2_1.4_224/resolve/main/config.json', 'google/mobilenet_v2_1.0_224': 'https://huggingface.co/google/mobilenet_v2_1.0_224/resolve/main/config.json', 'google/mobilenet_v2_0.75_160': 'https://huggingface.co/google/mobilenet_v2_0.75_160/resolve/main/config.json', 'google/mobilenet_v2_0.35_96': 'https://huggingface.co/google/mobilenet_v2_0.35_96/resolve/main/config.json', # See all MobileNetV2 models at https://huggingface.co/models?filter=mobilenet_v2 } class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : List[str] = "mobilenet_v2" def __init__( self, __magic_name__=3, __magic_name__=224, __magic_name__=1.0, __magic_name__=8, __magic_name__=8, __magic_name__=6, __magic_name__=32, __magic_name__=True, __magic_name__=True, __magic_name__="relu6", __magic_name__=True, __magic_name__=0.8, __magic_name__=0.02, __magic_name__=0.001, __magic_name__=255, **__magic_name__, ) -> List[Any]: """simple docstring""" super().__init__(**__magic_name__ ) if depth_multiplier <= 0: raise ValueError('''depth_multiplier must be greater than zero.''' ) UpperCamelCase__ : Union[str, Any] = num_channels UpperCamelCase__ : int = image_size UpperCamelCase__ : int = depth_multiplier UpperCamelCase__ : Tuple = depth_divisible_by UpperCamelCase__ : List[str] = min_depth UpperCamelCase__ : Optional[int] = expand_ratio UpperCamelCase__ : Optional[int] = output_stride UpperCamelCase__ : Tuple = first_layer_is_expansion UpperCamelCase__ : Union[str, Any] = finegrained_output UpperCamelCase__ : str = hidden_act UpperCamelCase__ : Optional[Any] = tf_padding UpperCamelCase__ : Optional[int] = classifier_dropout_prob UpperCamelCase__ : int = initializer_range UpperCamelCase__ : Union[str, Any] = layer_norm_eps UpperCamelCase__ : Tuple = semantic_loss_ignore_index class lowercase__ ( __lowerCamelCase ): '''simple docstring''' a : Union[str, Any] = version.parse("1.11" ) @property def UpperCamelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict([('''pixel_values''', {0: '''batch'''})] ) @property def UpperCamelCase__ ( self ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "image-classification": return OrderedDict([('''logits''', {0: '''batch'''})] ) else: return OrderedDict([('''last_hidden_state''', {0: '''batch'''}), ('''pooler_output''', {0: '''batch'''})] ) @property def UpperCamelCase__ ( self ) -> float: """simple docstring""" return 1E-4
201
1
'''simple docstring''' import argparse import json from collections import OrderedDict from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import PoolFormerConfig, PoolFormerForImageClassification, PoolFormerImageProcessor from transformers.utils import logging logging.set_verbosity_info() lowerCamelCase_ = logging.get_logger(__name__) def __lowercase ( __lowercase , __lowercase , __lowercase , __lowercase ) -> Optional[Any]: '''simple docstring''' _A = original_name.split("." )[0] _A = key.split("." ) _A = int(key_list[key_list.index(__lowercase ) - 2] ) _A = int(key_list[key_list.index(__lowercase ) - 1] ) _A = orig_block_num - offset _A = key.replace(F'''{orig_block_num}.{layer_num}.{original_name}''' , F'''block.{new_block_num}.{layer_num}.{new_name}''' ) return key def __lowercase ( __lowercase ) -> str: '''simple docstring''' _A = OrderedDict() _A , _A = 0, 0 for key, value in state_dict.items(): if key.startswith("network" ): _A = key.replace("network" , "poolformer.encoder" ) if "proj" in key: # Works for the first embedding as well as the internal embedding layers if key.endswith("bias" ) and "patch_embed" not in key: patch_emb_offset += 1 _A = key[: key.find("proj" )] _A = key.replace(__lowercase , F'''patch_embeddings.{total_embed_found}.''' ) _A = key.replace("proj" , "projection" ) if key.endswith("bias" ): total_embed_found += 1 if "patch_embeddings" in key: _A = "poolformer.encoder." + key if "mlp.fc1" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "mlp.fc1" , "output.conv1" ) if "mlp.fc2" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "mlp.fc2" , "output.conv2" ) if "norm1" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "norm1" , "before_norm" ) if "norm2" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "norm2" , "after_norm" ) if "layer_scale_1" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "layer_scale_1" , "layer_scale_1" ) if "layer_scale_2" in key: _A = replace_key_with_offset(__lowercase , __lowercase , "layer_scale_2" , "layer_scale_2" ) if "head" in key: _A = key.replace("head" , "classifier" ) _A = value return new_state_dict def __lowercase ( ) -> List[str]: '''simple docstring''' _A = "http://images.cocodataset.org/val2017/000000039769.jpg" _A = Image.open(requests.get(__lowercase , stream=__lowercase ).raw ) return image @torch.no_grad() def __lowercase ( __lowercase , __lowercase , __lowercase ) -> Any: '''simple docstring''' _A = PoolFormerConfig() # set attributes based on model_name _A = "huggingface/label-files" _A = model_name[-3:] _A = 1000 _A = "imagenet-1k-id2label.json" _A = (1, 1000) # set config attributes _A = json.load(open(hf_hub_download(__lowercase , __lowercase , repo_type="dataset" ) , "r" ) ) _A = {int(__lowercase ): v for k, v in idalabel.items()} _A = idalabel _A = {v: k for k, v in idalabel.items()} if size == "s12": _A = [2, 2, 6, 2] _A = [64, 128, 320, 512] _A = 4.0 _A = 0.9 elif size == "s24": _A = [4, 4, 12, 4] _A = [64, 128, 320, 512] _A = 4.0 _A = 0.9 elif size == "s36": _A = [6, 6, 18, 6] _A = [64, 128, 320, 512] _A = 4.0 _A = 1e-6 _A = 0.9 elif size == "m36": _A = [6, 6, 18, 6] _A = [96, 192, 384, 768] _A = 4.0 _A = 1e-6 _A = 0.95 elif size == "m48": _A = [8, 8, 24, 8] _A = [96, 192, 384, 768] _A = 4.0 _A = 1e-6 _A = 0.95 else: raise ValueError(F'''Size {size} not supported''' ) # load image processor _A = PoolFormerImageProcessor(crop_pct=__lowercase ) # Prepare image _A = prepare_img() _A = image_processor(images=__lowercase , return_tensors="pt" ).pixel_values logger.info(F'''Converting model {model_name}...''' ) # load original state dict _A = torch.load(__lowercase , map_location=torch.device("cpu" ) ) # rename keys _A = rename_keys(__lowercase ) # create HuggingFace model and load state dict _A = PoolFormerForImageClassification(__lowercase ) model.load_state_dict(__lowercase ) model.eval() # Define image processor _A = PoolFormerImageProcessor(crop_pct=__lowercase ) _A = image_processor(images=prepare_img() , return_tensors="pt" ).pixel_values # forward pass _A = model(__lowercase ) _A = outputs.logits # define expected logit slices for different models if size == "s12": _A = torch.tensor([-0.3045, -0.6758, -0.4869] ) elif size == "s24": _A = torch.tensor([0.4402, -0.1374, -0.8045] ) elif size == "s36": _A = torch.tensor([-0.6080, -0.5133, -0.5898] ) elif size == "m36": _A = torch.tensor([0.3952, 0.2263, -1.2668] ) elif size == "m48": _A = torch.tensor([0.1167, -0.0656, -0.3423] ) else: raise ValueError(F'''Size {size} not supported''' ) # verify logits assert logits.shape == expected_shape assert torch.allclose(logits[0, :3] , __lowercase , atol=1e-2 ) # finally, save model and image processor logger.info(F'''Saving PyTorch model and image processor to {pytorch_dump_folder_path}...''' ) Path(__lowercase ).mkdir(exist_ok=__lowercase ) model.save_pretrained(__lowercase ) print(F'''Saving image processor to {pytorch_dump_folder_path}''' ) image_processor.save_pretrained(__lowercase ) if __name__ == "__main__": lowerCamelCase_ = argparse.ArgumentParser() parser.add_argument( '''--model_name''', default='''poolformer_s12''', type=str, help='''Name of the model you\'d like to convert.''', ) parser.add_argument( '''--checkpoint_path''', default=None, type=str, help='''Path to the original PyTorch checkpoint (.pth file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the folder to output PyTorch model.''' ) lowerCamelCase_ = parser.parse_args() convert_poolformer_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path)
174
'''simple docstring''' from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
174
1
'''simple docstring''' import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class lowercase__ ( unittest.TestCase ): '''simple docstring''' def UpperCAmelCase_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() @property def UpperCAmelCase_ ( self ): torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE : List[Any] = UNetaDModel( sample_size=(32, 64) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return model @property def UpperCAmelCase_ ( self ): torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE : Dict = UNetaDConditionModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""CrossAttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """CrossAttnUpBlock2D""") , cross_attention_dim=10 , ) return model @property def UpperCAmelCase_ ( self ): torch.manual_seed(0 ) _SCREAMING_SNAKE_CASE : List[Any] = AutoencoderKL( sample_size=(128, 64) , in_channels=1 , out_channels=1 , latent_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""DownEncoderBlock2D""", """DownEncoderBlock2D""") , up_block_types=("""UpDecoderBlock2D""", """UpDecoderBlock2D""") , ) _SCREAMING_SNAKE_CASE : List[str] = UNetaDModel( sample_size=(64, 32) , in_channels=1 , out_channels=1 , layers_per_block=2 , block_out_channels=(128, 128) , down_block_types=("""AttnDownBlock2D""", """DownBlock2D""") , up_block_types=("""UpBlock2D""", """AttnUpBlock2D""") , ) return vqvae, unet @slow def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : str = """cpu""" # ensure determinism for the device-dependent torch.Generator _SCREAMING_SNAKE_CASE : str = Mel( x_res=self.dummy_unet.config.sample_size[1] , y_res=self.dummy_unet.config.sample_size[0] , ) _SCREAMING_SNAKE_CASE : Any = DDPMScheduler() _SCREAMING_SNAKE_CASE : Optional[int] = AudioDiffusionPipeline(vqvae=__snake_case , unet=self.dummy_unet , mel=__snake_case , scheduler=__snake_case ) _SCREAMING_SNAKE_CASE : Optional[Any] = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) _SCREAMING_SNAKE_CASE : Union[str, Any] = torch.Generator(device=__snake_case ).manual_seed(42 ) _SCREAMING_SNAKE_CASE : List[Any] = pipe(generator=__snake_case , steps=4 ) _SCREAMING_SNAKE_CASE : int = output.audios[0] _SCREAMING_SNAKE_CASE : Optional[Any] = output.images[0] _SCREAMING_SNAKE_CASE : Optional[int] = torch.Generator(device=__snake_case ).manual_seed(42 ) _SCREAMING_SNAKE_CASE : int = pipe(generator=__snake_case , steps=4 , return_dict=__snake_case ) _SCREAMING_SNAKE_CASE : int = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) _SCREAMING_SNAKE_CASE : List[str] = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] _SCREAMING_SNAKE_CASE : Union[str, Any] = np.frombuffer(image_from_tuple.tobytes() , dtype="""uint8""" )[:10] _SCREAMING_SNAKE_CASE : List[str] = np.array([69, 255, 255, 255, 0, 0, 77, 181, 12, 127] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 _SCREAMING_SNAKE_CASE : str = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] , y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] , ) _SCREAMING_SNAKE_CASE : Optional[Any] = DDIMScheduler() _SCREAMING_SNAKE_CASE : str = self.dummy_vqvae_and_unet _SCREAMING_SNAKE_CASE : Dict = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=dummy_vqvae_and_unet[1] , mel=__snake_case , scheduler=__snake_case ) _SCREAMING_SNAKE_CASE : Union[str, Any] = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) np.random.seed(0 ) _SCREAMING_SNAKE_CASE : str = np.random.uniform(-1 , 1 , ((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) _SCREAMING_SNAKE_CASE : int = torch.Generator(device=__snake_case ).manual_seed(42 ) _SCREAMING_SNAKE_CASE : Optional[int] = pipe(raw_audio=__snake_case , generator=__snake_case , start_step=5 , steps=10 ) _SCREAMING_SNAKE_CASE : List[str] = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) _SCREAMING_SNAKE_CASE : List[Any] = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] _SCREAMING_SNAKE_CASE : List[Any] = np.array([120, 117, 110, 109, 138, 167, 138, 148, 132, 121] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 _SCREAMING_SNAKE_CASE : Union[str, Any] = self.dummy_unet_condition _SCREAMING_SNAKE_CASE : str = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] , unet=__snake_case , mel=__snake_case , scheduler=__snake_case ) _SCREAMING_SNAKE_CASE : Any = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) np.random.seed(0 ) _SCREAMING_SNAKE_CASE : Tuple = torch.rand((1, 1, 10) ) _SCREAMING_SNAKE_CASE : int = pipe(generator=__snake_case , encoding=__snake_case ) _SCREAMING_SNAKE_CASE : Union[str, Any] = output.images[0] _SCREAMING_SNAKE_CASE : int = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] _SCREAMING_SNAKE_CASE : Union[str, Any] = np.array([107, 103, 120, 127, 142, 122, 113, 122, 97, 111] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class lowercase__ ( unittest.TestCase ): '''simple docstring''' def UpperCAmelCase_ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self ): _SCREAMING_SNAKE_CASE : Optional[Any] = torch_device _SCREAMING_SNAKE_CASE : str = DiffusionPipeline.from_pretrained("""teticio/audio-diffusion-ddim-256""" ) _SCREAMING_SNAKE_CASE : Dict = pipe.to(__snake_case ) pipe.set_progress_bar_config(disable=__snake_case ) _SCREAMING_SNAKE_CASE : Union[str, Any] = torch.Generator(device=__snake_case ).manual_seed(42 ) _SCREAMING_SNAKE_CASE : Optional[Any] = pipe(generator=__snake_case ) _SCREAMING_SNAKE_CASE : Optional[Any] = output.audios[0] _SCREAMING_SNAKE_CASE : int = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] _SCREAMING_SNAKE_CASE : Tuple = np.frombuffer(image.tobytes() , dtype="""uint8""" )[:10] _SCREAMING_SNAKE_CASE : int = np.array([151, 167, 154, 144, 122, 134, 121, 105, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
200
import os import numpy import onnx def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = a.name snake_case_ = b.name snake_case_ = '' snake_case_ = '' snake_case_ = a == b snake_case_ = name_a snake_case_ = name_b return res def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for i, input_name in enumerate(node_proto.input ): if input_name == name: node_proto.input.insert(UpperCamelCase__ , UpperCamelCase__ ) node_proto.input.pop(i + 1 ) if node_proto.op_type == "If": _graph_replace_input_with(node_proto.attribute[0].g , UpperCamelCase__ , UpperCamelCase__ ) _graph_replace_input_with(node_proto.attribute[1].g , UpperCamelCase__ , UpperCamelCase__ ) if node_proto.op_type == "Loop": _graph_replace_input_with(node_proto.attribute[0].g , UpperCamelCase__ , UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' for n in graph_proto.node: _node_replace_input_with(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ): '''simple docstring''' snake_case_ = list(model.graph.initializer ) snake_case_ = list(model_without_ext.graph.initializer ) for i, ref_i in ind_to_replace: assert inits_with_data[i].name == inits[i].name assert inits_with_data[ref_i].name == inits[ref_i].name assert i > ref_i snake_case_ = inits[i].name snake_case_ = inits[ref_i].name model_without_ext.graph.initializer.remove(inits[i] ) # for n in model.graph.node: _graph_replace_input_with(model_without_ext.graph , UpperCamelCase__ , UpperCamelCase__ ) def __lowerCamelCase ( UpperCamelCase__ ): '''simple docstring''' snake_case_ = os.path.dirname(UpperCamelCase__ ) snake_case_ = os.path.basename(UpperCamelCase__ ) snake_case_ = onnx.load(os.path.join(UpperCamelCase__ , UpperCamelCase__ ) ) snake_case_ = list(model.graph.initializer ) snake_case_ = set() snake_case_ = {} snake_case_ = [] snake_case_ = 0 for i in range(len(UpperCamelCase__ ) ): if i in dup_set: continue for j in range(i + 1 , len(UpperCamelCase__ ) ): if j in dup_set: continue if _is_equal_tensor_proto(inits[i] , inits[j] ): dup_set.add(UpperCamelCase__ ) dup_set.add(UpperCamelCase__ ) snake_case_ = inits[j].data_type snake_case_ = numpy.prod(inits[j].dims ) if dtype == 1: mem_size *= 4 elif dtype == 6: mem_size *= 4 elif dtype == 7 or dtype == 11: mem_size *= 8 else: print('unexpected data type: ' , UpperCamelCase__ ) total_reduced_size += mem_size snake_case_ = inits[i].name snake_case_ = inits[j].name if name_i in dup_map: dup_map[name_i].append(UpperCamelCase__ ) else: snake_case_ = [name_j] ind_to_replace.append((j, i) ) print('total reduced size: ' , total_reduced_size / 1024 / 1024 / 1024 , 'GB' ) snake_case_ = sorted(UpperCamelCase__ ) _remove_dup_initializers_from_model(UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ ) snake_case_ = 'optimized_' + model_file_name snake_case_ = os.path.join(UpperCamelCase__ , UpperCamelCase__ ) onnx.save(UpperCamelCase__ , UpperCamelCase__ ) return new_model
285
0
"""simple docstring""" import inspect import os import unittest from dataclasses import dataclass import torch from accelerate import Accelerator, DistributedDataParallelKwargs, GradScalerKwargs from accelerate.state import AcceleratorState from accelerate.test_utils import execute_subprocess_async, require_cuda, require_multi_gpu from accelerate.utils import KwargsHandler @dataclass class UpperCAmelCase_ ( _a): '''simple docstring''' __UpperCamelCase : int = 0 __UpperCamelCase : bool = False __UpperCamelCase : float = 3.0 class UpperCAmelCase_ ( unittest.TestCase): '''simple docstring''' def _lowercase ( self ): """simple docstring""" self.assertDictEqual(MockClass().to_kwargs() , {} ) self.assertDictEqual(MockClass(a=2 ).to_kwargs() , {'''a''': 2} ) self.assertDictEqual(MockClass(a=2 , b=__SCREAMING_SNAKE_CASE ).to_kwargs() , {'''a''': 2, '''b''': True} ) self.assertDictEqual(MockClass(a=2 , c=2.25 ).to_kwargs() , {'''a''': 2, '''c''': 2.25} ) @require_cuda def _lowercase ( self ): """simple docstring""" UpperCamelCase : Dict = GradScalerKwargs(init_scale=1_024 , growth_factor=2 ) AcceleratorState._reset_state() UpperCamelCase : str = Accelerator(mixed_precision='''fp16''' , kwargs_handlers=[scaler_handler] ) print(accelerator.use_fpaa ) UpperCamelCase : Optional[Any] = accelerator.scaler # Check the kwargs have been applied self.assertEqual(scaler._init_scale , 1_024.0 ) self.assertEqual(scaler._growth_factor , 2.0 ) # Check the other values are at the default self.assertEqual(scaler._backoff_factor , 0.5 ) self.assertEqual(scaler._growth_interval , 2_000 ) self.assertEqual(scaler._enabled , __SCREAMING_SNAKE_CASE ) @require_multi_gpu def _lowercase ( self ): """simple docstring""" UpperCamelCase : Dict = ['''torchrun''', f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] execute_subprocess_async(__SCREAMING_SNAKE_CASE , env=os.environ.copy() ) if __name__ == "__main__": __UpperCAmelCase : Any = DistributedDataParallelKwargs(bucket_cap_mb=15, find_unused_parameters=True) __UpperCAmelCase : Any = Accelerator(kwargs_handlers=[ddp_scaler]) __UpperCAmelCase : str = torch.nn.Linear(100, 200) __UpperCAmelCase : Any = accelerator.prepare(model) # Check the values changed in kwargs __UpperCAmelCase : Optional[Any] = "" __UpperCAmelCase : Dict = model.bucket_bytes_cap // (1024 * 1024) if observed_bucket_cap_map != 15: error_msg += f"Kwargs badly passed, should have `15` but found {observed_bucket_cap_map}.\n" if model.find_unused_parameters is not True: error_msg += f"Kwargs badly passed, should have `True` but found {model.find_unused_parameters}.\n" # Check the values of the defaults if model.dim != 0: error_msg += f"Default value not respected, should have `0` but found {model.dim}.\n" if model.broadcast_buffers is not True: error_msg += f"Default value not respected, should have `True` but found {model.broadcast_buffers}.\n" if model.gradient_as_bucket_view is not False: error_msg += f"Default value not respected, should have `False` but found {model.gradient_as_bucket_view}.\n" # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
358
import math def a ( SCREAMING_SNAKE_CASE_ : int ): """simple docstring""" assert isinstance(SCREAMING_SNAKE_CASE_ , SCREAMING_SNAKE_CASE_ ) and ( number >= 0 ), "'number' must been an int and positive" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or not number % 2: # Negatives, 0, 1 and all even numbers are not primes return False UpperCamelCase : Union[str, Any] = range(3 , int(math.sqrt(SCREAMING_SNAKE_CASE_ ) + 1 ) , 2 ) return not any(not number % i for i in odd_numbers ) def a ( SCREAMING_SNAKE_CASE_ : List[Any] , SCREAMING_SNAKE_CASE_ : Tuple=1 , **SCREAMING_SNAKE_CASE_ : Tuple ): """simple docstring""" UpperCamelCase : Tuple = factor * value UpperCamelCase : Optional[int] = value while not is_prime(SCREAMING_SNAKE_CASE_ ): value += 1 if not ("desc" in kwargs and kwargs["desc"] is True) else -1 if value == first_value_val: return next_prime(value + 1 , **SCREAMING_SNAKE_CASE_ ) return value
315
0
from __future__ import annotations import json import requests from bsa import BeautifulSoup from fake_useragent import UserAgent SCREAMING_SNAKE_CASE__ : Tuple = {'UserAgent': UserAgent().random} def __magic_name__ ( __lowerCAmelCase : Optional[Any] ) -> dict: __lowerCamelCase = script.contents[0] __lowerCamelCase = json.loads(data[data.find('''{\"config\"''' ) : -1] ) return info["entry_data"]["ProfilePage"][0]["graphql"]["user"] class lowerCAmelCase__ : def __init__( self : List[str] , SCREAMING_SNAKE_CASE__ : List[Any] ) -> str: __lowerCamelCase = f'''https://www.instagram.com/{username}/''' __lowerCamelCase = self.get_json() def __A ( self : Optional[Any] ) -> dict: __lowerCamelCase = requests.get(self.url , headers=_UpperCAmelCase ).text __lowerCamelCase = BeautifulSoup(_UpperCAmelCase , '''html.parser''' ).find_all('''script''' ) try: return extract_user_profile(scripts[4] ) except (json.decoder.JSONDecodeError, KeyError): return extract_user_profile(scripts[3] ) def __repr__( self : str ) -> str: return f'''{self.__class__.__name__}(\'{self.username}\')''' def __str__( self : Optional[Any] ) -> str: return f'''{self.fullname} ({self.username}) is {self.biography}''' @property def __A ( self : Optional[Any] ) -> str: return self.user_data["username"] @property def __A ( self : Union[str, Any] ) -> str: return self.user_data["full_name"] @property def __A ( self : str ) -> str: return self.user_data["biography"] @property def __A ( self : str ) -> str: return self.user_data["business_email"] @property def __A ( self : List[str] ) -> str: return self.user_data["external_url"] @property def __A ( self : Tuple ) -> int: return self.user_data["edge_followed_by"]["count"] @property def __A ( self : Any ) -> int: return self.user_data["edge_follow"]["count"] @property def __A ( self : str ) -> int: return self.user_data["edge_owner_to_timeline_media"]["count"] @property def __A ( self : Optional[int] ) -> str: return self.user_data["profile_pic_url_hd"] @property def __A ( self : List[str] ) -> bool: return self.user_data["is_verified"] @property def __A ( self : Optional[int] ) -> bool: return self.user_data["is_private"] def __magic_name__ ( __lowerCAmelCase : str = "github" ) -> None: import os if os.environ.get('''CI''' ): return # test failing on GitHub Actions __lowerCamelCase = InstagramUser(__lowerCAmelCase ) assert instagram_user.user_data assert isinstance(instagram_user.user_data , __lowerCAmelCase ) assert instagram_user.username == username if username != "github": return assert instagram_user.fullname == "GitHub" assert instagram_user.biography == "Built for developers." assert instagram_user.number_of_posts > 150 assert instagram_user.number_of_followers > 12_0000 assert instagram_user.number_of_followings > 15 assert instagram_user.email == "support@github.com" assert instagram_user.website == "https://github.com/readme" assert instagram_user.profile_picture_url.startswith('''https://instagram.''' ) assert instagram_user.is_verified is True assert instagram_user.is_private is False if __name__ == "__main__": import doctest doctest.testmod() SCREAMING_SNAKE_CASE__ : Optional[int] = InstagramUser("github") print(instagram_user) print(F'{instagram_user.number_of_posts = }') print(F'{instagram_user.number_of_followers = }') print(F'{instagram_user.number_of_followings = }') print(F'{instagram_user.email = }') print(F'{instagram_user.website = }') print(F'{instagram_user.profile_picture_url = }') print(F'{instagram_user.is_verified = }') print(F'{instagram_user.is_private = }')
270
import numpy as np import pandas as pd from sklearn.preprocessing import Normalizer from sklearn.svm import SVR from statsmodels.tsa.statespace.sarimax import SARIMAX def UpperCamelCase ( __magic_name__ : list , __magic_name__ : list , __magic_name__ : list , __magic_name__ : list , __magic_name__ : list ) -> float: """simple docstring""" lowercase__ = np.array([[1, item, train_mtch[i]] for i, item in enumerate(__magic_name__ )] ) lowercase__ = np.array(__magic_name__ ) lowercase__ = np.dot(np.dot(np.linalg.inv(np.dot(x.transpose() , __magic_name__ ) ) , x.transpose() ) , __magic_name__ ) return abs(beta[0] + test_dt[0] * beta[1] + test_mtch[0] + beta[2] ) def UpperCamelCase ( __magic_name__ : list , __magic_name__ : list , __magic_name__ : list ) -> float: """simple docstring""" lowercase__ = (1, 2, 1) lowercase__ = (1, 1, 0, 7) lowercase__ = SARIMAX( __magic_name__ , exog=__magic_name__ , order=__magic_name__ , seasonal_order=__magic_name__ ) lowercase__ = model.fit(disp=__magic_name__ , maxiter=600 , method="""nm""" ) lowercase__ = model_fit.predict(1 , len(__magic_name__ ) , exog=[test_match] ) return result[0] def UpperCamelCase ( __magic_name__ : list , __magic_name__ : list , __magic_name__ : list ) -> float: """simple docstring""" lowercase__ = SVR(kernel="""rbf""" , C=1 , gamma=0.1 , epsilon=0.1 ) regressor.fit(__magic_name__ , __magic_name__ ) lowercase__ = regressor.predict(__magic_name__ ) return y_pred[0] def UpperCamelCase ( __magic_name__ : list ) -> float: """simple docstring""" train_user.sort() lowercase__ = np.percentile(__magic_name__ , 25 ) lowercase__ = np.percentile(__magic_name__ , 75 ) lowercase__ = qa - qa lowercase__ = qa - (iqr * 0.1) return low_lim def UpperCamelCase ( __magic_name__ : list , __magic_name__ : float ) -> bool: """simple docstring""" lowercase__ = 0 lowercase__ = 0 for i in list_vote: if i > actual_result: lowercase__ = not_safe + 1 else: if abs(abs(__magic_name__ ) - abs(__magic_name__ ) ) <= 0.1: safe += 1 else: not_safe += 1 return safe > not_safe if __name__ == "__main__": # data_input_df = pd.read_csv("ex_data.csv", header=None) A : Dict = [[1_8_2_3_1, 0.0, 1], [2_2_6_2_1, 1.0, 2], [1_5_6_7_5, 0.0, 3], [2_3_5_8_3, 1.0, 4]] A : str = pd.DataFrame( data_input, columns=['total_user', 'total_even', 'days'] ) A : Any = Normalizer().fit_transform(data_input_df.values) # split data A : Optional[int] = normalize_df[:, 2].tolist() A : Any = normalize_df[:, 0].tolist() A : str = normalize_df[:, 1].tolist() # for svr (input variable = total date and total match) A : int = normalize_df[:, [1, 2]].tolist() A : Any = x[: len(x) - 1] A : Tuple = x[len(x) - 1 :] # for linear regression & sarimax A : Optional[int] = total_date[: len(total_date) - 1] A : Optional[int] = total_user[: len(total_user) - 1] A : str = total_match[: len(total_match) - 1] A : Union[str, Any] = total_date[len(total_date) - 1 :] A : List[str] = total_user[len(total_user) - 1 :] A : str = total_match[len(total_match) - 1 :] # voting system with forecasting A : int = [ linear_regression_prediction( trn_date, trn_user, trn_match, tst_date, tst_match ), sarimax_predictor(trn_user, trn_match, tst_match), support_vector_regressor(x_train, x_test, trn_user), ] # check the safety of today's data A : int = '' if data_safety_checker(res_vote, tst_user) else 'not ' print('Today\'s data is {not_str}safe.')
305
0
import gc import random import unittest import numpy as np import torch from PIL import Image from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, PNDMScheduler, StableDiffusionInpaintPipeline, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, slow from ..pipeline_params import TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS from ..test_pipelines_common import PipelineKarrasSchedulerTesterMixin, PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class __magic_name__ ( snake_case , snake_case , snake_case , unittest.TestCase ): UpperCamelCase_ :List[str] = StableDiffusionInpaintPipeline UpperCamelCase_ :str = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS UpperCamelCase_ :str = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS UpperCamelCase_ :Union[str, Any] = frozenset( [] ) # TO-DO: update image_params once pipeline is refactored with VaeImageProcessor.preprocess UpperCamelCase_ :str = frozenset([] ) def UpperCAmelCase_ ( self )-> Any: torch.manual_seed(0 ) UpperCamelCase_ = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=9 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=32 , attention_head_dim=(2, 4) , use_linear_projection=_lowercase , ) UpperCamelCase_ = PNDMScheduler(skip_prk_steps=_lowercase ) torch.manual_seed(0 ) UpperCamelCase_ = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) UpperCamelCase_ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-0_5 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_000 , hidden_act="gelu" , projection_dim=512 , ) UpperCamelCase_ = CLIPTextModel(_lowercase ) UpperCamelCase_ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCamelCase_ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def UpperCAmelCase_ ( self , _lowercase , _lowercase=0 )-> List[Any]: # TODO: use tensor inputs instead of PIL, this is here just to leave the old expected_slices untouched UpperCamelCase_ = floats_tensor((1, 3, 32, 32) , rng=random.Random(_lowercase ) ).to(_lowercase ) UpperCamelCase_ = image.cpu().permute(0 , 2 , 3 , 1 )[0] UpperCamelCase_ = Image.fromarray(np.uinta(_lowercase ) ).convert("RGB" ).resize((64, 64) ) UpperCamelCase_ = Image.fromarray(np.uinta(image + 4 ) ).convert("RGB" ).resize((64, 64) ) if str(_lowercase ).startswith("mps" ): UpperCamelCase_ = torch.manual_seed(_lowercase ) else: UpperCamelCase_ = torch.Generator(device=_lowercase ).manual_seed(_lowercase ) UpperCamelCase_ = { "prompt": "A painting of a squirrel eating a burger", "image": init_image, "mask_image": mask_image, "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "numpy", } return inputs def UpperCAmelCase_ ( self )-> Dict: UpperCamelCase_ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCamelCase_ = self.get_dummy_components() UpperCamelCase_ = StableDiffusionInpaintPipeline(**_lowercase ) UpperCamelCase_ = sd_pipe.to(_lowercase ) sd_pipe.set_progress_bar_config(disable=_lowercase ) UpperCamelCase_ = self.get_dummy_inputs(_lowercase ) UpperCamelCase_ = sd_pipe(**_lowercase ).images UpperCamelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) UpperCamelCase_ = np.array([0.4_727, 0.5_735, 0.3_941, 0.5_446, 0.5_926, 0.4_394, 0.5_062, 0.4_654, 0.4_476] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def UpperCAmelCase_ ( self )-> Tuple: super().test_inference_batch_single_identical(expected_max_diff=3e-3 ) @slow @require_torch_gpu class __magic_name__ ( unittest.TestCase ): def UpperCAmelCase_ ( self )-> Any: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase_ ( self )-> Optional[Any]: UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench.npy" ) UpperCamelCase_ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase_ = StableDiffusionInpaintPipeline.from_pretrained(_lowercase , safety_checker=_lowercase ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() UpperCamelCase_ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="np" , ) UpperCamelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 9e-3 def UpperCAmelCase_ ( self )-> Union[str, Any]: UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase_ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint" "/yellow_cat_sitting_on_a_park_bench_fp16.npy" ) UpperCamelCase_ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase_ = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , torch_dtype=torch.floataa , safety_checker=_lowercase , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing() UpperCamelCase_ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , output_type="np" , ) UpperCamelCase_ = output.images[0] assert image.shape == (512, 512, 3) assert np.abs(expected_image - image ).max() < 5e-1 def UpperCAmelCase_ ( self )-> Tuple: torch.cuda.empty_cache() torch.cuda.reset_max_memory_allocated() torch.cuda.reset_peak_memory_stats() UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/sd2-inpaint/init_image.png" ) UpperCamelCase_ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/sd2-inpaint/mask.png" ) UpperCamelCase_ = "stabilityai/stable-diffusion-2-inpainting" UpperCamelCase_ = PNDMScheduler.from_pretrained(_lowercase , subfolder="scheduler" ) UpperCamelCase_ = StableDiffusionInpaintPipeline.from_pretrained( _lowercase , safety_checker=_lowercase , scheduler=_lowercase , torch_dtype=torch.floataa , ) pipe.to(_lowercase ) pipe.set_progress_bar_config(disable=_lowercase ) pipe.enable_attention_slicing(1 ) pipe.enable_sequential_cpu_offload() UpperCamelCase_ = "Face of a yellow cat, high resolution, sitting on a park bench" UpperCamelCase_ = torch.manual_seed(0 ) UpperCamelCase_ = pipe( prompt=_lowercase , image=_lowercase , mask_image=_lowercase , generator=_lowercase , num_inference_steps=2 , output_type="np" , ) UpperCamelCase_ = torch.cuda.max_memory_allocated() # make sure that less than 2.65 GB is allocated assert mem_bytes < 2.65 * 10**9
60
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...utils import logging SCREAMING_SNAKE_CASE :Union[str, Any] = logging.get_logger(__name__) SCREAMING_SNAKE_CASE :Union[str, Any] = { """BridgeTower/bridgetower-base""": """https://huggingface.co/BridgeTower/bridgetower-base/blob/main/config.json""", """BridgeTower/bridgetower-base-itm-mlm""": ( """https://huggingface.co/BridgeTower/bridgetower-base-itm-mlm/blob/main/config.json""" ), } class __magic_name__ ( snake_case ): UpperCamelCase_ :List[Any] = """bridgetower_vision_model""" def __init__( self , _lowercase=768 , _lowercase=12 , _lowercase=3 , _lowercase=16 , _lowercase=288 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=False , _lowercase=True , _lowercase=False , **_lowercase , )-> Optional[Any]: super().__init__(**_lowercase ) UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_channels UpperCamelCase_ = patch_size UpperCamelCase_ = image_size UpperCamelCase_ = initializer_factor UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = stop_gradient UpperCamelCase_ = share_layernorm UpperCamelCase_ = remove_last_layer @classmethod def UpperCAmelCase_ ( cls , _lowercase , **_lowercase )-> "PretrainedConfig": UpperCamelCase_ , UpperCamelCase_ = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get("model_type" ) == "bridgetower": UpperCamelCase_ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_lowercase , **_lowercase ) class __magic_name__ ( snake_case ): UpperCamelCase_ :Optional[int] = """bridgetower_text_model""" def __init__( self , _lowercase=50_265 , _lowercase=768 , _lowercase=12 , _lowercase=12 , _lowercase=1 , _lowercase=3_072 , _lowercase="gelu" , _lowercase=0.1 , _lowercase=0.1 , _lowercase=514 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=1 , _lowercase=0 , _lowercase=2 , _lowercase="absolute" , _lowercase=True , **_lowercase , )-> Optional[int]: super().__init__(**_lowercase ) UpperCamelCase_ = vocab_size UpperCamelCase_ = hidden_size UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = num_attention_heads UpperCamelCase_ = hidden_act UpperCamelCase_ = initializer_factor UpperCamelCase_ = intermediate_size UpperCamelCase_ = hidden_dropout_prob UpperCamelCase_ = attention_probs_dropout_prob UpperCamelCase_ = max_position_embeddings UpperCamelCase_ = type_vocab_size UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = position_embedding_type UpperCamelCase_ = use_cache UpperCamelCase_ = pad_token_id UpperCamelCase_ = bos_token_id UpperCamelCase_ = eos_token_id @classmethod def UpperCAmelCase_ ( cls , _lowercase , **_lowercase )-> "PretrainedConfig": UpperCamelCase_ , UpperCamelCase_ = cls.get_config_dict(_lowercase , **_lowercase ) if config_dict.get("model_type" ) == "bridgetower": UpperCamelCase_ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( F"You are using a model of type {config_dict['model_type']} to instantiate a model of type " F"{cls.model_type}. This is not supported for all configurations of models and can yield errors." ) return cls.from_dict(_lowercase , **_lowercase ) class __magic_name__ ( snake_case ): UpperCamelCase_ :List[Any] = """bridgetower""" def __init__( self , _lowercase=True , _lowercase="gelu" , _lowercase=768 , _lowercase=1 , _lowercase=1e-0_5 , _lowercase=False , _lowercase="add" , _lowercase=12 , _lowercase=6 , _lowercase=False , _lowercase=False , _lowercase=None , _lowercase=None , **_lowercase , )-> List[Any]: # TODO: remove this once the Hub files are updated. UpperCamelCase_ = kwargs.pop("text_config_dict" , _lowercase ) UpperCamelCase_ = kwargs.pop("vision_config_dict" , _lowercase ) super().__init__(**_lowercase ) UpperCamelCase_ = share_cross_modal_transformer_layers UpperCamelCase_ = hidden_act UpperCamelCase_ = hidden_size UpperCamelCase_ = initializer_factor UpperCamelCase_ = layer_norm_eps UpperCamelCase_ = share_link_tower_layers UpperCamelCase_ = link_tower_type UpperCamelCase_ = num_attention_heads UpperCamelCase_ = num_hidden_layers UpperCamelCase_ = tie_word_embeddings UpperCamelCase_ = init_layernorm_from_vision_encoder if text_config is None: UpperCamelCase_ = {} logger.info("`text_config` is `None`. Initializing the `BridgeTowerTextConfig` with default values." ) if vision_config is None: UpperCamelCase_ = {} logger.info("`vision_config` is `None`. Initializing the `BridgeTowerVisionConfig` with default values." ) UpperCamelCase_ = BridgeTowerTextConfig(**_lowercase ) UpperCamelCase_ = BridgeTowerVisionConfig(**_lowercase ) @classmethod def UpperCAmelCase_ ( cls , _lowercase , _lowercase , **_lowercase )-> List[str]: return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **_lowercase ) def UpperCAmelCase_ ( self )-> Union[str, Any]: UpperCamelCase_ = copy.deepcopy(self.__dict__ ) UpperCamelCase_ = self.text_config.to_dict() UpperCamelCase_ = self.vision_config.to_dict() UpperCamelCase_ = self.__class__.model_type return output
60
1
"""simple docstring""" _UpperCamelCase : List[str] = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' _UpperCamelCase : Any = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] _UpperCamelCase : Union[str, Any] = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
220
from typing import Any, Callable, Dict, List, Optional, Union import torch from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, LMSDiscreteScheduler, PNDMScheduler, StableDiffusionPipeline, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput from diffusers.pipelines.stable_diffusion.safety_checker import StableDiffusionSafetyChecker a__ : List[str] = '''CompVis/stable-diffusion-v1-1''' a__ : Optional[Any] = '''CompVis/stable-diffusion-v1-2''' a__ : Any = '''CompVis/stable-diffusion-v1-3''' a__ : Optional[Any] = '''CompVis/stable-diffusion-v1-4''' class a_ ( a__ ): """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase = True , ) ->str: super()._init_() SCREAMING_SNAKE_CASE : Tuple = StableDiffusionPipeline.from_pretrained(_lowerCamelCase ) SCREAMING_SNAKE_CASE : List[Any] = StableDiffusionPipeline.from_pretrained(_lowerCamelCase ) SCREAMING_SNAKE_CASE : str = StableDiffusionPipeline.from_pretrained(_lowerCamelCase ) SCREAMING_SNAKE_CASE : Dict = StableDiffusionPipeline( vae=_lowerCamelCase , text_encoder=_lowerCamelCase , tokenizer=_lowerCamelCase , unet=_lowerCamelCase , scheduler=_lowerCamelCase , safety_checker=_lowerCamelCase , feature_extractor=_lowerCamelCase , requires_safety_checker=_lowerCamelCase , ) self.register_modules(pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea , pipelinea=self.pipea ) @property def __lowerCAmelCase ( self ) ->Dict[str, Any]: return {k: getattr(self , _lowerCamelCase ) for k in self.config.keys() if not k.startswith('''_''' )} def __lowerCAmelCase ( self , _lowerCamelCase = "auto" ) ->Optional[int]: if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory SCREAMING_SNAKE_CASE : Optional[int] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(_lowerCamelCase ) def __lowerCAmelCase ( self ) ->List[str]: self.enable_attention_slicing(_lowerCamelCase ) @torch.no_grad() def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ) ->str: return self.pipea( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) @torch.no_grad() def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ) ->Tuple: return self.pipea( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) @torch.no_grad() def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ) ->Dict: return self.pipea( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) @torch.no_grad() def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ) ->Optional[Any]: return self.pipea( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) @torch.no_grad() def __lowerCAmelCase ( self , _lowerCamelCase , _lowerCamelCase = 512 , _lowerCamelCase = 512 , _lowerCamelCase = 50 , _lowerCamelCase = 7.5 , _lowerCamelCase = None , _lowerCamelCase = 1 , _lowerCamelCase = 0.0 , _lowerCamelCase = None , _lowerCamelCase = None , _lowerCamelCase = "pil" , _lowerCamelCase = True , _lowerCamelCase = None , _lowerCamelCase = 1 , **_lowerCamelCase , ) ->Dict: SCREAMING_SNAKE_CASE : Optional[Any] = '''cuda''' if torch.cuda.is_available() else '''cpu''' self.to(_lowerCamelCase ) # Checks if the height and width are divisible by 8 or not if height % 8 != 0 or width % 8 != 0: raise ValueError(F"""`height` and `width` must be divisible by 8 but are {height} and {width}.""" ) # Get first result from Stable Diffusion Checkpoint v1.1 SCREAMING_SNAKE_CASE : Optional[Any] = self.textaimg_sda_a( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) # Get first result from Stable Diffusion Checkpoint v1.2 SCREAMING_SNAKE_CASE : Any = self.textaimg_sda_a( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) # Get first result from Stable Diffusion Checkpoint v1.3 SCREAMING_SNAKE_CASE : Optional[int] = self.textaimg_sda_a( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) # Get first result from Stable Diffusion Checkpoint v1.4 SCREAMING_SNAKE_CASE : str = self.textaimg_sda_a( prompt=_lowerCamelCase , height=_lowerCamelCase , width=_lowerCamelCase , num_inference_steps=_lowerCamelCase , guidance_scale=_lowerCamelCase , negative_prompt=_lowerCamelCase , num_images_per_prompt=_lowerCamelCase , eta=_lowerCamelCase , generator=_lowerCamelCase , latents=_lowerCamelCase , output_type=_lowerCamelCase , return_dict=_lowerCamelCase , callback=_lowerCamelCase , callback_steps=_lowerCamelCase , **_lowerCamelCase , ) # Get all result images into a single list and pass it via StableDiffusionPipelineOutput for final result return StableDiffusionPipelineOutput([resa[0], resa[0], resa[0], resa[0]] )
313
0
"""simple docstring""" import os import sys from contextlib import contextmanager # Windows only if os.name == "nt": import ctypes import msvcrt # noqa class __a ( ctypes.Structure ): """simple docstring""" SCREAMING_SNAKE_CASE_ = [('size', ctypes.c_int), ('visible', ctypes.c_byte)] def _lowerCAmelCase ( ): '''simple docstring''' if os.name == "nt": UpperCamelCase__ : List[Any] =CursorInfo() UpperCamelCase__ : int =ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(UpperCAmelCase__ , ctypes.byref(UpperCAmelCase__ ) ) UpperCamelCase__ : List[str] =False ctypes.windll.kernelaa.SetConsoleCursorInfo(UpperCAmelCase__ , ctypes.byref(UpperCAmelCase__ ) ) elif os.name == "posix": sys.stdout.write('''\033[?25l''' ) sys.stdout.flush() def _lowerCAmelCase ( ): '''simple docstring''' if os.name == "nt": UpperCamelCase__ : int =CursorInfo() UpperCamelCase__ : Optional[Any] =ctypes.windll.kernelaa.GetStdHandle(-11 ) ctypes.windll.kernelaa.GetConsoleCursorInfo(UpperCAmelCase__ , ctypes.byref(UpperCAmelCase__ ) ) UpperCamelCase__ : Optional[int] =True ctypes.windll.kernelaa.SetConsoleCursorInfo(UpperCAmelCase__ , ctypes.byref(UpperCAmelCase__ ) ) elif os.name == "posix": sys.stdout.write('''\033[?25h''' ) sys.stdout.flush() @contextmanager def _lowerCAmelCase ( ): '''simple docstring''' try: hide_cursor() yield finally: show_cursor()
371
"""simple docstring""" import torch from diffusers import StableDiffusionPipeline _SCREAMING_SNAKE_CASE : Optional[int] = """path-to-your-trained-model""" _SCREAMING_SNAKE_CASE : Optional[Any] = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.floataa).to("""cuda""") _SCREAMING_SNAKE_CASE : Dict = """A photo of sks dog in a bucket""" _SCREAMING_SNAKE_CASE : Any = pipe(prompt, num_inference_steps=5_0, guidance_scale=7.5).images[0] image.save("""dog-bucket.png""")
157
0
import warnings from ...configuration_utils import PretrainedConfig from ...utils import logging lowercase_ = logging.get_logger(__name__) lowercase_ = { """RUCAIBox/mvp""": """https://huggingface.co/RUCAIBox/mvp/resolve/main/config.json""", } class __UpperCamelCase ( lowerCAmelCase__ ): """simple docstring""" lowerCAmelCase_ = '''mvp''' lowerCAmelCase_ = ['''past_key_values'''] lowerCAmelCase_ = {'''num_attention_heads''': '''encoder_attention_heads''', '''hidden_size''': '''d_model'''} def __init__( self : int , _A : Optional[int]=5_0267 , _A : Dict=1024 , _A : Tuple=12 , _A : List[str]=4096 , _A : str=16 , _A : List[Any]=12 , _A : Optional[int]=4096 , _A : Dict=16 , _A : List[str]=0.0 , _A : str=0.0 , _A : List[str]="gelu" , _A : Tuple=1024 , _A : Tuple=0.1 , _A : List[str]=0.0 , _A : Optional[int]=0.0 , _A : Optional[int]=0.02 , _A : List[str]=0.0 , _A : Tuple=False , _A : Optional[int]=True , _A : List[Any]=1 , _A : Dict=0 , _A : Union[str, Any]=2 , _A : str=True , _A : Optional[Any]=2 , _A : Optional[Any]=2 , _A : Union[str, Any]=False , _A : str=100 , _A : Dict=800 , **_A : Optional[int] , ): """simple docstring""" __SCREAMING_SNAKE_CASE : str = vocab_size __SCREAMING_SNAKE_CASE : Optional[Any] = max_position_embeddings __SCREAMING_SNAKE_CASE : Optional[int] = d_model __SCREAMING_SNAKE_CASE : Union[str, Any] = encoder_ffn_dim __SCREAMING_SNAKE_CASE : Optional[Any] = encoder_layers __SCREAMING_SNAKE_CASE : Tuple = encoder_attention_heads __SCREAMING_SNAKE_CASE : str = decoder_ffn_dim __SCREAMING_SNAKE_CASE : Union[str, Any] = decoder_layers __SCREAMING_SNAKE_CASE : str = decoder_attention_heads __SCREAMING_SNAKE_CASE : Optional[int] = dropout __SCREAMING_SNAKE_CASE : str = attention_dropout __SCREAMING_SNAKE_CASE : str = activation_dropout __SCREAMING_SNAKE_CASE : Any = activation_function __SCREAMING_SNAKE_CASE : Tuple = init_std __SCREAMING_SNAKE_CASE : Any = encoder_layerdrop __SCREAMING_SNAKE_CASE : Union[str, Any] = decoder_layerdrop __SCREAMING_SNAKE_CASE : Optional[int] = classifier_dropout __SCREAMING_SNAKE_CASE : Tuple = use_cache __SCREAMING_SNAKE_CASE : List[Any] = encoder_layers __SCREAMING_SNAKE_CASE : Optional[int] = scale_embedding # scale factor will be sqrt(d_model) if True __SCREAMING_SNAKE_CASE : Any = use_prompt __SCREAMING_SNAKE_CASE : Dict = prompt_length __SCREAMING_SNAKE_CASE : Dict = prompt_mid_dim super().__init__( pad_token_id=_A , bos_token_id=_A , eos_token_id=_A , is_encoder_decoder=_A , decoder_start_token_id=_A , forced_eos_token_id=_A , **_A , ) if self.forced_bos_token_id is None and kwargs.get('''force_bos_token_to_be_generated''' , _A ): __SCREAMING_SNAKE_CASE : Tuple = self.bos_token_id warnings.warn( F'''Please make sure the config includes `forced_bos_token_id={self.bos_token_id}` in future versions. ''' '''The config can simply be saved and uploaded again to be fixed.''' )
303
'''simple docstring''' from typing import List, Optional, Union import numpy as np from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase_ = logging.get_logger(__name__) class _a ( SCREAMING_SNAKE_CASE ): '''simple docstring''' A : str = ['''input_values''', '''padding_mask'''] def __init__( self, A = 1, A = 24_000, A = 0.0, A = None, A = None, **A, ): '''simple docstring''' super().__init__(feature_size=A, sampling_rate=A, padding_value=A, **A ) SCREAMING_SNAKE_CASE : Any = chunk_length_s SCREAMING_SNAKE_CASE : Dict = overlap @property def UpperCamelCase_ ( self ): '''simple docstring''' if self.chunk_length_s is None: return None else: return int(self.chunk_length_s * self.sampling_rate ) @property def UpperCamelCase_ ( self ): '''simple docstring''' if self.chunk_length_s is None or self.overlap is None: return None else: return max(1, int((1.0 - self.overlap) * self.chunk_length ) ) def __call__( self, A, A = None, A = False, A = None, A = None, A = None, ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F"The model corresponding to this feature extractor: {self} was trained using a sampling rate of" F" {self.sampling_rate}. Please make sure that the provided audio input was sampled with" F" {self.sampling_rate} and not {sampling_rate}." ) else: logger.warning( 'It is strongly recommended to pass the `sampling_rate` argument to this function. ' 'Failing to do so can result in silent errors that might be hard to debug.' ) if padding and truncation: raise ValueError('Both padding and truncation were set. Make sure you only set one.' ) elif padding is None: # by default let's pad the inputs SCREAMING_SNAKE_CASE : Tuple = True SCREAMING_SNAKE_CASE : Optional[int] = bool( isinstance(A, (list, tuple) ) and (isinstance(raw_audio[0], (np.ndarray, tuple, list) )) ) if is_batched: SCREAMING_SNAKE_CASE : Union[str, Any] = [np.asarray(A, dtype=np.floataa ).T for audio in raw_audio] elif not is_batched and not isinstance(A, np.ndarray ): SCREAMING_SNAKE_CASE : Optional[int] = np.asarray(A, dtype=np.floataa ) elif isinstance(A, np.ndarray ) and raw_audio.dtype is np.dtype(np.floataa ): SCREAMING_SNAKE_CASE : Optional[Any] = raw_audio.astype(np.floataa ) # always return batch if not is_batched: SCREAMING_SNAKE_CASE : str = [np.asarray(A ).T] # verify inputs are valid for idx, example in enumerate(A ): if example.ndim > 2: raise ValueError(F"Expected input shape (channels, length) but got shape {example.shape}" ) if self.feature_size == 1 and example.ndim != 1: raise ValueError(F"Expected mono audio but example has {example.shape[-1]} channels" ) if self.feature_size == 2 and example.shape[-1] != 2: raise ValueError(F"Expected stereo audio but example has {example.shape[-1]} channels" ) SCREAMING_SNAKE_CASE : List[str] = None SCREAMING_SNAKE_CASE : List[str] = BatchFeature({'input_values': raw_audio} ) if self.chunk_stride is not None and self.chunk_length is not None and max_length is None: if truncation: SCREAMING_SNAKE_CASE : Optional[int] = min(array.shape[0] for array in raw_audio ) SCREAMING_SNAKE_CASE : Tuple = int(np.floor(max_length / self.chunk_stride ) ) SCREAMING_SNAKE_CASE : int = (nb_step - 1) * self.chunk_stride + self.chunk_length elif padding: SCREAMING_SNAKE_CASE : str = max(array.shape[0] for array in raw_audio ) SCREAMING_SNAKE_CASE : Tuple = int(np.ceil(max_length / self.chunk_stride ) ) SCREAMING_SNAKE_CASE : Dict = (nb_step - 1) * self.chunk_stride + self.chunk_length SCREAMING_SNAKE_CASE : List[str] = 'max_length' else: SCREAMING_SNAKE_CASE : List[Any] = input_values # normal padding on batch if padded_inputs is None: SCREAMING_SNAKE_CASE : int = self.pad( A, max_length=A, truncation=A, padding=A, return_attention_mask=A, ) if padding: SCREAMING_SNAKE_CASE : Dict = padded_inputs.pop('attention_mask' ) SCREAMING_SNAKE_CASE : Optional[int] = [] for example in padded_inputs.pop('input_values' ): if self.feature_size == 1: SCREAMING_SNAKE_CASE : List[str] = example[..., None] input_values.append(example.T ) SCREAMING_SNAKE_CASE : Dict = input_values if return_tensors is not None: SCREAMING_SNAKE_CASE : int = padded_inputs.convert_to_tensors(A ) return padded_inputs
251
0
from ...utils import is_note_seq_available, is_transformers_available, is_torch_available from ...utils import OptionalDependencyNotAvailable try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .notes_encoder import SpectrogramNotesEncoder from .continous_encoder import SpectrogramContEncoder from .pipeline_spectrogram_diffusion import ( SpectrogramContEncoder, SpectrogramDiffusionPipeline, TaFilmDecoder, ) try: if not (is_transformers_available() and is_torch_available() and is_note_seq_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_transformers_and_torch_and_note_seq_objects import * # noqa F403 else: from .midi_utils import MidiProcessor
306
import collections import tempfile import unittest import numpy as np from transformers.testing_utils import ( is_pt_flax_cross_test, require_flax, require_torch, require_vision, slow, torch_device, ) from transformers.utils import is_flax_available, is_torch_available, is_vision_available from ...test_modeling_flax_common import floats_tensor, ids_tensor, random_attention_mask from ..bert.test_modeling_flax_bert import FlaxBertModelTester from ..clip.test_modeling_flax_clip import FlaxCLIPVisionModelTester from ..vit.test_modeling_flax_vit import FlaxViTModelTester if is_flax_available(): from transformers import ( FlaxBertModel, FlaxCLIPVisionModel, FlaxVisionTextDualEncoderModel, FlaxViTModel, VisionTextDualEncoderConfig, VisionTextDualEncoderProcessor, ) from transformers.modeling_flax_pytorch_utils import ( convert_pytorch_state_dict_to_flax, load_flax_weights_in_pytorch_model, ) if is_torch_available(): import torch from transformers import VisionTextDualEncoderModel if is_vision_available(): from PIL import Image def snake_case_ ( lowerCAmelCase_ : Tuple ): if isinstance(lowerCAmelCase_ , collections.abc.Iterable ): return x return (x, x) @require_flax class lowerCAmelCase : '''simple docstring''' def lowerCAmelCase ( self : Any , __a : Any , __a : List[Any] ) -> Optional[Any]: """simple docstring""" pass def lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" pass def lowerCAmelCase ( self : Union[str, Any] ) -> Tuple: """simple docstring""" pass def lowerCAmelCase ( self : Tuple , __a : np.ndarray , __a : np.ndarray , __a : float ) -> List[Any]: """simple docstring""" __lowercase : List[str] = np.abs((a - b) ).max() self.assertLessEqual(__a , __a , F"Difference between torch and flax is {diff} (>= {tol})." ) def lowerCAmelCase ( self : Tuple , __a : int , __a : str , __a : Union[str, Any] , __a : Optional[Any] , __a : Optional[Any]=None , **__a : Tuple ) -> Optional[Any]: """simple docstring""" __lowercase : str = VisionTextDualEncoderConfig.from_vision_text_configs(__a , __a ) __lowercase : str = FlaxVisionTextDualEncoderModel(__a ) __lowercase : Optional[Any] = model(input_ids=__a , pixel_values=__a , attention_mask=__a ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], config.projection_dim) ) def lowerCAmelCase ( self : Optional[int] , __a : Optional[int] , __a : Dict , __a : Dict , __a : List[str] , __a : Optional[Any]=None , **__a : str ) -> str: """simple docstring""" __lowercase , __lowercase : List[str] = self.get_vision_text_model(__a , __a ) __lowercase : Union[str, Any] = {"""vision_model""": vision_model, """text_model""": text_model} __lowercase : str = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__a ) __lowercase : Any = model(input_ids=__a , pixel_values=__a , attention_mask=__a ) self.assertEqual(output["""text_embeds"""].shape , (input_ids.shape[0], model.config.projection_dim) ) self.assertEqual(output["""image_embeds"""].shape , (pixel_values.shape[0], model.config.projection_dim) ) def lowerCAmelCase ( self : Tuple , __a : Union[str, Any] , __a : Union[str, Any] , __a : Union[str, Any] , __a : Dict , __a : int=None , **__a : int ) -> List[Any]: """simple docstring""" __lowercase , __lowercase : Tuple = self.get_vision_text_model(__a , __a ) __lowercase : Union[str, Any] = {"""vision_model""": vision_model, """text_model""": text_model} __lowercase : List[str] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__a ) __lowercase : List[Any] = model(input_ids=__a , pixel_values=__a , attention_mask=__a ) __lowercase : int = output[0] with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__a ) __lowercase : int = FlaxVisionTextDualEncoderModel.from_pretrained(__a ) __lowercase : Tuple = model(input_ids=__a , pixel_values=__a , attention_mask=__a ) __lowercase : int = after_output[0] __lowercase : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__a , 1E-3 ) def lowerCAmelCase ( self : List[Any] , __a : Any , __a : Tuple , __a : Optional[int] , __a : str , __a : Optional[Any]=None , **__a : Optional[Any] ) -> List[Any]: """simple docstring""" __lowercase , __lowercase : str = self.get_vision_text_model(__a , __a ) __lowercase : Optional[Any] = {"""vision_model""": vision_model, """text_model""": text_model} __lowercase : Dict = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained(**__a ) __lowercase : Union[str, Any] = model( input_ids=__a , pixel_values=__a , attention_mask=__a , output_attentions=__a ) __lowercase : Optional[int] = output.vision_model_output.attentions self.assertEqual(len(__a ) , vision_config.num_hidden_layers ) # in ViT, the seq_len equals the number of patches + 1 (we add 1 for the [CLS] token) __lowercase : Optional[int] = to_atuple(vision_model.config.image_size ) __lowercase : List[str] = to_atuple(vision_model.config.patch_size ) __lowercase : Optional[Any] = (image_size[1] // patch_size[1]) * (image_size[0] // patch_size[0]) __lowercase : int = num_patches + 1 self.assertEqual(vision_attentions[0].shape[-3:] , (vision_config.num_attention_heads, seq_len, seq_len) ) __lowercase : Dict = output.text_model_output.attentions self.assertEqual(len(__a ) , text_config.num_hidden_layers ) self.assertEqual( text_attentions[0].shape[-3:] , (text_config.num_attention_heads, input_ids.shape[-1], input_ids.shape[-1]) , ) def lowerCAmelCase ( self : Optional[int] , __a : List[str] , __a : List[Any] , __a : Optional[Any] ) -> Optional[int]: """simple docstring""" pt_model.to(__a ) pt_model.eval() # prepare inputs __lowercase : Union[str, Any] = inputs_dict __lowercase : List[Any] = {k: torch.tensor(v.tolist() ) for k, v in flax_inputs.items()} with torch.no_grad(): __lowercase : Union[str, Any] = pt_model(**__a ).to_tuple() __lowercase : Tuple = fx_model(**__a ).to_tuple() self.assertEqual(len(__a ) , len(__a ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output in zip(fx_outputs[:4] , pt_outputs[:4] ): self.assert_almost_equals(__a , pt_output.numpy() , 4E-2 ) # PT -> Flax with tempfile.TemporaryDirectory() as tmpdirname: pt_model.save_pretrained(__a ) __lowercase : Any = FlaxVisionTextDualEncoderModel.from_pretrained(__a , from_pt=__a ) __lowercase : Dict = fx_model_loaded(**__a ).to_tuple() self.assertEqual(len(__a ) , len(__a ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output_loaded, pt_output in zip(fx_outputs_loaded[:4] , pt_outputs[:4] ): self.assert_almost_equals(__a , pt_output.numpy() , 4E-2 ) # Flax -> PT with tempfile.TemporaryDirectory() as tmpdirname: fx_model.save_pretrained(__a ) __lowercase : str = VisionTextDualEncoderModel.from_pretrained(__a , from_flax=__a ) pt_model_loaded.to(__a ) pt_model_loaded.eval() with torch.no_grad(): __lowercase : List[Any] = pt_model_loaded(**__a ).to_tuple() self.assertEqual(len(__a ) , len(__a ) , """Output lengths differ between Flax and PyTorch""" ) for fx_output, pt_output_loaded in zip(fx_outputs[:4] , pt_outputs_loaded[:4] ): self.assert_almost_equals(__a , pt_output_loaded.numpy() , 4E-2 ) def lowerCAmelCase ( self : Optional[int] , __a : List[Any] , __a : int , __a : Optional[int] ) -> Optional[int]: """simple docstring""" __lowercase : Union[str, Any] = VisionTextDualEncoderConfig.from_vision_text_configs(__a , __a ) __lowercase : str = VisionTextDualEncoderModel(__a ) __lowercase : Union[str, Any] = FlaxVisionTextDualEncoderModel(__a ) __lowercase : List[str] = convert_pytorch_state_dict_to_flax(pt_model.state_dict() , __a ) __lowercase : Any = fx_state self.check_pt_flax_equivalence(__a , __a , __a ) def lowerCAmelCase ( self : Any , __a : Any , __a : Dict , __a : Tuple ) -> str: """simple docstring""" __lowercase : int = VisionTextDualEncoderConfig.from_vision_text_configs(__a , __a ) __lowercase : Union[str, Any] = VisionTextDualEncoderModel(__a ) __lowercase : Dict = FlaxVisionTextDualEncoderModel(__a ) __lowercase : Tuple = load_flax_weights_in_pytorch_model(__a , fx_model.params ) self.check_pt_flax_equivalence(__a , __a , __a ) def lowerCAmelCase ( self : str ) -> Optional[Any]: """simple docstring""" __lowercase : Optional[Any] = self.prepare_config_and_inputs() self.check_model_from_pretrained_configs(**__a ) def lowerCAmelCase ( self : Optional[Any] ) -> Optional[int]: """simple docstring""" __lowercase : int = self.prepare_config_and_inputs() self.check_vision_text_dual_encoder_from_pretrained(**__a ) def lowerCAmelCase ( self : List[Any] ) -> Dict: """simple docstring""" __lowercase : List[str] = self.prepare_config_and_inputs() self.check_save_load(**__a ) def lowerCAmelCase ( self : Any ) -> Dict: """simple docstring""" __lowercase : str = self.prepare_config_and_inputs() self.check_vision_text_output_attention(**__a ) @is_pt_flax_cross_test def lowerCAmelCase ( self : List[str] ) -> Tuple: """simple docstring""" __lowercase : Optional[Any] = self.prepare_config_and_inputs() __lowercase : Optional[int] = config_inputs_dict.pop("""vision_config""" ) __lowercase : Optional[int] = config_inputs_dict.pop("""text_config""" ) __lowercase : Dict = config_inputs_dict self.check_equivalence_pt_to_flax(__a , __a , __a ) self.check_equivalence_flax_to_pt(__a , __a , __a ) @slow def lowerCAmelCase ( self : Union[str, Any] ) -> str: """simple docstring""" __lowercase , __lowercase : List[Any] = self.get_pretrained_model_and_inputs() __lowercase : Dict = model_a(**__a ) __lowercase : Any = outputs[0] with tempfile.TemporaryDirectory() as tmp_dirname: model_a.save_pretrained(__a ) __lowercase : Tuple = FlaxVisionTextDualEncoderModel.from_pretrained(__a ) __lowercase : Optional[int] = model_a(**__a ) __lowercase : Tuple = after_outputs[0] __lowercase : Union[str, Any] = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(__a , 1E-5 ) @require_flax class lowerCAmelCase ( __a , unittest.TestCase ): '''simple docstring''' def lowerCAmelCase ( self : Dict ) -> Dict: """simple docstring""" __lowercase : int = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( """hf-internal-testing/tiny-random-vit""" , """hf-internal-testing/tiny-bert""" , vision_from_pt=__a , text_from_pt=__a , ) __lowercase : int = 13 __lowercase : Union[str, Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) __lowercase : Dict = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) __lowercase : Tuple = random_attention_mask([batch_size, 4] ) __lowercase : str = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def lowerCAmelCase ( self : Optional[Any] , __a : Union[str, Any] , __a : int ) -> Dict: """simple docstring""" __lowercase : int = FlaxViTModel(__a ) __lowercase : List[Any] = FlaxBertModel(__a ) return vision_model, text_model def lowerCAmelCase ( self : Tuple ) -> Optional[Any]: """simple docstring""" __lowercase : Tuple = FlaxViTModelTester(self ) __lowercase : str = FlaxBertModelTester(self ) __lowercase : List[str] = vit_model_tester.prepare_config_and_inputs() __lowercase : Union[str, Any] = bert_model_tester.prepare_config_and_inputs() __lowercase , __lowercase : Optional[int] = vision_config_and_inputs __lowercase , __lowercase , __lowercase , __lowercase : Any = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_torch class lowerCAmelCase ( __a , unittest.TestCase ): '''simple docstring''' def lowerCAmelCase ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" __lowercase : List[Any] = FlaxVisionTextDualEncoderModel.from_vision_text_pretrained( """hf-internal-testing/tiny-random-clip""" , """hf-internal-testing/tiny-bert""" , vision_from_pt=__a , text_from_pt=__a , ) __lowercase : Tuple = 13 __lowercase : Optional[Any] = floats_tensor( [ batch_size, model.config.vision_config.num_channels, model.config.vision_config.image_size, model.config.vision_config.image_size, ] ) __lowercase : Tuple = ids_tensor([batch_size, 4] , model.config.text_config.vocab_size ) __lowercase : List[Any] = random_attention_mask([batch_size, 4] ) __lowercase : int = {"""pixel_values""": pixel_values, """input_ids""": input_ids, """attention_mask""": attention_mask} return model, inputs def lowerCAmelCase ( self : str , __a : str , __a : Union[str, Any] ) -> Any: """simple docstring""" __lowercase : Dict = FlaxCLIPVisionModel(__a ) __lowercase : Optional[Any] = FlaxBertModel(__a ) return vision_model, text_model def lowerCAmelCase ( self : List[Any] ) -> List[str]: """simple docstring""" __lowercase : List[Any] = FlaxCLIPVisionModelTester(self ) __lowercase : Optional[Any] = FlaxBertModelTester(self ) __lowercase : Any = clip_model_tester.prepare_config_and_inputs() __lowercase : Optional[Any] = bert_model_tester.prepare_config_and_inputs() __lowercase , __lowercase : Dict = vision_config_and_inputs __lowercase , __lowercase , __lowercase , __lowercase : Optional[int] = text_config_and_inputs # make sure that cross attention layers are added return { "text_config": text_config, "vision_config": vision_config, "pixel_values": pixel_values, "attention_mask": attention_mask, "input_ids": input_ids, "token_type_ids": token_type_ids, } @require_flax @require_vision class lowerCAmelCase ( unittest.TestCase ): '''simple docstring''' @slow def lowerCAmelCase ( self : List[str] ) -> Optional[Any]: """simple docstring""" __lowercase : Any = FlaxVisionTextDualEncoderModel.from_pretrained("""clip-italian/clip-italian""" , logit_scale_init_value=1.0 ) __lowercase : int = VisionTextDualEncoderProcessor.from_pretrained("""clip-italian/clip-italian""" ) __lowercase : Union[str, Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) __lowercase : Tuple = processor( text=["""una foto di un gatto""", """una foto di un cane"""] , images=__a , padding=__a , return_tensors="""np""" ) __lowercase : Optional[int] = model(**__a ) # verify the logits self.assertEqual(outputs.logits_per_image.shape , (inputs.pixel_values.shape[0], inputs.input_ids.shape[0]) ) self.assertEqual( outputs.logits_per_text.shape , (inputs.input_ids.shape[0], inputs.pixel_values.shape[0]) , ) __lowercase : Optional[Any] = np.array([[1.2284727, 0.3104122]] ) self.assertTrue(np.allclose(outputs.logits_per_image , __a , atol=1E-3 ) )
306
1
import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor lowercase__ : Optional[int] = logging.get_logger(__name__) class lowercase_ ( UpperCamelCase_ ): """simple docstring""" def __init__( self , *__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) ->None: warnings.warn( '''The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.''' ''' Please use LayoutLMv2ImageProcessor instead.''' , __SCREAMING_SNAKE_CASE , ) super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE )
338
import os import re import shutil import sys import tempfile import unittest import black lowercase__ : List[str] = os.path.abspath(os.path.dirname(os.path.dirname(os.path.dirname(__file__)))) sys.path.append(os.path.join(git_repo_path, '''utils''')) import check_copies # noqa: E402 # This is the reference code that will be used in the tests. # If BertLMPredictionHead is changed in modeling_bert.py, this code needs to be manually updated. lowercase__ : Dict = ''' def __init__(self, config): super().__init__() self.transform = BertPredictionHeadTransform(config) # The output weights are the same as the input embeddings, but there is # an output-only bias for each token. self.decoder = nn.Linear(config.hidden_size, config.vocab_size, bias=False) self.bias = nn.Parameter(torch.zeros(config.vocab_size)) # Need a link between the two variables so that the bias is correctly resized with `resize_token_embeddings` self.decoder.bias = self.bias def forward(self, hidden_states): hidden_states = self.transform(hidden_states) hidden_states = self.decoder(hidden_states) return hidden_states ''' class lowercase_ ( unittest.TestCase ): """simple docstring""" def SCREAMING_SNAKE_CASE_ ( self ) ->Dict: lowerCAmelCase = tempfile.mkdtemp() os.makedirs(os.path.join(self.transformer_dir , '''models/bert/''' ) ) lowerCAmelCase = self.transformer_dir shutil.copy( os.path.join(__SCREAMING_SNAKE_CASE , '''src/transformers/models/bert/modeling_bert.py''' ) , os.path.join(self.transformer_dir , '''models/bert/modeling_bert.py''' ) , ) def SCREAMING_SNAKE_CASE_ ( self ) ->Optional[Any]: lowerCAmelCase = '''src/transformers''' shutil.rmtree(self.transformer_dir ) def SCREAMING_SNAKE_CASE_ ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None ) ->Union[str, Any]: lowerCAmelCase = comment + F"\nclass {class_name}(nn.Module):\n" + class_code if overwrite_result is not None: lowerCAmelCase = comment + F"\nclass {class_name}(nn.Module):\n" + overwrite_result lowerCAmelCase = black.Mode(target_versions={black.TargetVersion.PYaa} , line_length=119 ) lowerCAmelCase = black.format_str(__SCREAMING_SNAKE_CASE , mode=__SCREAMING_SNAKE_CASE ) lowerCAmelCase = os.path.join(self.transformer_dir , '''new_code.py''' ) with open(__SCREAMING_SNAKE_CASE , '''w''' , newline='''\n''' ) as f: f.write(__SCREAMING_SNAKE_CASE ) if overwrite_result is None: self.assertTrue(len(check_copies.is_copy_consistent(__SCREAMING_SNAKE_CASE ) ) == 0 ) else: check_copies.is_copy_consistent(f.name , overwrite=__SCREAMING_SNAKE_CASE ) with open(__SCREAMING_SNAKE_CASE , '''r''' ) as f: self.assertTrue(f.read() , __SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self ) ->int: lowerCAmelCase = check_copies.find_code_in_transformers('''models.bert.modeling_bert.BertLMPredictionHead''' ) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) def SCREAMING_SNAKE_CASE_ ( self ) ->Tuple: # Base copy consistency self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , REFERENCE_CODE + '''\n''' , ) # With no empty line at the end self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead''' , '''BertLMPredictionHead''' , __SCREAMING_SNAKE_CASE , ) # Copy consistency with rename self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , re.sub('''Bert''' , '''TestModel''' , __SCREAMING_SNAKE_CASE ) , ) # Copy consistency with a really long name lowerCAmelCase = '''TestModelWithAReallyLongNameBecauseSomePeopleLikeThatForSomeReason''' self.check_copy_consistency( F"# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->{long_class_name}" , F"{long_class_name}LMPredictionHead" , re.sub('''Bert''' , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) , ) # Copy consistency with overwrite self.check_copy_consistency( '''# Copied from transformers.models.bert.modeling_bert.BertLMPredictionHead with Bert->TestModel''' , '''TestModelLMPredictionHead''' , __SCREAMING_SNAKE_CASE , overwrite_result=re.sub('''Bert''' , '''TestModel''' , __SCREAMING_SNAKE_CASE ) , ) def SCREAMING_SNAKE_CASE_ ( self ) ->Tuple: lowerCAmelCase = check_copies.LOCALIZED_READMES['''README_zh-hans.md'''] lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (from HuggingFace),''' ''' released together with the paper [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) by Victor Sanh, Lysandre Debut and Thomas Wolf. The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)**''' ''' (from Google Research/Stanford University) released with the paper [ELECTRA: Pre-training text encoders''' ''' as discriminators rather than generators](https://arxiv.org/abs/2003.10555) by Kevin Clark, Minh-Thang''' ''' Luong, Quoc V. Le, Christopher D. Manning.''' ) lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n1.''' ''' **[DistilBERT](https://huggingface.co/transformers/model_doc/distilbert.html)** (来自 HuggingFace) 伴随论文''' ''' [DistilBERT, a distilled version of BERT: smaller, faster, cheaper and''' ''' lighter](https://arxiv.org/abs/1910.01108) 由 Victor Sanh, Lysandre Debut and Thomas Wolf 发布。 The same''' ''' method has been applied to compress GPT2 into''' ''' [DistilGPT2](https://github.com/huggingface/transformers/tree/main/examples/distillation), RoBERTa into''' ''' [DistilRoBERTa](https://github.com/huggingface/transformers/tree/main/examples/distillation),''' ''' Multilingual BERT into''' ''' [DistilmBERT](https://github.com/huggingface/transformers/tree/main/examples/distillation) and a German''' ''' version of DistilBERT.\n1. **[ELECTRA](https://huggingface.co/transformers/model_doc/electra.html)** (来自''' ''' Google Research/Stanford University) 伴随论文 [ELECTRA: Pre-training text encoders as discriminators rather''' ''' than generators](https://arxiv.org/abs/2003.10555) 由 Kevin Clark, Minh-Thang Luong, Quoc V. Le,''' ''' Christopher D. Manning 发布。\n''' ) lowerCAmelCase , lowerCAmelCase = check_copies.convert_to_localized_md( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , localized_readme['''format_model_list'''] ) self.assertFalse(__SCREAMING_SNAKE_CASE ) self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowerCAmelCase , lowerCAmelCase = check_copies.convert_to_localized_md( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , localized_readme['''format_model_list'''] ) # Check whether the number of models is equal to README.md after conversion. self.assertTrue(__SCREAMING_SNAKE_CASE ) lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (from Google Research and the''' ''' Toyota Technological Institute at Chicago) released with the paper [ALBERT: A Lite BERT for''' ''' Self-supervised Learning of Language Representations](https://arxiv.org/abs/1909.11942), by Zhenzhong''' ''' Lan, Mingda Chen, Sebastian Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut.''' ) lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/main/model_doc/albert.html)** (来自 Google Research and''' ''' the Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) lowerCAmelCase = ( '''1. **[ALBERT](https://huggingface.co/transformers/model_doc/albert.html)** (来自 Google Research and the''' ''' Toyota Technological Institute at Chicago) 伴随论文 [ALBERT: A Lite BERT for Self-supervised Learning of''' ''' Language Representations](https://arxiv.org/abs/1909.11942), 由 Zhenzhong Lan, Mingda Chen, Sebastian''' ''' Goodman, Kevin Gimpel, Piyush Sharma, Radu Soricut 发布。\n''' ) lowerCAmelCase , lowerCAmelCase = check_copies.convert_to_localized_md( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , localized_readme['''format_model_list'''] ) # Check if the model link is synchronized. self.assertEqual(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE )
338
1
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class lowerCAmelCase : def __init__( self :int , _lowercase :List[Any] , _lowercase :int=13 , _lowercase :List[Any]=30 , _lowercase :Tuple=2 , _lowercase :Optional[Any]=3 , _lowercase :str=True , _lowercase :Dict=True , _lowercase :int=32 , _lowercase :Tuple=5 , _lowercase :Dict=4 , _lowercase :List[Any]=37 , _lowercase :List[Any]="gelu" , _lowercase :List[Any]=0.1 , _lowercase :int=0.1 , _lowercase :str=10 , _lowercase :Optional[int]=0.02 , _lowercase :Any=3 , _lowercase :List[Any]=0.6 , _lowercase :List[str]=None , ): '''simple docstring''' lowercase__ = parent lowercase__ = batch_size lowercase__ = image_size lowercase__ = patch_size lowercase__ = num_channels lowercase__ = is_training lowercase__ = use_labels lowercase__ = hidden_size lowercase__ = num_hidden_layers lowercase__ = num_attention_heads lowercase__ = intermediate_size lowercase__ = hidden_act lowercase__ = hidden_dropout_prob lowercase__ = attention_probs_dropout_prob lowercase__ = type_sequence_label_size lowercase__ = initializer_range lowercase__ = mask_ratio lowercase__ = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) lowercase__ = (image_size // patch_size) ** 2 lowercase__ = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' lowercase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) lowercase__ = None if self.use_labels: lowercase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) lowercase__ = self.get_config() return config, pixel_values, labels def UpperCAmelCase ( self :Dict ): '''simple docstring''' return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=_lowercase , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :Optional[int] , _lowercase :int , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = ViTMAEModel(config=_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def UpperCAmelCase ( self :Dict , _lowercase :Dict , _lowercase :Union[str, Any] , _lowercase :Optional[int] ): '''simple docstring''' lowercase__ = ViTMAEForPreTraining(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = model(_lowercase ) lowercase__ = (self.image_size // self.patch_size) ** 2 lowercase__ = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images lowercase__ = 1 lowercase__ = ViTMAEForPreTraining(_lowercase ) model.to(_lowercase ) model.eval() lowercase__ = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) lowercase__ = model(_lowercase ) lowercase__ = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' lowercase__ = self.prepare_config_and_inputs() lowercase__ , lowercase__ , lowercase__ = config_and_inputs lowercase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class lowerCAmelCase ( lowercase_ , lowercase_ , unittest.TestCase ): __lowerCamelCase = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () __lowerCamelCase = {'feature-extraction': ViTMAEModel} if is_torch_available() else {} __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False __lowerCamelCase = False def UpperCAmelCase ( self :List[str] ): '''simple docstring''' lowercase__ = ViTMAEModelTester(self ) lowercase__ = ConfigTester(self , config_class=_lowercase , has_text_modality=_lowercase , hidden_size=37 ) def UpperCAmelCase ( self :int ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="ViTMAE does not use inputs_embeds" ) def UpperCAmelCase ( self :Optional[Any] ): '''simple docstring''' pass def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) lowercase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(_lowercase , nn.Linear ) ) def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) lowercase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic lowercase__ = [*signature.parameters.keys()] lowercase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , _lowercase ) def UpperCAmelCase ( self :Tuple ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*_lowercase ) def UpperCAmelCase ( self :Optional[Any] , _lowercase :Optional[int] , _lowercase :Optional[Any] , _lowercase :Any ): '''simple docstring''' np.random.seed(2 ) lowercase__ = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) lowercase__ = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) lowercase__ = torch.from_numpy(_lowercase ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument lowercase__ = pt_noise super().check_pt_tf_models(_lowercase , _lowercase , _lowercase ) def UpperCAmelCase ( self :int ): '''simple docstring''' lowercase__ , lowercase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: lowercase__ = model_class(_lowercase ) model.to(_lowercase ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowercase__ = model(**self._prepare_for_class(_lowercase , _lowercase ) ) lowercase__ = outputs[0].cpu().numpy() lowercase__ = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(_lowercase ) lowercase__ = model_class.from_pretrained(_lowercase ) model.to(_lowercase ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): lowercase__ = model(**self._prepare_for_class(_lowercase , _lowercase ) ) # Make sure we don't have nans lowercase__ = after_outputs[0].cpu().numpy() lowercase__ = 0 lowercase__ = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(_lowercase , 1e-5 ) @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' pass @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def UpperCAmelCase ( self :Any ): '''simple docstring''' pass @unittest.skip( reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results." ) def UpperCAmelCase ( self :Optional[int] ): '''simple docstring''' pass @unittest.skip(reason="ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load" ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' pass @unittest.skip("Will be fixed soon by reducing the size of the model used for common tests." ) def UpperCAmelCase ( self :Dict ): '''simple docstring''' pass @slow def UpperCAmelCase ( self :List[Any] ): '''simple docstring''' for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: lowercase__ = ViTMAEModel.from_pretrained(_lowercase ) self.assertIsNotNone(_lowercase ) def _A ( ): lowercase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class lowerCAmelCase ( unittest.TestCase ): @cached_property def UpperCAmelCase ( self :Any ): '''simple docstring''' return ViTImageProcessor.from_pretrained("facebook/vit-mae-base" ) if is_vision_available() else None @slow def UpperCAmelCase ( self :Union[str, Any] ): '''simple docstring''' np.random.seed(2 ) lowercase__ = ViTMAEForPreTraining.from_pretrained("facebook/vit-mae-base" ).to(_lowercase ) lowercase__ = self.default_image_processor lowercase__ = prepare_img() lowercase__ = image_processor(images=_lowercase , return_tensors="pt" ).to(_lowercase ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) lowercase__ = ViTMAEConfig() lowercase__ = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) lowercase__ = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): lowercase__ = model(**_lowercase , noise=torch.from_numpy(_lowercase ).to(device=_lowercase ) ) # verify the logits lowercase__ = torch.Size((1, 1_96, 7_68) ) self.assertEqual(outputs.logits.shape , _lowercase ) lowercase__ = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(_lowercase ) , atol=1e-4 ) )
201
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _snake_case = { """configuration_autoformer""": [ """AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """AutoformerConfig""", ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case = [ """AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST""", """AutoformerForPrediction""", """AutoformerModel""", """AutoformerPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_autoformer import ( AUTOFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, AutoformerConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_autoformer import ( AUTOFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, AutoformerForPrediction, AutoformerModel, AutoformerPreTrainedModel, ) else: import sys _snake_case = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
201
1
import pytest from datasets.utils.sharding import _distribute_shards, _number_of_shards_in_gen_kwargs, _split_gen_kwargs @pytest.mark.parametrize( """kwargs, expected""" , [ ({"""num_shards""": 0, """max_num_jobs""": 1}, []), ({"""num_shards""": 10, """max_num_jobs""": 1}, [range(10 )]), ({"""num_shards""": 10, """max_num_jobs""": 10}, [range(snake_case__ , i + 1 ) for i in range(10 )]), ({"""num_shards""": 1, """max_num_jobs""": 10}, [range(1 )]), ({"""num_shards""": 10, """max_num_jobs""": 3}, [range(0 , 4 ), range(4 , 7 ), range(7 , 10 )]), ({"""num_shards""": 3, """max_num_jobs""": 10}, [range(0 , 1 ), range(1 , 2 ), range(2 , 3 )]), ] , ) def _UpperCAmelCase ( snake_case , snake_case ): """simple docstring""" _lowerCAmelCase = _distribute_shards(**snake_case__ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, max_num_jobs, expected""" , [ ({"""foo""": 0}, 10, [{"""foo""": 0}]), ({"""shards""": [0, 1, 2, 3]}, 1, [{"""shards""": [0, 1, 2, 3]}]), ({"""shards""": [0, 1, 2, 3]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}, {"""shards""": [2]}, {"""shards""": [3]}]), ({"""shards""": [0, 1]}, 4, [{"""shards""": [0]}, {"""shards""": [1]}]), ({"""shards""": [0, 1, 2, 3]}, 2, [{"""shards""": [0, 1]}, {"""shards""": [2, 3]}]), ] , ) def _UpperCAmelCase ( snake_case , snake_case , snake_case ): """simple docstring""" _lowerCAmelCase = _split_gen_kwargs(snake_case__ , snake_case__ ) assert out == expected @pytest.mark.parametrize( """gen_kwargs, expected""" , [ ({"""foo""": 0}, 1), ({"""shards""": [0]}, 1), ({"""shards""": [0, 1, 2, 3]}, 4), ({"""shards""": [0, 1, 2, 3], """foo""": 0}, 4), ({"""shards""": [0, 1, 2, 3], """other""": (0, 1)}, 4), ({"""shards""": [0, 1, 2, 3], """shards2""": [0, 1]}, RuntimeError), ] , ) def _UpperCAmelCase ( snake_case , snake_case ): """simple docstring""" if expected is RuntimeError: with pytest.raises(snake_case__ ): _number_of_shards_in_gen_kwargs(snake_case__ ) else: _lowerCAmelCase = _number_of_shards_in_gen_kwargs(snake_case__ ) assert out == expected
82
'''simple docstring''' import torch from torch import nn from torch.nn import CrossEntropyLoss, MSELoss from transformers.file_utils import add_start_docstrings, add_start_docstrings_to_model_forward from transformers.models.bert.modeling_bert import ( BERT_INPUTS_DOCSTRING, BERT_START_DOCSTRING, BertEmbeddings, BertLayer, BertPooler, BertPreTrainedModel, ) def __lowerCAmelCase ( snake_case__ ): __UpperCamelCase : Tuple = torch.exp(snake_case__ ) __UpperCamelCase : str = torch.sum(snake_case__ , dim=1 ) # sum of exp(x_i) __UpperCamelCase : int = torch.sum(x * exp_x , dim=1 ) # sum of x_i * exp(x_i) return torch.log(snake_case__ ) - B / A class A ( nn.Module ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Union[str, Any]: super().__init__() __UpperCamelCase : Any = config.output_attentions __UpperCamelCase : Dict = config.output_hidden_states __UpperCamelCase : Union[str, Any] = nn.ModuleList([BertLayer(_UpperCAmelCase ) for _ in range(config.num_hidden_layers )] ) __UpperCamelCase : Tuple = nn.ModuleList([BertHighway(_UpperCAmelCase ) for _ in range(config.num_hidden_layers )] ) __UpperCamelCase : Optional[int] = [-1 for _ in range(config.num_hidden_layers )] def a_ (self , _UpperCAmelCase ) -> int: if (type(_UpperCAmelCase ) is float) or (type(_UpperCAmelCase ) is int): for i in range(len(self.early_exit_entropy ) ): __UpperCamelCase : str = x else: __UpperCamelCase : List[Any] = x def a_ (self , _UpperCAmelCase ) -> str: __UpperCamelCase : Tuple = pooler.state_dict() for highway in self.highway: for name, param in highway.pooler.state_dict().items(): param.copy_(loaded_model[name] ) def a_ (self , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> List[Any]: __UpperCamelCase : Optional[Any] = () __UpperCamelCase : Tuple = () __UpperCamelCase : Dict = () for i, layer_module in enumerate(self.layer ): if self.output_hidden_states: __UpperCamelCase : Tuple = all_hidden_states + (hidden_states,) __UpperCamelCase : Optional[int] = layer_module( _UpperCAmelCase , _UpperCAmelCase , head_mask[i] , _UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase : Tuple = layer_outputs[0] if self.output_attentions: __UpperCamelCase : Optional[Any] = all_attentions + (layer_outputs[1],) __UpperCamelCase : Any = (hidden_states,) if self.output_hidden_states: __UpperCamelCase : Any = current_outputs + (all_hidden_states,) if self.output_attentions: __UpperCamelCase : int = current_outputs + (all_attentions,) __UpperCamelCase : Optional[int] = self.highway[i](_UpperCAmelCase ) # logits, pooled_output if not self.training: __UpperCamelCase : Dict = highway_exit[0] __UpperCamelCase : Any = entropy(_UpperCAmelCase ) __UpperCamelCase : str = highway_exit + (highway_entropy,) # logits, hidden_states(?), entropy __UpperCamelCase : Optional[Any] = all_highway_exits + (highway_exit,) if highway_entropy < self.early_exit_entropy[i]: __UpperCamelCase : str = (highway_logits,) + current_outputs[1:] + (all_highway_exits,) raise HighwayException(_UpperCAmelCase , i + 1 ) else: __UpperCamelCase : Optional[int] = all_highway_exits + (highway_exit,) # Add last layer if self.output_hidden_states: __UpperCamelCase : int = all_hidden_states + (hidden_states,) __UpperCamelCase : Dict = (hidden_states,) if self.output_hidden_states: __UpperCamelCase : Union[str, Any] = outputs + (all_hidden_states,) if self.output_attentions: __UpperCamelCase : Optional[int] = outputs + (all_attentions,) __UpperCamelCase : List[Any] = outputs + (all_highway_exits,) return outputs # last-layer hidden state, (all hidden states), (all attentions), all highway exits @add_start_docstrings( "The Bert Model transformer with early exiting (DeeBERT). " , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Dict: super().__init__(_UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = config __UpperCamelCase : Dict = BertEmbeddings(_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = DeeBertEncoder(_UpperCAmelCase ) __UpperCamelCase : str = BertPooler(_UpperCAmelCase ) self.init_weights() def a_ (self ) -> Any: self.encoder.init_highway_pooler(self.pooler ) def a_ (self ) -> Optional[int]: return self.embeddings.word_embeddings def a_ (self , _UpperCAmelCase ) -> Dict: __UpperCamelCase : int = value def a_ (self , _UpperCAmelCase ) -> Tuple: for layer, heads in heads_to_prune.items(): self.encoder.layer[layer].attention.prune_heads(_UpperCAmelCase ) @add_start_docstrings_to_model_forward(_UpperCAmelCase ) def a_ (self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , ) -> Union[str, Any]: if input_ids is not None and inputs_embeds is not None: raise ValueError("You cannot specify both input_ids and inputs_embeds at the same time" ) elif input_ids is not None: __UpperCamelCase : Tuple = input_ids.size() elif inputs_embeds is not None: __UpperCamelCase : Optional[int] = inputs_embeds.size()[:-1] else: raise ValueError("You have to specify either input_ids or inputs_embeds" ) __UpperCamelCase : List[str] = input_ids.device if input_ids is not None else inputs_embeds.device if attention_mask is None: __UpperCamelCase : int = torch.ones(_UpperCAmelCase , device=_UpperCAmelCase ) if encoder_attention_mask is None: __UpperCamelCase : Tuple = torch.ones(_UpperCAmelCase , device=_UpperCAmelCase ) if token_type_ids is None: __UpperCamelCase : Optional[Any] = torch.zeros(_UpperCAmelCase , dtype=torch.long , device=_UpperCAmelCase ) # We can provide a self-attention mask of dimensions [batch_size, from_seq_length, to_seq_length] # ourselves in which case we just need to make it broadcastable to all heads. __UpperCamelCase : torch.Tensor = self.get_extended_attention_mask(_UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # If a 2D ou 3D attention mask is provided for the cross-attention # we need to make broadcastable to [batch_size, num_heads, seq_length, seq_length] if encoder_attention_mask.dim() == 3: __UpperCamelCase : Tuple = encoder_attention_mask[:, None, :, :] if encoder_attention_mask.dim() == 2: __UpperCamelCase : Any = encoder_attention_mask[:, None, None, :] __UpperCamelCase : List[Any] = encoder_extended_attention_mask.to( dtype=next(self.parameters() ).dtype ) # fp16 compatibility __UpperCamelCase : Dict = (1.0 - encoder_extended_attention_mask) * -10_000.0 # Prepare head mask if needed # 1.0 in head_mask indicate we keep the head # attention_probs has shape bsz x n_heads x N x N # input head_mask has shape [num_heads] or [num_hidden_layers x num_heads] # and head_mask is converted to shape [num_hidden_layers x batch x num_heads x seq_length x seq_length] __UpperCamelCase : Dict = self.get_head_mask(_UpperCAmelCase , self.config.num_hidden_layers ) __UpperCamelCase : Optional[int] = self.embeddings( input_ids=_UpperCAmelCase , position_ids=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , inputs_embeds=_UpperCAmelCase ) __UpperCamelCase : List[Any] = self.encoder( _UpperCAmelCase , attention_mask=_UpperCAmelCase , head_mask=_UpperCAmelCase , encoder_hidden_states=_UpperCAmelCase , encoder_attention_mask=_UpperCAmelCase , ) __UpperCamelCase : Union[str, Any] = encoder_outputs[0] __UpperCamelCase : Any = self.pooler(_UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = ( sequence_output, pooled_output, ) + encoder_outputs[ 1: ] # add hidden_states and attentions if they are here return outputs # sequence_output, pooled_output, (hidden_states), (attentions), highway exits class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase , _UpperCAmelCase ) -> Optional[Any]: __UpperCamelCase : Tuple = message __UpperCamelCase : Union[str, Any] = exit_layer # start from 1! class A ( nn.Module ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Dict: super().__init__() __UpperCamelCase : Union[str, Any] = BertPooler(_UpperCAmelCase ) __UpperCamelCase : int = nn.Dropout(config.hidden_dropout_prob ) __UpperCamelCase : Union[str, Any] = nn.Linear(config.hidden_size , config.num_labels ) def a_ (self , _UpperCAmelCase ) -> Any: # Pooler __UpperCamelCase : Optional[int] = encoder_outputs[0] __UpperCamelCase : str = self.pooler(_UpperCAmelCase ) # "return" pooler_output # BertModel __UpperCamelCase : Tuple = (pooler_input, pooler_output) + encoder_outputs[1:] # "return" bmodel_output # Dropout and classification __UpperCamelCase : Dict = bmodel_output[1] __UpperCamelCase : List[Any] = self.dropout(_UpperCAmelCase ) __UpperCamelCase : Any = self.classifier(_UpperCAmelCase ) return logits, pooled_output @add_start_docstrings( "Bert Model (with early exiting - DeeBERT) with a classifier on top,\n also takes care of multi-layer training. " , SCREAMING_SNAKE_CASE__ , ) class A ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__(self , _UpperCAmelCase ) -> Any: super().__init__(_UpperCAmelCase ) __UpperCamelCase : List[Any] = config.num_labels __UpperCamelCase : List[Any] = config.num_hidden_layers __UpperCamelCase : Optional[int] = DeeBertModel(_UpperCAmelCase ) __UpperCamelCase : List[str] = nn.Dropout(config.hidden_dropout_prob ) __UpperCamelCase : str = nn.Linear(config.hidden_size , self.config.num_labels ) self.init_weights() @add_start_docstrings_to_model_forward(_UpperCAmelCase ) def a_ (self , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=None , _UpperCAmelCase=-1 , _UpperCAmelCase=False , ) -> int: __UpperCamelCase : int = self.num_layers try: __UpperCamelCase : Tuple = self.bert( _UpperCAmelCase , attention_mask=_UpperCAmelCase , token_type_ids=_UpperCAmelCase , position_ids=_UpperCAmelCase , head_mask=_UpperCAmelCase , inputs_embeds=_UpperCAmelCase , ) # sequence_output, pooled_output, (hidden_states), (attentions), highway exits __UpperCamelCase : str = outputs[1] __UpperCamelCase : List[Any] = self.dropout(_UpperCAmelCase ) __UpperCamelCase : Dict = self.classifier(_UpperCAmelCase ) __UpperCamelCase : Tuple = (logits,) + outputs[2:] # add hidden states and attention if they are here except HighwayException as e: __UpperCamelCase : int = e.message __UpperCamelCase : Optional[Any] = e.exit_layer __UpperCamelCase : Optional[int] = outputs[0] if not self.training: __UpperCamelCase : Optional[int] = entropy(_UpperCAmelCase ) __UpperCamelCase : Optional[Any] = [] __UpperCamelCase : Any = [] if labels is not None: if self.num_labels == 1: # We are doing regression __UpperCamelCase : List[str] = MSELoss() __UpperCamelCase : Tuple = loss_fct(logits.view(-1 ) , labels.view(-1 ) ) else: __UpperCamelCase : Dict = CrossEntropyLoss() __UpperCamelCase : Any = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) # work with highway exits __UpperCamelCase : List[Any] = [] for highway_exit in outputs[-1]: __UpperCamelCase : Union[str, Any] = highway_exit[0] if not self.training: highway_logits_all.append(_UpperCAmelCase ) highway_entropy.append(highway_exit[2] ) if self.num_labels == 1: # We are doing regression __UpperCamelCase : Union[str, Any] = MSELoss() __UpperCamelCase : str = loss_fct(highway_logits.view(-1 ) , labels.view(-1 ) ) else: __UpperCamelCase : Optional[Any] = CrossEntropyLoss() __UpperCamelCase : List[str] = loss_fct(highway_logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) highway_losses.append(_UpperCAmelCase ) if train_highway: __UpperCamelCase : int = (sum(highway_losses[:-1] ),) + outputs # exclude the final highway, of course else: __UpperCamelCase : Dict = (loss,) + outputs if not self.training: __UpperCamelCase : Optional[int] = outputs + ((original_entropy, highway_entropy), exit_layer) if output_layer >= 0: __UpperCamelCase : int = ( (outputs[0],) + (highway_logits_all[output_layer],) + outputs[2:] ) # use the highway of the last layer return outputs # (loss), logits, (hidden_states), (attentions), (highway_exits)
298
0
"""simple docstring""" import os from pickle import UnpicklingError from typing import Dict, Tuple import jax import jax.numpy as jnp import numpy as np from flax.serialization import from_bytes from flax.traverse_util import flatten_dict, unflatten_dict import transformers from .utils import logging __UpperCamelCase : Any = logging.get_logger(__name__) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : int , _UpperCAmelCase : Tuple , _UpperCAmelCase : int , _UpperCAmelCase : str=False ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a PyTorch model in Flax, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise if not is_sharded: lowerCAmelCase = os.path.abspath(_UpperCAmelCase ) logger.info(F'Loading PyTorch weights from {pt_path}' ) lowerCAmelCase = torch.load(_UpperCAmelCase , map_location='cpu' ) logger.info(F'PyTorch checkpoint contains {sum(t.numel() for t in pt_state_dict.values() ):,} parameters.' ) lowerCAmelCase = convert_pytorch_state_dict_to_flax(_UpperCAmelCase , _UpperCAmelCase ) else: # model is sharded and pytorch_checkpoint_path already contains the list of .pt shard files lowerCAmelCase = convert_pytorch_sharded_state_dict_to_flax(_UpperCAmelCase , _UpperCAmelCase ) return flax_state_dict def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Tuple[str] , _UpperCAmelCase : np.ndarray , _UpperCAmelCase : Dict[str, jnp.ndarray] , _UpperCAmelCase : str , ): def is_key_or_prefix_key_in_dict(_UpperCAmelCase : Tuple[str] ) -> bool: return len(set(_UpperCAmelCase ) & {key, (model_prefix,) + key} ) > 0 # layer norm lowerCAmelCase = pt_tuple_key[:-1] + ('scale',) if pt_tuple_key[-1] in ["weight", "gamma"] and is_key_or_prefix_key_in_dict(_UpperCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer mean lowerCAmelCase = pt_tuple_key[:-1] + ('mean',) if pt_tuple_key[-1] == "running_mean" and not is_key_or_prefix_key_in_dict(_UpperCAmelCase ): return renamed_pt_tuple_key, pt_tensor # batch norm layer var lowerCAmelCase = pt_tuple_key[:-1] + ('var',) if pt_tuple_key[-1] == "running_var" and not is_key_or_prefix_key_in_dict(_UpperCAmelCase ): return renamed_pt_tuple_key, pt_tensor # embedding lowerCAmelCase = pt_tuple_key[:-1] + ('embedding',) if pt_tuple_key[-1] == "weight" and is_key_or_prefix_key_in_dict(_UpperCAmelCase ): return renamed_pt_tuple_key, pt_tensor # conv layer lowerCAmelCase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and pt_tensor.ndim == 4 and not is_key_or_prefix_key_in_dict(_UpperCAmelCase ): lowerCAmelCase = pt_tensor.transpose(2 , 3 , 1 , 0 ) return renamed_pt_tuple_key, pt_tensor # linear layer lowerCAmelCase = pt_tuple_key[:-1] + ('kernel',) if pt_tuple_key[-1] == "weight" and not is_key_or_prefix_key_in_dict(_UpperCAmelCase ): lowerCAmelCase = pt_tensor.T return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm weight lowerCAmelCase = pt_tuple_key[:-1] + ('weight',) if pt_tuple_key[-1] == "gamma": return renamed_pt_tuple_key, pt_tensor # old PyTorch layer norm bias lowerCAmelCase = pt_tuple_key[:-1] + ('bias',) if pt_tuple_key[-1] == "beta": return renamed_pt_tuple_key, pt_tensor # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 lowerCAmelCase = None if pt_tuple_key[-3::2] == ("parametrizations", "original0"): lowerCAmelCase = pt_tuple_key[-2] + '_g' elif pt_tuple_key[-3::2] == ("parametrizations", "original1"): lowerCAmelCase = pt_tuple_key[-2] + '_v' if name is not None: lowerCAmelCase = pt_tuple_key[:-3] + (name,) return renamed_pt_tuple_key, pt_tensor return pt_tuple_key, pt_tensor def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Dict , _UpperCAmelCase : List[Any] ): # convert pytorch tensor to numpy lowerCAmelCase = {k: v.numpy() for k, v in pt_state_dict.items()} lowerCAmelCase = flax_model.base_model_prefix # use params dict if the model contains batch norm layers if "params" in flax_model.params: lowerCAmelCase = flax_model.params['params'] else: lowerCAmelCase = flax_model.params lowerCAmelCase = flatten_dict(_UpperCAmelCase ) # add batch_stats keys,values to dict if "batch_stats" in flax_model.params: lowerCAmelCase = flatten_dict(flax_model.params['batch_stats'] ) random_flax_state_dict.update(_UpperCAmelCase ) lowerCAmelCase = {} lowerCAmelCase = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) lowerCAmelCase = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowerCAmelCase = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary lowerCAmelCase = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: lowerCAmelCase = pt_tuple_key[1:] # Correctly rename weight parameters lowerCAmelCase ,lowerCAmelCase = rename_key_and_reshape_tensor( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # add model prefix if necessary lowerCAmelCase = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: lowerCAmelCase = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1] or "var" in flax_key[-1]: lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) continue # also add unexpected weight so that warning is thrown lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) else: # also add unexpected weight so that warning is thrown lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Optional[int] , _UpperCAmelCase : Optional[int] ): import torch # Load the index lowerCAmelCase = {} for shard_file in shard_filenames: # load using msgpack utils lowerCAmelCase = torch.load(_UpperCAmelCase ) lowerCAmelCase = {k: v.numpy() for k, v in pt_state_dict.items()} lowerCAmelCase = flax_model.base_model_prefix # use params dict if the model contains batch norm layers and then add batch_stats keys,values to dict if "batch_stats" in flax_model.params: lowerCAmelCase = flax_model.params['params'] lowerCAmelCase = flatten_dict(_UpperCAmelCase ) random_flax_state_dict.update(flatten_dict(flax_model.params['batch_stats'] ) ) else: lowerCAmelCase = flax_model.params lowerCAmelCase = flatten_dict(_UpperCAmelCase ) lowerCAmelCase = (model_prefix not in flax_model_params) and ( model_prefix in {k.split('.' )[0] for k in pt_state_dict.keys()} ) lowerCAmelCase = (model_prefix in flax_model_params) and ( model_prefix not in {k.split('.' )[0] for k in pt_state_dict.keys()} ) # Need to change some parameters name to match Flax names for pt_key, pt_tensor in pt_state_dict.items(): lowerCAmelCase = tuple(pt_key.split('.' ) ) # remove base model prefix if necessary lowerCAmelCase = pt_tuple_key[0] == model_prefix if load_model_with_head_into_base_model and has_base_model_prefix: lowerCAmelCase = pt_tuple_key[1:] # Correctly rename weight parameters lowerCAmelCase ,lowerCAmelCase = rename_key_and_reshape_tensor( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) # add model prefix if necessary lowerCAmelCase = (model_prefix,) + flax_key in random_flax_state_dict if load_base_model_into_model_with_head and require_base_model_prefix: lowerCAmelCase = (model_prefix,) + flax_key if flax_key in random_flax_state_dict: if flax_tensor.shape != random_flax_state_dict[flax_key].shape: raise ValueError( F'PyTorch checkpoint seems to be incorrect. Weight {pt_key} was expected to be of shape ' F'{random_flax_state_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) # add batch stats if the model contains batchnorm layers if "batch_stats" in flax_model.params: if "mean" in flax_key[-1]: lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) continue if "var" in flax_key[-1]: lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) continue # remove num_batches_tracked key if "num_batches_tracked" in flax_key[-1]: flax_state_dict.pop(_UpperCAmelCase , _UpperCAmelCase ) continue # also add unexpected weight so that warning is thrown lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) else: # also add unexpected weight so that warning is thrown lowerCAmelCase = jnp.asarray(_UpperCAmelCase ) return unflatten_dict(_UpperCAmelCase ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Dict , _UpperCAmelCase : Optional[int] ): lowerCAmelCase = os.path.abspath(_UpperCAmelCase ) logger.info(F'Loading Flax weights from {flax_checkpoint_path}' ) # import correct flax class lowerCAmelCase = getattr(_UpperCAmelCase , 'Flax' + model.__class__.__name__ ) # load flax weight dict with open(_UpperCAmelCase , 'rb' ) as state_f: try: lowerCAmelCase = from_bytes(_UpperCAmelCase , state_f.read() ) except UnpicklingError: raise EnvironmentError(F'Unable to convert {flax_checkpoint_path} to Flax deserializable object. ' ) return load_flax_weights_in_pytorch_model(_UpperCAmelCase , _UpperCAmelCase ) def _SCREAMING_SNAKE_CASE (_UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] ): try: import torch # noqa: F401 except ImportError: logger.error( 'Loading a Flax weights in PyTorch, requires both PyTorch and Flax to be installed. Please see' ' https://pytorch.org/ and https://flax.readthedocs.io/en/latest/installation.html for installation' ' instructions.' ) raise # check if we have bf16 weights lowerCAmelCase = flatten_dict(jax.tree_util.tree_map(lambda _UpperCAmelCase : x.dtype == jnp.bfloataa , _UpperCAmelCase ) ).values() if any(_UpperCAmelCase ): # convert all weights to fp32 if the are bf16 since torch.from_numpy can-not handle bf16 # and bf16 is not fully supported in PT yet. logger.warning( 'Found ``bfloat16`` weights in Flax model. Casting all ``bfloat16`` weights to ``float32`` ' 'before loading those in PyTorch model.' ) lowerCAmelCase = jax.tree_util.tree_map( lambda _UpperCAmelCase : params.astype(np.floataa ) if params.dtype == jnp.bfloataa else params , _UpperCAmelCase ) lowerCAmelCase = flatten_dict(_UpperCAmelCase ) lowerCAmelCase = pt_model.state_dict() lowerCAmelCase = (pt_model.base_model_prefix in flax_state) and ( pt_model.base_model_prefix not in {k.split('.' )[0] for k in pt_model_dict.keys()} ) lowerCAmelCase = (pt_model.base_model_prefix not in flax_state) and ( pt_model.base_model_prefix in {k.split('.' )[0] for k in pt_model_dict.keys()} ) # keep track of unexpected & missing keys lowerCAmelCase = [] lowerCAmelCase = set(pt_model_dict.keys() ) for flax_key_tuple, flax_tensor in flax_state_dict.items(): lowerCAmelCase = flax_key_tuple[0] == pt_model.base_model_prefix lowerCAmelCase = '.'.join((pt_model.base_model_prefix,) + flax_key_tuple ) in pt_model_dict # adapt flax_key to prepare for loading from/to base model only if load_model_with_head_into_base_model and has_base_model_prefix: lowerCAmelCase = flax_key_tuple[1:] elif load_base_model_into_model_with_head and require_base_model_prefix: lowerCAmelCase = (pt_model.base_model_prefix,) + flax_key_tuple # rename flax weights to PyTorch format if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 4 and ".".join(_UpperCAmelCase ) not in pt_model_dict: # conv layer lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) lowerCAmelCase = jnp.transpose(_UpperCAmelCase , (3, 2, 0, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(_UpperCAmelCase ) not in pt_model_dict: # linear layer lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) lowerCAmelCase = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: lowerCAmelCase = flax_key_tuple[:-1] + ('weight',) # adding batch stats from flax batch norm to pt elif "mean" in flax_key_tuple[-1]: lowerCAmelCase = flax_key_tuple[:-1] + ('running_mean',) elif "var" in flax_key_tuple[-1]: lowerCAmelCase = flax_key_tuple[:-1] + ('running_var',) if "batch_stats" in flax_state: lowerCAmelCase = '.'.join(flax_key_tuple[1:] ) # Remove the params/batch_stats header else: lowerCAmelCase = '.'.join(_UpperCAmelCase ) # We also need to look at `pt_model_dict` and see if there are keys requiring further transformation. lowerCAmelCase = {} # New `weight_norm` from https://github.com/huggingface/transformers/pull/24030 for key in pt_model_dict: lowerCAmelCase = key.split('.' ) lowerCAmelCase = None if key_components[-3::2] == ["parametrizations", "original0"]: lowerCAmelCase = key_components[-2] + '_g' elif key_components[-3::2] == ["parametrizations", "original1"]: lowerCAmelCase = key_components[-2] + '_v' if name is not None: lowerCAmelCase = key_components[:-3] + [name] lowerCAmelCase = '.'.join(_UpperCAmelCase ) lowerCAmelCase = key if flax_key in special_pt_names: lowerCAmelCase = special_pt_names[flax_key] if flax_key in pt_model_dict: if flax_tensor.shape != pt_model_dict[flax_key].shape: raise ValueError( F'Flax checkpoint seems to be incorrect. Weight {flax_key_tuple} was expected ' F'to be of shape {pt_model_dict[flax_key].shape}, but is {flax_tensor.shape}.' ) else: # add weight to pytorch dict lowerCAmelCase = np.asarray(_UpperCAmelCase ) if not isinstance(_UpperCAmelCase , np.ndarray ) else flax_tensor lowerCAmelCase = torch.from_numpy(_UpperCAmelCase ) # remove from missing keys missing_keys.remove(_UpperCAmelCase ) else: # weight is not expected by PyTorch model unexpected_keys.append(_UpperCAmelCase ) pt_model.load_state_dict(_UpperCAmelCase ) # re-transform missing_keys to list lowerCAmelCase = list(_UpperCAmelCase ) if len(_UpperCAmelCase ) > 0: logger.warning( 'Some weights of the Flax model were not used when initializing the PyTorch model' F' {pt_model.__class__.__name__}: {unexpected_keys}\n- This IS expected if you are initializing' F' {pt_model.__class__.__name__} from a Flax model trained on another task or with another architecture' ' (e.g. initializing a BertForSequenceClassification model from a FlaxBertForPreTraining model).\n- This' F' IS NOT expected if you are initializing {pt_model.__class__.__name__} from a Flax model that you expect' ' to be exactly identical (e.g. initializing a BertForSequenceClassification model from a' ' FlaxBertForSequenceClassification model).' ) else: logger.warning(F'All Flax model weights were used when initializing {pt_model.__class__.__name__}.\n' ) if len(_UpperCAmelCase ) > 0: logger.warning( F'Some weights of {pt_model.__class__.__name__} were not initialized from the Flax model and are newly' F' initialized: {missing_keys}\nYou should probably TRAIN this model on a down-stream task to be able to' ' use it for predictions and inference.' ) else: logger.warning( F'All the weights of {pt_model.__class__.__name__} were initialized from the Flax model.\n' 'If your task is similar to the task the model of the checkpoint was trained on, ' F'you can already use {pt_model.__class__.__name__} for predictions without further training.' ) return pt_model
309
"""simple docstring""" import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () __UpperCamelCase : List[Any] = np.linspace(start=0, stop=75, num=75, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). __UpperCamelCase : str = [0, 25, 50] __UpperCamelCase : int = [25, 50, 75] __UpperCamelCase : str = fuzz.membership.trimf(X, abca) __UpperCamelCase : Tuple = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. __UpperCamelCase : Dict = np.ones(75) __UpperCamelCase : str = np.zeros((75,)) # 1. Union = max(µA(x), µB(x)) __UpperCamelCase : Optional[Any] = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) __UpperCamelCase : Dict = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) __UpperCamelCase : Dict = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) __UpperCamelCase : List[str] = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] __UpperCamelCase : List[str] = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) __UpperCamelCase : Tuple = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] __UpperCamelCase : Union[str, Any] = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] __UpperCamelCase : Dict = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('''Young''') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('''Middle aged''') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('''union''') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('''intersection''') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('''complement_a''') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('''difference a/b''') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('''alg_sum''') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('''alg_product''') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('''bdd_sum''') plt.grid(True) plt.subplot(4, 3, 10) plt.plot(X, bdd_difference) plt.title('''bdd_difference''') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
309
1
'''simple docstring''' import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: lowercase__ : Any = None lowercase__ : Dict = logging.get_logger(__name__) lowercase__ : List[Any] = '▁' lowercase__ : List[str] = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} lowercase__ : Optional[int] = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'}, 'tokenizer_file': { 'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json' }, } lowercase__ : int = { 'google/pegasus-xsum': 5_12, } class __lowerCAmelCase ( __magic_name__ ): """simple docstring""" _snake_case : List[str] = VOCAB_FILES_NAMES _snake_case : Tuple = PRETRAINED_VOCAB_FILES_MAP _snake_case : Tuple = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES _snake_case : List[Any] = PegasusTokenizer _snake_case : Any = ['input_ids', 'attention_mask'] def __init__( self : int , lowerCAmelCase__ : Optional[int]=None , lowerCAmelCase__ : Optional[Any]=None , lowerCAmelCase__ : Dict="<pad>" , lowerCAmelCase__ : int="</s>" , lowerCAmelCase__ : str="<unk>" , lowerCAmelCase__ : List[str]="<mask_2>" , lowerCAmelCase__ : Any="<mask_1>" , lowerCAmelCase__ : str=None , lowerCAmelCase__ : Any=103 , **lowerCAmelCase__ : Dict , ) -> Optional[int]: '''simple docstring''' _UpperCamelCase = offset if additional_special_tokens is not None: if not isinstance(lowerCAmelCase__ , lowerCAmelCase__ ): raise TypeError( f"""additional_special_tokens should be of type {type(lowerCAmelCase__ )}, but is""" f""" {type(lowerCAmelCase__ )}""" ) _UpperCamelCase = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f"""<unk_{i}>""" for i in range(len(lowerCAmelCase__ ) , self.offset - 1 ) ] if len(set(lowerCAmelCase__ ) ) != len(lowerCAmelCase__ ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f""" shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.""" ) _UpperCamelCase = additional_special_tokens_extended else: _UpperCamelCase = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f"""<unk_{i}>""" for i in range(2 , self.offset )] super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , eos_token=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , mask_token_sent=lowerCAmelCase__ , offset=lowerCAmelCase__ , additional_special_tokens=lowerCAmelCase__ , **lowerCAmelCase__ , ) _UpperCamelCase = vocab_file _UpperCamelCase = False if not self.vocab_file else True def snake_case__ ( self : Optional[int] , lowerCAmelCase__ : Dict ) -> List[str]: '''simple docstring''' _UpperCamelCase = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( '''There should be 3 special tokens: mask_token, pad_token, and eos_token +''' f""" {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}""" ) return [1 if x in all_special_ids else 0 for x in seq] def snake_case__ ( self : List[str] , lowerCAmelCase__ : List , lowerCAmelCase__ : Optional[List] = None , lowerCAmelCase__ : bool = False ) -> List[int]: '''simple docstring''' if already_has_special_tokens: return self._special_token_mask(lowerCAmelCase__ ) elif token_ids_a is None: return self._special_token_mask(lowerCAmelCase__ ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def snake_case__ ( self : Union[str, Any] , lowerCAmelCase__ : List[str] , lowerCAmelCase__ : Union[str, Any]=None ) -> List[int]: '''simple docstring''' if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def snake_case__ ( self : int , lowerCAmelCase__ : str , lowerCAmelCase__ : Optional[str] = None ) -> Tuple[str]: '''simple docstring''' if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(lowerCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _UpperCamelCase = os.path.join( lowerCAmelCase__ , (filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(lowerCAmelCase__ ): copyfile(self.vocab_file , lowerCAmelCase__ ) return (out_vocab_file,)
324
'''simple docstring''' import argparse import torch from transformers import BlenderbotConfig, BlenderbotForConditionalGeneration from transformers.utils import logging logging.set_verbosity_info() lowercase__ : Optional[int] = logging.get_logger(__name__) lowercase__ : str = [ ['attention', 'attn'], ['encoder_attention', 'encoder_attn'], ['q_lin', 'q_proj'], ['k_lin', 'k_proj'], ['v_lin', 'v_proj'], ['out_lin', 'out_proj'], ['norm_embeddings', 'layernorm_embedding'], ['position_embeddings', 'embed_positions'], ['embeddings', 'embed_tokens'], ['ffn.lin', 'fc'], ] def a__ ( lowercase : str ) -> Dict: """simple docstring""" if k == "embeddings.weight": return "shared.weight" for parlai_name, hf_name in PATTERNS: _UpperCamelCase = k.replace(lowercase, lowercase ) if k.startswith('''encoder''' ): _UpperCamelCase = k.replace('''.attn''', '''.self_attn''' ) _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''final_layer_norm''' ) elif k.startswith('''decoder''' ): _UpperCamelCase = k.replace('''norm1''', '''self_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm2''', '''encoder_attn_layer_norm''' ) _UpperCamelCase = k.replace('''norm3''', '''final_layer_norm''' ) return k def a__ ( lowercase : List[str] ) -> List[Any]: """simple docstring""" _UpperCamelCase = [ '''model.encoder.layernorm_embedding.weight''', '''model.encoder.layernorm_embedding.bias''', '''model.decoder.layernorm_embedding.weight''', '''model.decoder.layernorm_embedding.bias''', ] for k in keys: _UpperCamelCase = sd.pop(lowercase ) _UpperCamelCase = k.replace('''layernorm_embedding''', '''layer_norm''' ) assert new_k not in sd _UpperCamelCase = v lowercase__ : str = ['START'] @torch.no_grad() def a__ ( lowercase : Optional[int], lowercase : List[str], lowercase : List[str] ) -> Dict: """simple docstring""" _UpperCamelCase = torch.load(lowercase, map_location='''cpu''' ) _UpperCamelCase = model['''model'''] _UpperCamelCase = BlenderbotConfig.from_json_file(lowercase ) _UpperCamelCase = BlenderbotForConditionalGeneration(lowercase ) _UpperCamelCase = m.model.state_dict().keys() _UpperCamelCase = [] _UpperCamelCase = {} for k, v in sd.items(): if k in IGNORE_KEYS: continue _UpperCamelCase = rename_state_dict_key(lowercase ) if new_k not in valid_keys: failures.append([k, new_k] ) else: _UpperCamelCase = v if cfg.normalize_before: # Blenderbot-3B checkpoints. Rename layernorm_embedding -> layer_norm rename_layernorm_keys(lowercase ) m.model.load_state_dict(lowercase, strict=lowercase ) m.half() m.save_pretrained(lowercase ) if __name__ == "__main__": lowercase__ : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument('--src_path', type=str, help='like blenderbot-model.bin') parser.add_argument('--save_dir', default='hf_blenderbot', type=str, help='Where to save converted model.') parser.add_argument( '--hf_config_json', default='blenderbot-3b-config.json', type=str, help='Path to config to use' ) lowercase__ : Optional[Any] = parser.parse_args() convert_parlai_checkpoint(args.src_path, args.save_dir, args.hf_config_json)
324
1
"""simple docstring""" import numpy as np from scipy.spatial.distance import cdist from sklearn.metrics import fa_score import datasets __SCREAMING_SNAKE_CASE : Dict = '''\ @inproceedings{kakwani2020indicnlpsuite, title={{IndicNLPSuite: Monolingual Corpora, Evaluation Benchmarks and Pre-trained Multilingual Language Models for Indian Languages}}, author={Divyanshu Kakwani and Anoop Kunchukuttan and Satish Golla and Gokul N.C. and Avik Bhattacharyya and Mitesh M. Khapra and Pratyush Kumar}, year={2020}, booktitle={Findings of EMNLP}, } ''' __SCREAMING_SNAKE_CASE : str = '''\ IndicGLUE is a natural language understanding benchmark for Indian languages. It contains a wide variety of tasks and covers 11 major Indian languages - as, bn, gu, hi, kn, ml, mr, or, pa, ta, te. ''' __SCREAMING_SNAKE_CASE : List[Any] = ''' Compute IndicGLUE evaluation metric associated to each IndicGLUE dataset. Args: predictions: list of predictions to score (as int64), except for \'cvit-mkb-clsr\' where each prediction is a vector (of float32). references: list of ground truth labels corresponding to the predictions (as int64), except for \'cvit-mkb-clsr\' where each reference is a vector (of float32). Returns: depending on the IndicGLUE subset, one or several of: "accuracy": Accuracy "f1": F1 score "precision": Precision@10 Examples: >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wnli\') # \'wnli\' or any of ["copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md"] >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'wiki-ner\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> indic_glue_metric = datasets.load_metric(\'indic_glue\', \'cvit-mkb-clsr\') >>> references = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> predictions = [[0.5, 0.5, 0.5], [0.1, 0.2, 0.3]] >>> results = indic_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'precision@10\': 1.0} ''' def lowerCAmelCase_( lowercase_ : Dict , lowercase_ : Tuple ) -> List[str]: return float((preds == labels).mean() ) def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Union[str, Any] ) -> Dict: _lowerCamelCase = simple_accuracy(lowercase_ , lowercase_ ) _lowerCamelCase = float(fa_score(y_true=lowercase_ , y_pred=lowercase_ ) ) return { "accuracy": acc, "f1": fa, } def lowerCAmelCase_( lowercase_ : Optional[int] , lowercase_ : Any ) -> Optional[int]: _lowerCamelCase = np.array(lowercase_ ) _lowerCamelCase = np.array(lowercase_ ) _lowerCamelCase = en_sentvecs.shape[0] # mean centering _lowerCamelCase = en_sentvecs - np.mean(lowercase_ , axis=0 ) _lowerCamelCase = in_sentvecs - np.mean(lowercase_ , axis=0 ) _lowerCamelCase = cdist(lowercase_ , lowercase_ , '''cosine''' ) _lowerCamelCase = np.array(range(lowercase_ ) ) _lowerCamelCase = sim.argsort(axis=1 )[:, :10] _lowerCamelCase = np.any(preds == actual[:, None] , axis=1 ) return float(matches.mean() ) @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION, _KWARGS_DESCRIPTION ) class lowerCamelCase_( datasets.Metric ): '''simple docstring''' def snake_case__ ( self ): if self.config_name not in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", "wiki-ner", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''predictions''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), '''references''': datasets.Value('''int64''' ) if self.config_name != '''cvit-mkb-clsr''' else datasets.Sequence(datasets.Value('''float32''' ) ), } ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if self.config_name != '''cvit-mkb-clsr''' else None , ) def snake_case__ ( self , lowerCamelCase__ , lowerCamelCase__ ): if self.config_name == "cvit-mkb-clsr": return {"precision@10": precision_at_aa(lowerCamelCase__ , lowerCamelCase__ )} elif self.config_name in ["wiki-ner"]: return acc_and_fa(lowerCamelCase__ , lowerCamelCase__ ) elif self.config_name in [ "wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", "iitp-mr", "iitp-pr", "actsa-sc", "md", ]: return {"accuracy": simple_accuracy(lowerCamelCase__ , lowerCamelCase__ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["wnli", "copa", "sna", "csqa", "wstp", "inltkh", "bbca", ''' '''"cvit-mkb-clsr", "iitp-mr", "iitp-pr", "actsa-sc", "md", ''' '''"wiki-ner"]''' )
73
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging __SCREAMING_SNAKE_CASE : int = logging.get_logger(__name__) __SCREAMING_SNAKE_CASE : Dict = { '''alibaba-damo/mgp-str-base''': '''https://huggingface.co/alibaba-damo/mgp-str-base/resolve/main/config.json''', } class lowerCamelCase_( A__ ): '''simple docstring''' lowercase__ : List[str] = 'mgp-str' def __init__( self , lowerCamelCase__=[3_2, 1_2_8] , lowerCamelCase__=4 , lowerCamelCase__=3 , lowerCamelCase__=2_7 , lowerCamelCase__=3_8 , lowerCamelCase__=5_0_2_5_7 , lowerCamelCase__=3_0_5_2_2 , lowerCamelCase__=7_6_8 , lowerCamelCase__=1_2 , lowerCamelCase__=1_2 , lowerCamelCase__=4.0 , lowerCamelCase__=True , lowerCamelCase__=False , lowerCamelCase__=1e-5 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=0.0 , lowerCamelCase__=False , lowerCamelCase__=0.0_2 , **lowerCamelCase__ , ): super().__init__(**lowerCamelCase__ ) _lowerCamelCase = image_size _lowerCamelCase = patch_size _lowerCamelCase = num_channels _lowerCamelCase = max_token_length _lowerCamelCase = num_character_labels _lowerCamelCase = num_bpe_labels _lowerCamelCase = num_wordpiece_labels _lowerCamelCase = hidden_size _lowerCamelCase = num_hidden_layers _lowerCamelCase = num_attention_heads _lowerCamelCase = mlp_ratio _lowerCamelCase = distilled _lowerCamelCase = layer_norm_eps _lowerCamelCase = drop_rate _lowerCamelCase = qkv_bias _lowerCamelCase = attn_drop_rate _lowerCamelCase = drop_path_rate _lowerCamelCase = output_aa_attentions _lowerCamelCase = initializer_range
73
1
from collections import defaultdict from typing import Optional from ..image_utils import load_image from ..utils import ( add_end_docstrings, is_torch_available, logging, requires_backends, ) from .base import PIPELINE_INIT_ARGS, ChunkPipeline if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_MASK_GENERATION_MAPPING lowercase : Dict = logging.get_logger(__name__) @add_end_docstrings(_UpperCAmelCase) class lowerCamelCase__ ( _UpperCAmelCase): '''simple docstring''' def __init__( self :Optional[int] , **a :str ) -> Dict: super().__init__(**_UpperCAmelCase ) requires_backends(self , "vision" ) requires_backends(self , "torch" ) if self.framework != "pt": raise ValueError(f'The {self.__class__} is only available in PyTorch.' ) self.check_model_type(_UpperCAmelCase ) def _lowerCamelCase ( self :int , **a :Any ) -> Optional[Any]: __UpperCamelCase : List[Any] = {} __UpperCamelCase : List[Any] = {} __UpperCamelCase : Dict = {} # preprocess args if "points_per_batch" in kwargs: __UpperCamelCase : Dict = kwargs['''points_per_batch'''] if "points_per_crop" in kwargs: __UpperCamelCase : Any = kwargs['''points_per_crop'''] if "crops_n_layers" in kwargs: __UpperCamelCase : Union[str, Any] = kwargs['''crops_n_layers'''] if "crop_overlap_ratio" in kwargs: __UpperCamelCase : Optional[Any] = kwargs['''crop_overlap_ratio'''] if "crop_n_points_downscale_factor" in kwargs: __UpperCamelCase : Union[str, Any] = kwargs['''crop_n_points_downscale_factor'''] # postprocess args if "pred_iou_thresh" in kwargs: __UpperCamelCase : Any = kwargs['''pred_iou_thresh'''] if "stability_score_offset" in kwargs: __UpperCamelCase : Tuple = kwargs['''stability_score_offset'''] if "mask_threshold" in kwargs: __UpperCamelCase : List[str] = kwargs['''mask_threshold'''] if "stability_score_thresh" in kwargs: __UpperCamelCase : str = kwargs['''stability_score_thresh'''] if "crops_nms_thresh" in kwargs: __UpperCamelCase : List[str] = kwargs['''crops_nms_thresh'''] if "output_rle_mask" in kwargs: __UpperCamelCase : Dict = kwargs['''output_rle_mask'''] if "output_bboxes_mask" in kwargs: __UpperCamelCase : int = kwargs['''output_bboxes_mask'''] return preprocess_kwargs, forward_params, postprocess_kwargs def __call__( self :Optional[int] , a :Tuple , *a :str , a :int=None , a :List[Any]=None , **a :Union[str, Any] ) -> List[str]: return super().__call__(_UpperCAmelCase , *_UpperCAmelCase , num_workers=_UpperCAmelCase , batch_size=_UpperCAmelCase , **_UpperCAmelCase ) def _lowerCamelCase ( self :Any , a :Optional[int] , a :Union[str, Any]=6_4 , a :Dict = 0 , a :str = 5_1_2 / 1_5_0_0 , a :Optional[int] = 3_2 , a :int = 1 , ) -> Any: __UpperCamelCase : Union[str, Any] = load_image(_UpperCAmelCase ) __UpperCamelCase : Dict = self.image_processor.size['''longest_edge'''] __UpperCamelCase : Optional[Any] = self.image_processor.generate_crop_boxes( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase : List[Any] = self.image_processor(images=_UpperCAmelCase , return_tensors="pt" ) with self.device_placement(): if self.framework == "pt": __UpperCamelCase : Tuple = self.get_inference_context() with inference_context(): __UpperCamelCase : Optional[Any] = self._ensure_tensor_on_device(_UpperCAmelCase , device=self.device ) __UpperCamelCase : Any = self.model.get_image_embeddings(model_inputs.pop("pixel_values" ) ) __UpperCamelCase : Tuple = image_embeddings __UpperCamelCase : Optional[Any] = grid_points.shape[1] __UpperCamelCase : Tuple = points_per_batch if points_per_batch is not None else n_points if points_per_batch <= 0: raise ValueError( "Cannot have points_per_batch<=0. Must be >=1 to returned batched outputs. " "To return all points at once, set points_per_batch to None" ) for i in range(0 , _UpperCAmelCase , _UpperCAmelCase ): __UpperCamelCase : Dict = grid_points[:, i : i + points_per_batch, :, :] __UpperCamelCase : int = input_labels[:, i : i + points_per_batch] __UpperCamelCase : Any = i == n_points - points_per_batch yield { "input_points": batched_points, "input_labels": labels, "input_boxes": crop_boxes, "is_last": is_last, **model_inputs, } def _lowerCamelCase ( self :Optional[Any] , a :List[str] , a :Optional[Any]=0.88 , a :str=0.95 , a :Union[str, Any]=0 , a :Dict=1 , ) -> List[Any]: __UpperCamelCase : List[Any] = model_inputs.pop("input_boxes" ) __UpperCamelCase : List[Any] = model_inputs.pop("is_last" ) __UpperCamelCase : Any = model_inputs.pop("original_sizes" ).tolist() __UpperCamelCase : Union[str, Any] = model_inputs.pop("reshaped_input_sizes" ).tolist() __UpperCamelCase : List[Any] = self.model(**_UpperCAmelCase ) # post processing happens here in order to avoid CPU GPU copies of ALL the masks __UpperCamelCase : int = model_outputs['''pred_masks'''] __UpperCamelCase : str = self.image_processor.post_process_masks( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , binarize=_UpperCAmelCase ) __UpperCamelCase : str = model_outputs['''iou_scores'''] __UpperCamelCase : Optional[int] = self.image_processor.filter_masks( masks[0] , iou_scores[0] , original_sizes[0] , input_boxes[0] , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , ) return { "masks": masks, "is_last": is_last, "boxes": boxes, "iou_scores": iou_scores, } def _lowerCamelCase ( self :Union[str, Any] , a :List[Any] , a :int=False , a :Optional[int]=False , a :List[str]=0.7 , ) -> Optional[int]: __UpperCamelCase : int = [] __UpperCamelCase : str = [] __UpperCamelCase : List[Any] = [] for model_output in model_outputs: all_scores.append(model_output.pop("iou_scores" ) ) all_masks.extend(model_output.pop("masks" ) ) all_boxes.append(model_output.pop("boxes" ) ) __UpperCamelCase : Any = torch.cat(_UpperCAmelCase ) __UpperCamelCase : Dict = torch.cat(_UpperCAmelCase ) __UpperCamelCase : Any = self.image_processor.post_process_for_mask_generation( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) __UpperCamelCase : Union[str, Any] = defaultdict(_UpperCAmelCase ) for output in model_outputs: for k, v in output.items(): extra[k].append(_UpperCAmelCase ) __UpperCamelCase : Any = {} if output_rle_mask: __UpperCamelCase : Optional[Any] = rle_mask if output_bboxes_mask: __UpperCamelCase : Optional[int] = bounding_boxes return {"masks": output_masks, "scores": iou_scores, **optional, **extra}
232
"""simple docstring""" from typing import List, Optional, Tuple, Union import torch from torch import nn from torch.nn import CrossEntropyLoss from ... import AutoBackbone from ...modeling_outputs import SemanticSegmenterOutput from ...modeling_utils import PreTrainedModel from ...utils import add_start_docstrings, add_start_docstrings_to_model_forward, replace_return_docstrings from ...utils.backbone_utils import BackboneMixin from .configuration_upernet import UperNetConfig __A = [ "openmmlab/upernet-convnext-tiny", # See all UperNet models at https://huggingface.co/models?filter=upernet ] # General docstring __A = "UperNetConfig" class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase = 0 , _UpperCAmelCase = False , _UpperCAmelCase = 1 , ): super().__init__() lowercase__: List[Any] = nn.Convad( in_channels=_UpperCAmelCase , out_channels=_UpperCAmelCase , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , bias=_UpperCAmelCase , dilation=_UpperCAmelCase , ) lowercase__: List[Any] = nn.BatchNormad(_UpperCAmelCase ) lowercase__: int = nn.ReLU() def _snake_case ( self , _UpperCAmelCase ): lowercase__: Dict = self.conv(_UpperCAmelCase ) lowercase__: Optional[int] = self.batch_norm(_UpperCAmelCase ) lowercase__: List[Any] = self.activation(_UpperCAmelCase ) return output class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: int = [ nn.AdaptiveAvgPoolad(_UpperCAmelCase ), UperNetConvModule(_UpperCAmelCase , _UpperCAmelCase , kernel_size=1 ), ] for i, layer in enumerate(self.layers ): self.add_module(str(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase ): lowercase__: Any = input for layer in self.layers: lowercase__: Any = layer(_UpperCAmelCase ) return hidden_state class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: int = pool_scales lowercase__: Optional[Any] = align_corners lowercase__: Optional[int] = in_channels lowercase__: Optional[Any] = channels lowercase__: List[Any] = [] for i, pool_scale in enumerate(_UpperCAmelCase ): lowercase__: Optional[int] = UperNetPyramidPoolingBlock(pool_scale=_UpperCAmelCase , in_channels=_UpperCAmelCase , channels=_UpperCAmelCase ) self.blocks.append(_UpperCAmelCase ) self.add_module(str(_UpperCAmelCase ) , _UpperCAmelCase ) def _snake_case ( self , _UpperCAmelCase ): lowercase__: Union[str, Any] = [] for ppm in self.blocks: lowercase__: Tuple = ppm(_UpperCAmelCase ) lowercase__: Any = nn.functional.interpolate( _UpperCAmelCase , size=x.size()[2:] , mode='''bilinear''' , align_corners=self.align_corners ) ppm_outs.append(_UpperCAmelCase ) return ppm_outs class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase ): super().__init__() lowercase__: Optional[int] = config lowercase__: int = config.pool_scales # e.g. (1, 2, 3, 6) lowercase__: Optional[int] = in_channels lowercase__: List[str] = config.hidden_size lowercase__: List[str] = False lowercase__: List[str] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) # PSP Module lowercase__: Dict = UperNetPyramidPoolingModule( self.pool_scales , self.in_channels[-1] , self.channels , align_corners=self.align_corners , ) lowercase__: int = UperNetConvModule( self.in_channels[-1] + len(self.pool_scales ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) # FPN Module lowercase__: List[Any] = nn.ModuleList() lowercase__: Union[str, Any] = nn.ModuleList() for in_channels in self.in_channels[:-1]: # skip the top layer lowercase__: int = UperNetConvModule(_UpperCAmelCase , self.channels , kernel_size=1 ) lowercase__: Dict = UperNetConvModule(self.channels , self.channels , kernel_size=3 , padding=1 ) self.lateral_convs.append(_UpperCAmelCase ) self.fpn_convs.append(_UpperCAmelCase ) lowercase__: Any = UperNetConvModule( len(self.in_channels ) * self.channels , self.channels , kernel_size=3 , padding=1 , ) def _snake_case ( self ): self.apply(self._init_weights ) def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def _snake_case ( self , _UpperCAmelCase ): lowercase__: List[str] = inputs[-1] lowercase__: str = [x] psp_outs.extend(self.psp_modules(_UpperCAmelCase ) ) lowercase__: Dict = torch.cat(_UpperCAmelCase , dim=1 ) lowercase__: Tuple = self.bottleneck(_UpperCAmelCase ) return output def _snake_case ( self , _UpperCAmelCase ): # build laterals lowercase__: Dict = [lateral_conv(encoder_hidden_states[i] ) for i, lateral_conv in enumerate(self.lateral_convs )] laterals.append(self.psp_forward(_UpperCAmelCase ) ) # build top-down path lowercase__: int = len(_UpperCAmelCase ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowercase__: str = laterals[i - 1].shape[2:] lowercase__: Optional[int] = laterals[i - 1] + nn.functional.interpolate( laterals[i] , size=_UpperCAmelCase , mode='''bilinear''' , align_corners=self.align_corners ) # build outputs lowercase__: str = [self.fpn_convs[i](laterals[i] ) for i in range(used_backbone_levels - 1 )] # append psp feature fpn_outs.append(laterals[-1] ) for i in range(used_backbone_levels - 1 , 0 , -1 ): lowercase__: Any = nn.functional.interpolate( fpn_outs[i] , size=fpn_outs[0].shape[2:] , mode='''bilinear''' , align_corners=self.align_corners ) lowercase__: int = torch.cat(_UpperCAmelCase , dim=1 ) lowercase__: Tuple = self.fpn_bottleneck(_UpperCAmelCase ) lowercase__: Dict = self.classifier(_UpperCAmelCase ) return output class UpperCAmelCase (nn.Module ): """simple docstring""" def __init__( self , _UpperCAmelCase , _UpperCAmelCase = 2 , _UpperCAmelCase = 3 , _UpperCAmelCase = 1 ): super().__init__() lowercase__: Optional[Any] = config lowercase__: Optional[Any] = config.auxiliary_in_channels lowercase__: List[Any] = config.auxiliary_channels lowercase__: Tuple = config.auxiliary_num_convs lowercase__: Any = config.auxiliary_concat_input lowercase__: Optional[int] = in_index lowercase__: Tuple = (kernel_size // 2) * dilation lowercase__: Tuple = [] convs.append( UperNetConvModule( self.in_channels , self.channels , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , dilation=_UpperCAmelCase ) ) for i in range(self.num_convs - 1 ): convs.append( UperNetConvModule( self.channels , self.channels , kernel_size=_UpperCAmelCase , padding=_UpperCAmelCase , dilation=_UpperCAmelCase ) ) if self.num_convs == 0: lowercase__: List[Any] = nn.Identity() else: lowercase__: Union[str, Any] = nn.Sequential(*_UpperCAmelCase ) if self.concat_input: lowercase__: Dict = UperNetConvModule( self.in_channels + self.channels , self.channels , kernel_size=_UpperCAmelCase , padding=kernel_size // 2 ) lowercase__: Union[str, Any] = nn.Convad(self.channels , config.num_labels , kernel_size=1 ) def _snake_case ( self ): self.apply(self._init_weights ) def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , nn.Convad ): module.weight.data.normal_(mean=0.0 , std=self.config.initializer_range ) if module.bias is not None: module.bias.data.zero_() def _snake_case ( self , _UpperCAmelCase ): # just take the relevant feature maps lowercase__: Dict = encoder_hidden_states[self.in_index] lowercase__: Optional[int] = self.convs(_UpperCAmelCase ) if self.concat_input: lowercase__: Optional[int] = self.conv_cat(torch.cat([hidden_states, output] , dim=1 ) ) lowercase__: Dict = self.classifier(_UpperCAmelCase ) return output class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" _UpperCAmelCase :Dict = UperNetConfig _UpperCAmelCase :int = "pixel_values" _UpperCAmelCase :Optional[Any] = True def _snake_case ( self , _UpperCAmelCase ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): module.backbone.init_weights() module.decode_head.init_weights() module.auxiliary_head.init_weights() def _snake_case ( self ): self.backbone.init_weights() self.decode_head.init_weights() self.auxiliary_head.init_weights() def _snake_case ( self , _UpperCAmelCase , _UpperCAmelCase=False ): if isinstance(_UpperCAmelCase , _UpperCAmelCase ): lowercase__: Any = value __A = R"\n Parameters:\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) sub-class. Use\n it as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n config ([`UperNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n" __A = R"\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Padding will be ignored by default should you provide it. Pixel values can be obtained using\n [`AutoImageProcessor`]. See [`SegformerImageProcessor.__call__`] for details.\n output_attentions (`bool`, *optional*):\n Whether or not to return the attentions tensors of all attention layers in case the backbone has them. See\n `attentions` under returned tensors for more detail.\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers of the backbone. See `hidden_states` under\n returned tensors for more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n" @add_start_docstrings( "UperNet framework leveraging any vision backbone e.g. for ADE20k, CityScapes." ,_UpperCAmelCase ,) class UpperCAmelCase (_UpperCAmelCase ): """simple docstring""" def __init__( self , _UpperCAmelCase ): super().__init__(_UpperCAmelCase ) lowercase__: Optional[int] = AutoBackbone.from_config(config.backbone_config ) # Semantic segmentation head(s) lowercase__: Any = UperNetHead(_UpperCAmelCase , in_channels=self.backbone.channels ) lowercase__: Tuple = UperNetFCNHead(_UpperCAmelCase ) if config.use_auxiliary_head else None # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(UPERNET_INPUTS_DOCSTRING.format('''batch_size, sequence_length''' ) ) @replace_return_docstrings(output_type=_UpperCAmelCase , config_class=_CONFIG_FOR_DOC ) def _snake_case ( self , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , _UpperCAmelCase = None , ): lowercase__: Tuple = return_dict if return_dict is not None else self.config.use_return_dict lowercase__: Any = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) lowercase__: str = output_attentions if output_attentions is not None else self.config.output_attentions lowercase__: List[str] = self.backbone.forward_with_filtered_kwargs( _UpperCAmelCase , output_hidden_states=_UpperCAmelCase , output_attentions=_UpperCAmelCase ) lowercase__: Tuple = outputs.feature_maps lowercase__: Union[str, Any] = self.decode_head(_UpperCAmelCase ) lowercase__: str = nn.functional.interpolate(_UpperCAmelCase , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=_UpperCAmelCase ) lowercase__: Any = None if self.auxiliary_head is not None: lowercase__: Union[str, Any] = self.auxiliary_head(_UpperCAmelCase ) lowercase__: Tuple = nn.functional.interpolate( _UpperCAmelCase , size=pixel_values.shape[2:] , mode='''bilinear''' , align_corners=_UpperCAmelCase ) lowercase__: List[Any] = None if labels is not None: if self.config.num_labels == 1: raise ValueError('''The number of labels should be greater than one''' ) else: # compute weighted loss lowercase__: List[str] = CrossEntropyLoss(ignore_index=self.config.loss_ignore_index ) lowercase__: Optional[Any] = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: Dict = loss_fct(_UpperCAmelCase , _UpperCAmelCase ) lowercase__: int = main_loss + self.config.auxiliary_loss_weight * auxiliary_loss if not return_dict: if output_hidden_states: lowercase__: Tuple = (logits,) + outputs[1:] else: lowercase__: Optional[int] = (logits,) + outputs[2:] return ((loss,) + output) if loss is not None else output return SemanticSegmenterOutput( loss=_UpperCAmelCase , logits=_UpperCAmelCase , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
177
0
'''simple docstring''' import os def a__ ( ): """simple docstring""" __SCREAMING_SNAKE_CASE = os.path.dirname(os.path.realpath(a__ ) ) __SCREAMING_SNAKE_CASE = os.path.join(a__ , """triangle.txt""" ) with open(a__ ) as f: __SCREAMING_SNAKE_CASE = f.readlines() __SCREAMING_SNAKE_CASE = [] for line in triangle: __SCREAMING_SNAKE_CASE = [] for number in line.strip().split(""" """ ): numbers_from_line.append(int(a__ ) ) a.append(a__ ) for i in range(1 , len(a__ ) ): for j in range(len(a[i] ) ): __SCREAMING_SNAKE_CASE = a[i - 1][j] if j != len(a[i - 1] ) else 0 __SCREAMING_SNAKE_CASE = a[i - 1][j - 1] if j > 0 else 0 a[i][j] += max(a__ , a__ ) return max(a[-1] ) if __name__ == "__main__": print(solution())
331
'''simple docstring''' class lowerCAmelCase__ : # Public class to implement a graph """simple docstring""" def __init__( self : Dict , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : list[list[bool]] ) -> None: """simple docstring""" __SCREAMING_SNAKE_CASE = row __SCREAMING_SNAKE_CASE = col __SCREAMING_SNAKE_CASE = graph def UpperCAmelCase__ ( self : List[str] , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : list[list[bool]] ) -> bool: """simple docstring""" return ( 0 <= i < self.ROW and 0 <= j < self.COL and not visited[i][j] and self.graph[i][j] ) def UpperCAmelCase__ ( self : int , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : int , __SCREAMING_SNAKE_CASE : list[list[bool]] ) -> None: """simple docstring""" __SCREAMING_SNAKE_CASE = [-1, -1, -1, 0, 0, 1, 1, 1] # Coordinate order __SCREAMING_SNAKE_CASE = [-1, 0, 1, -1, 1, -1, 0, 1] __SCREAMING_SNAKE_CASE = True # Make those cells visited for k in range(8 ): if self.is_safe(i + row_nbr[k] , j + col_nbr[k] , __SCREAMING_SNAKE_CASE ): self.diffs(i + row_nbr[k] , j + col_nbr[k] , __SCREAMING_SNAKE_CASE ) def UpperCAmelCase__ ( self : Tuple ) -> int: # And finally, count all islands. """simple docstring""" __SCREAMING_SNAKE_CASE = [[False for j in range(self.COL )] for i in range(self.ROW )] __SCREAMING_SNAKE_CASE = 0 for i in range(self.ROW ): for j in range(self.COL ): if visited[i][j] is False and self.graph[i][j] == 1: self.diffs(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) count += 1 return count
331
1
from __future__ import annotations import matplotlib.pyplot as plt # type: ignore import numpy # initial triangle of Koch snowflake lowerCamelCase = numpy.array([0, 0]) lowerCamelCase = numpy.array([0.5, 0.866_0254]) lowerCamelCase = numpy.array([1, 0]) lowerCamelCase = [VECTOR_1, VECTOR_2, VECTOR_3, VECTOR_1] def a_ ( SCREAMING_SNAKE_CASE__ : List[Any] , SCREAMING_SNAKE_CASE__ : Tuple ): '''simple docstring''' _lowerCamelCase : Optional[Any] =initial_vectors for _ in range(UpperCAmelCase_ ): _lowerCamelCase : Any =iteration_step(UpperCAmelCase_ ) return vectors def a_ ( SCREAMING_SNAKE_CASE__ : Union[str, Any] ): '''simple docstring''' _lowerCamelCase : int =[] for i, start_vector in enumerate(vectors[:-1] ): _lowerCamelCase : Union[str, Any] =vectors[i + 1] new_vectors.append(UpperCAmelCase_ ) _lowerCamelCase : Optional[Any] =end_vector - start_vector new_vectors.append(start_vector + difference_vector / 3 ) new_vectors.append( start_vector + difference_vector / 3 + rotate(difference_vector / 3 , 60 ) ) new_vectors.append(start_vector + difference_vector * 2 / 3 ) new_vectors.append(vectors[-1] ) return new_vectors def a_ ( SCREAMING_SNAKE_CASE__ : List[str] , SCREAMING_SNAKE_CASE__ : List[str] ): '''simple docstring''' _lowerCamelCase : Dict =numpy.radians(UpperCAmelCase_ ) _lowerCamelCase : Any =numpy.cos(UpperCAmelCase_ ), numpy.sin(UpperCAmelCase_ ) _lowerCamelCase : int =numpy.array(((c, -s), (s, c)) ) return numpy.dot(UpperCAmelCase_ , UpperCAmelCase_ ) def a_ ( SCREAMING_SNAKE_CASE__ : List[Any] ): '''simple docstring''' _lowerCamelCase : Any =plt.gca() axes.set_aspect('equal' ) # matplotlib.pyplot.plot takes a list of all x-coordinates and a list of all # y-coordinates as inputs, which are constructed from the vector-list using # zip() _lowerCamelCase : Optional[Any] =zip(*UpperCAmelCase_ ) plt.plot(UpperCAmelCase_ , UpperCAmelCase_ ) plt.show() if __name__ == "__main__": import doctest doctest.testmod() lowerCamelCase = iterate(INITIAL_VECTORS, 5) plot(processed_vectors)
199
'''simple docstring''' import unittest import numpy as np from transformers import DistilBertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.distilbert.modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, ) class lowercase__ ( unittest.TestCase ): def __init__( self : List[str] ,lowerCamelCase__ : List[str] ,lowerCamelCase__ : List[str]=13 ,lowerCamelCase__ : Dict=7 ,lowerCamelCase__ : Union[str, Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : List[Any]=True ,lowerCamelCase__ : Any=True ,lowerCamelCase__ : Dict=99 ,lowerCamelCase__ : int=32 ,lowerCamelCase__ : Tuple=5 ,lowerCamelCase__ : Dict=4 ,lowerCamelCase__ : Any=37 ,lowerCamelCase__ : str="gelu" ,lowerCamelCase__ : List[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=0.1 ,lowerCamelCase__ : Optional[Any]=512 ,lowerCamelCase__ : Any=16 ,lowerCamelCase__ : Tuple=2 ,lowerCamelCase__ : int=0.0_2 ,lowerCamelCase__ : int=4 ,): '''simple docstring''' _UpperCamelCase : List[Any] = parent _UpperCamelCase : Dict = batch_size _UpperCamelCase : Union[str, Any] = seq_length _UpperCamelCase : Optional[Any] = is_training _UpperCamelCase : Optional[int] = use_attention_mask _UpperCamelCase : Any = use_token_type_ids _UpperCamelCase : str = use_labels _UpperCamelCase : Any = vocab_size _UpperCamelCase : List[Any] = hidden_size _UpperCamelCase : Dict = num_hidden_layers _UpperCamelCase : Dict = num_attention_heads _UpperCamelCase : str = intermediate_size _UpperCamelCase : int = hidden_act _UpperCamelCase : Any = hidden_dropout_prob _UpperCamelCase : Any = attention_probs_dropout_prob _UpperCamelCase : List[str] = max_position_embeddings _UpperCamelCase : Optional[int] = type_vocab_size _UpperCamelCase : str = type_sequence_label_size _UpperCamelCase : Dict = initializer_range _UpperCamelCase : List[Any] = num_choices def UpperCamelCase_ ( self : List[str] ): '''simple docstring''' _UpperCamelCase : int = ids_tensor([self.batch_size, self.seq_length] ,self.vocab_size ) _UpperCamelCase : Union[str, Any] = None if self.use_attention_mask: _UpperCamelCase : Union[str, Any] = random_attention_mask([self.batch_size, self.seq_length] ) _UpperCamelCase : Any = DistilBertConfig( vocab_size=self.vocab_size ,dim=self.hidden_size ,n_layers=self.num_hidden_layers ,n_heads=self.num_attention_heads ,hidden_dim=self.intermediate_size ,hidden_act=self.hidden_act ,dropout=self.hidden_dropout_prob ,attention_dropout=self.attention_probs_dropout_prob ,max_position_embeddings=self.max_position_embeddings ,initializer_range=self.initializer_range ,tie_weights_=lowerCamelCase__ ,) return config, input_ids, attention_mask def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = self.prepare_config_and_inputs() _UpperCamelCase , _UpperCamelCase , _UpperCamelCase : List[Any] = config_and_inputs _UpperCamelCase : Optional[int] = {'input_ids': input_ids, 'attention_mask': attention_mask} return config, inputs_dict @require_flax class lowercase__ ( lowercase , unittest.TestCase ): lowercase__ = ( ( FlaxDistilBertModel, FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertForQuestionAnswering, ) if is_flax_available() else () ) def UpperCamelCase_ ( self : Optional[int] ): '''simple docstring''' _UpperCamelCase : List[str] = FlaxDistilBertModelTester(self ) @slow def UpperCamelCase_ ( self : List[Any] ): '''simple docstring''' for model_class_name in self.all_model_classes: _UpperCamelCase : Dict = model_class_name.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : Optional[int] = model(np.ones((1, 1) ) ) self.assertIsNotNone(lowerCamelCase__ ) @require_flax class lowercase__ ( unittest.TestCase ): @slow def UpperCamelCase_ ( self : str ): '''simple docstring''' _UpperCamelCase : Optional[Any] = FlaxDistilBertModel.from_pretrained('distilbert-base-uncased' ) _UpperCamelCase : List[Any] = np.array([[0, 345, 232, 328, 740, 140, 1695, 69, 6078, 1588, 2]] ) _UpperCamelCase : Tuple = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) _UpperCamelCase : Dict = model(lowerCamelCase__ ,attention_mask=lowerCamelCase__ )[0] _UpperCamelCase : Any = (1, 11, 768) self.assertEqual(output.shape ,lowerCamelCase__ ) _UpperCamelCase : Union[str, Any] = np.array([[[-0.1_6_3_9, 0.3_2_9_9, 0.1_6_4_8], [-0.1_7_4_6, 0.3_2_8_9, 0.1_7_1_0], [-0.1_8_8_4, 0.3_3_5_7, 0.1_8_1_0]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] ,lowerCamelCase__ ,atol=1E-4 ) )
83
0
"""simple docstring""" from ..utils import DummyObject, requires_backends class UpperCamelCase ( metaclass=snake_case_ ): UpperCamelCase : str = ['''speech'''] def __init__( self : Optional[Any] , *UpperCAmelCase__ : List[Any] , **UpperCAmelCase__ : List[Any] ) -> List[str]: requires_backends(self , ["""speech"""] ) class UpperCamelCase ( metaclass=snake_case_ ): UpperCamelCase : Union[str, Any] = ['''speech'''] def __init__( self : List[str] , *UpperCAmelCase__ : Union[str, Any] , **UpperCAmelCase__ : List[Any] ) -> List[Any]: requires_backends(self , ["""speech"""] )
324
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: _snake_case = None _snake_case = logging.get_logger(__name__) _snake_case = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} _snake_case = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } _snake_case = { 'camembert-base': 512, } _snake_case = '▁' class UpperCamelCase ( snake_case_ ): UpperCamelCase : Any = VOCAB_FILES_NAMES UpperCamelCase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP UpperCamelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES UpperCamelCase : Dict = ['''input_ids''', '''attention_mask'''] UpperCamelCase : Optional[Any] = CamembertTokenizer def __init__( self : int , UpperCAmelCase__ : List[Any]=None , UpperCAmelCase__ : Optional[int]=None , UpperCAmelCase__ : Optional[Any]="<s>" , UpperCAmelCase__ : Optional[int]="</s>" , UpperCAmelCase__ : Tuple="</s>" , UpperCAmelCase__ : Tuple="<s>" , UpperCAmelCase__ : Tuple="<unk>" , UpperCAmelCase__ : Tuple="<pad>" , UpperCAmelCase__ : int="<mask>" , UpperCAmelCase__ : Optional[int]=["<s>NOTUSED", "</s>NOTUSED"] , **UpperCAmelCase__ : Optional[Any] , ) -> Union[str, Any]: # Mask token behave like a normal word, i.e. include the space before it _a : List[Any] = AddedToken(UpperCAmelCase__ , lstrip=UpperCAmelCase__ , rstrip=UpperCAmelCase__ ) if isinstance(UpperCAmelCase__ , UpperCAmelCase__ ) else mask_token super().__init__( UpperCAmelCase__ , tokenizer_file=UpperCAmelCase__ , bos_token=UpperCAmelCase__ , eos_token=UpperCAmelCase__ , sep_token=UpperCAmelCase__ , cls_token=UpperCAmelCase__ , unk_token=UpperCAmelCase__ , pad_token=UpperCAmelCase__ , mask_token=UpperCAmelCase__ , additional_special_tokens=UpperCAmelCase__ , **UpperCAmelCase__ , ) _a : int = vocab_file _a : int = False if not self.vocab_file else True def _lowercase ( self : Union[str, Any] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _a : List[Any] = [self.cls_token_id] _a : Dict = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def _lowercase ( self : Optional[int] , UpperCAmelCase__ : List[int] , UpperCAmelCase__ : Optional[List[int]] = None ) -> List[int]: _a : Union[str, Any] = [self.sep_token_id] _a : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _lowercase ( self : Optional[int] , UpperCAmelCase__ : str , UpperCAmelCase__ : Optional[str] = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( """Your fast tokenizer does not have the necessary information to save the vocabulary for a slow """ """tokenizer.""" ) if not os.path.isdir(UpperCAmelCase__ ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return _a : List[str] = os.path.join( UpperCAmelCase__ , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(UpperCAmelCase__ ): copyfile(self.vocab_file , UpperCAmelCase__ ) return (out_vocab_file,)
324
1
"""simple docstring""" def _A ( UpperCamelCase_ : int, UpperCamelCase_ : int) -> str: '''simple docstring''' if a < 0 or b < 0: raise ValueError("the value of both inputs must be positive") __lowercase = str(bin(UpperCamelCase_))[2:] # remove the leading "0b" __lowercase = str(bin(UpperCamelCase_))[2:] # remove the leading "0b" __lowercase = max(len(UpperCamelCase_), len(UpperCamelCase_)) return "0b" + "".join( str(int(char_a == "1" and char_b == "1")) for char_a, char_b in zip(a_binary.zfill(UpperCamelCase_), b_binary.zfill(UpperCamelCase_))) if __name__ == "__main__": import doctest doctest.testmod()
17
"""simple docstring""" import time from contextlib import contextmanager from pathlib import Path import pytest import requests from huggingface_hub.hf_api import HfApi, HfFolder _a = '__DUMMY_TRANSFORMERS_USER__' _a = 'Dummy User' _a = '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' _a = 'https://hub-ci.huggingface.co' _a = CI_HUB_ENDPOINT + '/datasets/{repo_id}/resolve/{revision}/{path}' _a = CI_HUB_ENDPOINT + '/{repo_id}/resolve/{revision}/{filename}' _a = Path('~/.huggingface/hub_ci_token').expanduser() @pytest.fixture def _A ( UpperCamelCase_ : List[Any]) -> Tuple: '''simple docstring''' monkeypatch.setattr( "huggingface_hub.file_download.HUGGINGFACE_CO_URL_TEMPLATE", UpperCamelCase_) @pytest.fixture def _A ( UpperCamelCase_ : int) -> List[Any]: '''simple docstring''' monkeypatch.setattr("datasets.config.HF_ENDPOINT", UpperCamelCase_) monkeypatch.setattr("datasets.config.HUB_DATASETS_URL", UpperCamelCase_) @pytest.fixture def _A ( UpperCamelCase_ : str) -> Dict: '''simple docstring''' monkeypatch.setattr("huggingface_hub.hf_api.HfFolder.path_token", UpperCamelCase_) @pytest.fixture def _A ( UpperCamelCase_ : Optional[Any], UpperCamelCase_ : List[Any]) -> List[str]: '''simple docstring''' HfFolder.save_token(UpperCamelCase_) yield HfFolder.delete_token() @pytest.fixture(scope="session") def _A ( ) -> List[Any]: '''simple docstring''' return HfApi(endpoint=UpperCamelCase_) @pytest.fixture(scope="session") def _A ( UpperCamelCase_ : HfApi) -> List[Any]: '''simple docstring''' __lowercase = HfFolder.get_token() HfFolder.save_token(UpperCamelCase_) yield CI_HUB_USER_TOKEN if previous_token is not None: HfFolder.save_token(UpperCamelCase_) @pytest.fixture def _A ( UpperCamelCase_ : Dict) -> int: '''simple docstring''' def _cleanup_repo(UpperCamelCase_ : Optional[int]): hf_api.delete_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset") return _cleanup_repo @pytest.fixture def _A ( UpperCamelCase_ : str) -> Any: '''simple docstring''' @contextmanager def _temporary_repo(UpperCamelCase_ : Any): try: yield repo_id finally: cleanup_repo(UpperCamelCase_) return _temporary_repo @pytest.fixture(scope="session") def _A ( UpperCamelCase_ : HfApi, UpperCamelCase_ : str, UpperCamelCase_ : Optional[int]) -> List[Any]: '''simple docstring''' __lowercase = F"""repo_txt_data-{int(time.time() * 10E3)}""" __lowercase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset", private=UpperCamelCase_) hf_api.upload_file( token=UpperCamelCase_, path_or_fileobj=str(UpperCamelCase_), path_in_repo="data/text_data.txt", repo_id=UpperCamelCase_, repo_type="dataset", ) yield repo_id try: hf_api.delete_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset") except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _A ( UpperCamelCase_ : Tuple, UpperCamelCase_ : Any, UpperCamelCase_ : Dict) -> Optional[int]: '''simple docstring''' return hf_private_dataset_repo_txt_data_ @pytest.fixture(scope="session") def _A ( UpperCamelCase_ : HfApi, UpperCamelCase_ : int, UpperCamelCase_ : Optional[int]) -> int: '''simple docstring''' __lowercase = F"""repo_zipped_txt_data-{int(time.time() * 10E3)}""" __lowercase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset", private=UpperCamelCase_) hf_api.upload_file( token=UpperCamelCase_, path_or_fileobj=str(UpperCamelCase_), path_in_repo="data.zip", repo_id=UpperCamelCase_, repo_type="dataset", ) yield repo_id try: hf_api.delete_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset") except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _A ( UpperCamelCase_ : List[str], UpperCamelCase_ : Dict, UpperCamelCase_ : Any) -> int: '''simple docstring''' return hf_private_dataset_repo_zipped_txt_data_ @pytest.fixture(scope="session") def _A ( UpperCamelCase_ : HfApi, UpperCamelCase_ : List[str], UpperCamelCase_ : List[str]) -> List[Any]: '''simple docstring''' __lowercase = F"""repo_zipped_img_data-{int(time.time() * 10E3)}""" __lowercase = F"""{CI_HUB_USER}/{repo_name}""" hf_api.create_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset", private=UpperCamelCase_) hf_api.upload_file( token=UpperCamelCase_, path_or_fileobj=str(UpperCamelCase_), path_in_repo="data.zip", repo_id=UpperCamelCase_, repo_type="dataset", ) yield repo_id try: hf_api.delete_repo(UpperCamelCase_, token=UpperCamelCase_, repo_type="dataset") except (requests.exceptions.HTTPError, ValueError): # catch http error and token invalid error pass @pytest.fixture() def _A ( UpperCamelCase_ : Union[str, Any], UpperCamelCase_ : List[str], UpperCamelCase_ : List[str]) -> str: '''simple docstring''' return hf_private_dataset_repo_zipped_img_data_
17
1
"""simple docstring""" from __future__ import annotations def __A ( a_ :float , a_ :float , a_ :float , ) -> tuple[str, float]: if (stress, tangential_force, area).count(0) != 1: raise ValueError('''You cannot supply more or less than 2 values''') elif stress < 0: raise ValueError('''Stress cannot be negative''') elif tangential_force < 0: raise ValueError('''Tangential Force cannot be negative''') elif area < 0: raise ValueError('''Area cannot be negative''') elif stress == 0: return ( "stress", tangential_force / area, ) elif tangential_force == 0: return ( "tangential_force", stress * area, ) else: return ( "area", tangential_force / stress, ) if __name__ == "__main__": import doctest doctest.testmod()
354
"""simple docstring""" import comet # From: unbabel-comet import torch import datasets A = datasets.logging.get_logger(__name__) A = '''\ @inproceedings{rei-EtAl:2020:WMT, author = {Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon}, title = {Unbabel\'s Participation in the WMT20 Metrics Shared Task}, booktitle = {Proceedings of the Fifth Conference on Machine Translation}, month = {November}, year = {2020}, address = {Online}, publisher = {Association for Computational Linguistics}, pages = {909--918}, } @inproceedings{rei-etal-2020-comet, title = "{COMET}: A Neural Framework for {MT} Evaluation", author = "Rei, Ricardo and Stewart, Craig and Farinha, Ana C and Lavie, Alon", booktitle = "Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP)", month = nov, year = "2020", address = "Online", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/2020.emnlp-main.213", pages = "2685--2702", } ''' A = '''\ Crosslingual Optimized Metric for Evaluation of Translation (COMET) is an open-source framework used to train Machine Translation metrics that achieve high levels of correlation with different types of human judgments (HTER, DA\'s or MQM). With the release of the framework the authors also released fully trained models that were used to compete in the WMT20 Metrics Shared Task achieving SOTA in that years competition. See the [README.md] file at https://unbabel.github.io/COMET/html/models.html for more information. ''' A = ''' COMET score. Args: `sources` (list of str): Source sentences `predictions` (list of str): candidate translations `references` (list of str): reference translations `cuda` (bool): If set to True, runs COMET using GPU `show_progress` (bool): Shows progress `model`: COMET model to be used. Will default to `wmt-large-da-estimator-1719` if None. Returns: `samples`: List of dictionaries with `src`, `mt`, `ref` and `score`. `scores`: List of scores. Examples: >>> comet_metric = datasets.load_metric(\'comet\') >>> # comet_metric = load_metric(\'comet\', \'wmt20-comet-da\') # you can also choose which model to use >>> source = ["Dem Feuer konnte Einhalt geboten werden", "Schulen und Kindergärten wurden eröffnet."] >>> hypothesis = ["The fire could be stopped", "Schools and kindergartens were open"] >>> reference = ["They were able to control the fire.", "Schools and kindergartens opened"] >>> results = comet_metric.compute(predictions=hypothesis, references=reference, sources=source) >>> print([round(v, 2) for v in results["scores"]]) [0.19, 0.92] ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __lowercase ( datasets.Metric ): '''simple docstring''' def _lowerCamelCase ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , homepage='''https://unbabel.github.io/COMET/html/index.html''' , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { '''sources''': datasets.Value('''string''' , id='''sequence''' ), '''predictions''': datasets.Value('''string''' , id='''sequence''' ), '''references''': datasets.Value('''string''' , id='''sequence''' ), } ) , codebase_urls=['''https://github.com/Unbabel/COMET'''] , reference_urls=[ '''https://github.com/Unbabel/COMET''', '''https://www.aclweb.org/anthology/2020.emnlp-main.213/''', '''http://www.statmt.org/wmt20/pdf/2020.wmt-1.101.pdf6''', ] , ) def _lowerCamelCase ( self , _UpperCAmelCase ): if self.config_name == "default": __a : List[str] = comet.load_from_checkpoint(comet.download_model('''wmt20-comet-da''' ) ) else: __a : List[str] = comet.load_from_checkpoint(comet.download_model(self.config_name ) ) def _lowerCamelCase ( self , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase=None , _UpperCAmelCase=False ): if gpus is None: __a : str = 1 if torch.cuda.is_available() else 0 __a : Optional[Any] = {'''src''': sources, '''mt''': predictions, '''ref''': references} __a : Dict = [dict(zip(_UpperCAmelCase , _UpperCAmelCase ) ) for t in zip(*data.values() )] __a , __a : int = self.scorer.predict(_UpperCAmelCase , gpus=_UpperCAmelCase , progress_bar=_UpperCAmelCase ) return {"mean_score": mean_score, "scores": scores}
188
0
import argparse import requests import torch from PIL import Image from transformers import CLIPProcessor, GroupViTConfig, GroupViTModel def A (__A : Optional[int] ) -> str: """simple docstring""" if "img_encoder.pos_embed" in name: UpperCAmelCase_ = name.replace('''img_encoder.pos_embed''' , '''vision_model.embeddings.position_embeddings''' ) if "img_encoder.patch_embed.proj" in name: UpperCAmelCase_ = name.replace('''img_encoder.patch_embed.proj''' , '''vision_model.embeddings.patch_embeddings.projection''' ) if "img_encoder.patch_embed.norm" in name: UpperCAmelCase_ = name.replace('''img_encoder.patch_embed.norm''' , '''vision_model.embeddings.layernorm''' ) if "img_encoder.layers" in name: UpperCAmelCase_ = name.replace('''img_encoder.layers''' , '''vision_model.encoder.stages''' ) if "blocks" in name and "res" not in name: UpperCAmelCase_ = name.replace('''blocks''' , '''layers''' ) if "attn" in name and "pre_assign" not in name: UpperCAmelCase_ = name.replace('''attn''' , '''self_attn''' ) if "proj" in name and "self_attn" in name and "text" not in name: UpperCAmelCase_ = name.replace('''proj''' , '''out_proj''' ) if "pre_assign_attn.attn.proj" in name: UpperCAmelCase_ = name.replace('''pre_assign_attn.attn.proj''' , '''pre_assign_attn.attn.out_proj''' ) if "norm1" in name: UpperCAmelCase_ = name.replace('''norm1''' , '''layer_norm1''' ) if "norm2" in name and "pre_assign" not in name: UpperCAmelCase_ = name.replace('''norm2''' , '''layer_norm2''' ) if "img_encoder.norm" in name: UpperCAmelCase_ = name.replace('''img_encoder.norm''' , '''vision_model.layernorm''' ) # text encoder if "text_encoder.token_embedding" in name: UpperCAmelCase_ = name.replace('''text_encoder.token_embedding''' , '''text_model.embeddings.token_embedding''' ) if "text_encoder.positional_embedding" in name: UpperCAmelCase_ = name.replace('''text_encoder.positional_embedding''' , '''text_model.embeddings.position_embedding.weight''' ) if "text_encoder.transformer.resblocks." in name: UpperCAmelCase_ = name.replace('''text_encoder.transformer.resblocks.''' , '''text_model.encoder.layers.''' ) if "ln_1" in name: UpperCAmelCase_ = name.replace('''ln_1''' , '''layer_norm1''' ) if "ln_2" in name: UpperCAmelCase_ = name.replace('''ln_2''' , '''layer_norm2''' ) if "c_fc" in name: UpperCAmelCase_ = name.replace('''c_fc''' , '''fc1''' ) if "c_proj" in name: UpperCAmelCase_ = name.replace('''c_proj''' , '''fc2''' ) if "text_encoder" in name: UpperCAmelCase_ = name.replace('''text_encoder''' , '''text_model''' ) if "ln_final" in name: UpperCAmelCase_ = name.replace('''ln_final''' , '''final_layer_norm''' ) # projection layers if "img_projector.linear_hidden." in name: UpperCAmelCase_ = name.replace('''img_projector.linear_hidden.''' , '''visual_projection.''' ) if "img_projector.linear_out." in name: UpperCAmelCase_ = name.replace('''img_projector.linear_out.''' , '''visual_projection.3.''' ) if "text_projector.linear_hidden" in name: UpperCAmelCase_ = name.replace('''text_projector.linear_hidden''' , '''text_projection''' ) if "text_projector.linear_out" in name: UpperCAmelCase_ = name.replace('''text_projector.linear_out''' , '''text_projection.3''' ) return name def A (__A : Union[str, Any] , __A : Union[str, Any] ) -> Tuple: """simple docstring""" for key in orig_state_dict.copy().keys(): UpperCAmelCase_ = orig_state_dict.pop(__lowerCamelCase ) if "qkv" in key: # weights and biases of the key, value and query projections of vision encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors UpperCAmelCase_ = key.split('''.''' ) UpperCAmelCase_ , UpperCAmelCase_ = int(key_split[2] ), int(key_split[4] ) UpperCAmelCase_ = config.vision_config.hidden_size if "weight" in key: UpperCAmelCase_ = val[:dim, :] UpperCAmelCase_ = val[dim : dim * 2, :] UpperCAmelCase_ = val[-dim:, :] else: UpperCAmelCase_ = val[:dim] UpperCAmelCase_ = val[dim : dim * 2] UpperCAmelCase_ = val[-dim:] elif "in_proj" in key: # weights and biases of the key, value and query projections of text encoder's attention layers require special treatment: # we need to split them up into separate matrices/vectors UpperCAmelCase_ = key.split('''.''' ) UpperCAmelCase_ = int(key_split[3] ) UpperCAmelCase_ = config.text_config.hidden_size if "weight" in key: UpperCAmelCase_ = val[:dim, :] UpperCAmelCase_ = val[ dim : dim * 2, : ] UpperCAmelCase_ = val[-dim:, :] else: UpperCAmelCase_ = val[:dim] UpperCAmelCase_ = val[dim : dim * 2] UpperCAmelCase_ = val[-dim:] else: UpperCAmelCase_ = rename_key(__lowerCamelCase ) # squeeze if necessary if ( "text_projection.0" in new_name or "text_projection.3" in new_name or "visual_projection.0" in new_name or "visual_projection.3" in new_name ): UpperCAmelCase_ = val.squeeze_() else: UpperCAmelCase_ = val return orig_state_dict def A () -> List[str]: """simple docstring""" UpperCAmelCase_ = '''http://images.cocodataset.org/val2017/000000039769.jpg''' UpperCAmelCase_ = Image.open(requests.get(__lowerCamelCase , stream=__lowerCamelCase ).raw ) return im @torch.no_grad() def A (__A : Tuple , __A : List[Any] , __A : Optional[Any]="groupvit-gcc-yfcc" , __A : str=False ) -> int: """simple docstring""" UpperCAmelCase_ = GroupViTConfig() UpperCAmelCase_ = GroupViTModel(__lowerCamelCase ).eval() UpperCAmelCase_ = torch.load(__lowerCamelCase , map_location='''cpu''' )['''model'''] UpperCAmelCase_ = convert_state_dict(__lowerCamelCase , __lowerCamelCase ) UpperCAmelCase_ , UpperCAmelCase_ = model.load_state_dict(__lowerCamelCase , strict=__lowerCamelCase ) assert missing_keys == ["text_model.embeddings.position_ids"] assert (unexpected_keys == ["multi_label_logit_scale"]) or (len(__lowerCamelCase ) == 0) # verify result UpperCAmelCase_ = CLIPProcessor.from_pretrained('''openai/clip-vit-base-patch32''' ) UpperCAmelCase_ = prepare_img() UpperCAmelCase_ = processor(text=['''a photo of a cat''', '''a photo of a dog'''] , images=__lowerCamelCase , padding=__lowerCamelCase , return_tensors='''pt''' ) with torch.no_grad(): UpperCAmelCase_ = model(**__lowerCamelCase ) if model_name == "groupvit-gcc-yfcc": UpperCAmelCase_ = torch.tensor([[13.3_523, 6.3_629]] ) elif model_name == "groupvit-gcc-redcaps": UpperCAmelCase_ = torch.tensor([[16.1_873, 8.6_230]] ) else: raise ValueError(F"""Model name {model_name} not supported.""" ) assert torch.allclose(outputs.logits_per_image , __lowerCamelCase , atol=1E-3 ) processor.save_pretrained(__lowerCamelCase ) model.save_pretrained(__lowerCamelCase ) print('''Successfully saved processor and model to''' , __lowerCamelCase ) if push_to_hub: print('''Pushing to the hub...''' ) processor.push_to_hub(__lowerCamelCase , organization='''nielsr''' ) model.push_to_hub(__lowerCamelCase , organization='''nielsr''' ) if __name__ == "__main__": snake_case_ : Optional[Any] = argparse.ArgumentParser() parser.add_argument( "--pytorch_dump_folder_path", default=None, type=str, help="Path to dump the processor and PyTorch model." ) parser.add_argument("--checkpoint_path", default=None, type=str, help="Path to GroupViT checkpoint") parser.add_argument( "--model_name", default="groupvit-gccy-fcc", type=str, help="Name of the model. Expecting either \'groupvit-gcc-yfcc\' or \'groupvit-gcc-redcaps\'", ) parser.add_argument( "--push_to_hub", action="store_true", help="Whether or not to push the converted model and processor to the 🤗 hub using the provided `model_name`.", ) snake_case_ : Optional[int] = parser.parse_args() convert_groupvit_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.model_name, args.push_to_hub)
51
"""simple docstring""" import math import numpy as np import qiskit from qiskit import Aer, ClassicalRegister, QuantumCircuit, QuantumRegister, execute def _UpperCAmelCase ( __lowerCamelCase : int = 3 ) -> qiskit.result.counts.Counts: if isinstance(__lowerCamelCase , __lowerCamelCase ): raise TypeError('''number of qubits must be a integer.''' ) if number_of_qubits <= 0: raise ValueError('''number of qubits must be > 0.''' ) if math.floor(__lowerCamelCase ) != number_of_qubits: raise ValueError('''number of qubits must be exact integer.''' ) if number_of_qubits > 10: raise ValueError('''number of qubits too large to simulate(>10).''' ) _snake_case = QuantumRegister(__lowerCamelCase , '''qr''' ) _snake_case = ClassicalRegister(__lowerCamelCase , '''cr''' ) _snake_case = QuantumCircuit(__lowerCamelCase , __lowerCamelCase ) _snake_case = number_of_qubits for i in range(__lowerCamelCase ): quantum_circuit.h(number_of_qubits - i - 1 ) counter -= 1 for j in range(__lowerCamelCase ): quantum_circuit.cp(np.pi / 2 ** (counter - j) , __lowerCamelCase , __lowerCamelCase ) for k in range(number_of_qubits // 2 ): quantum_circuit.swap(__lowerCamelCase , number_of_qubits - k - 1 ) # measure all the qubits quantum_circuit.measure(__lowerCamelCase , __lowerCamelCase ) # simulate with 10000 shots _snake_case = Aer.get_backend('''qasm_simulator''' ) _snake_case = execute(__lowerCamelCase , __lowerCamelCase , shots=1_00_00 ) return job.result().get_counts(__lowerCamelCase ) if __name__ == "__main__": print( F"Total count for quantum fourier transform state is: \ {quantum_fourier_transform(3)}" )
288
0
'''simple docstring''' import random import unittest import torch from diffusers import IFInpaintingSuperResolutionPipeline from diffusers.utils import floats_tensor from diffusers.utils.import_utils import is_xformers_available from diffusers.utils.testing_utils import skip_mps, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS, TEXT_GUIDED_IMAGE_INPAINTING_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin from . import IFPipelineTesterMixin @skip_mps class __magic_name__ ( _UpperCAmelCase, _UpperCAmelCase, unittest.TestCase): UpperCamelCase__ = IFInpaintingSuperResolutionPipeline UpperCamelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_PARAMS - {'''width''', '''height'''} UpperCamelCase__ = TEXT_GUIDED_IMAGE_INPAINTING_BATCH_PARAMS.union({'''original_image'''}) UpperCamelCase__ = PipelineTesterMixin.required_optional_params - {'''latents'''} def SCREAMING_SNAKE_CASE_ ( self : Tuple ): return self._get_superresolution_dummy_components() def SCREAMING_SNAKE_CASE_ ( self : List[Any] , lowercase_ : str , lowercase_ : List[Any]=0 ): if str(lowercase_ ).startswith("""mps""" ): lowercase_ : List[str] = torch.manual_seed(lowercase_ ) else: lowercase_ : int = torch.Generator(device=lowercase_ ).manual_seed(lowercase_ ) lowercase_ : List[Any] = floats_tensor((1, 3, 16, 16) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) lowercase_ : Dict = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) lowercase_ : Optional[int] = floats_tensor((1, 3, 32, 32) , rng=random.Random(lowercase_ ) ).to(lowercase_ ) lowercase_ : int = { """prompt""": """A painting of a squirrel eating a burger""", """image""": image, """original_image""": original_image, """mask_image""": mask_image, """generator""": generator, """num_inference_steps""": 2, """output_type""": """numpy""", } return inputs @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def SCREAMING_SNAKE_CASE_ ( self : Any ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): self._test_save_load_optional_components() @unittest.skipIf(torch_device != """cuda""" , reason="""float16 requires CUDA""" ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): # Due to non-determinism in save load of the hf-internal-testing/tiny-random-t5 text encoder super().test_save_load_floataa(expected_max_diff=1E-1 ) def SCREAMING_SNAKE_CASE_ ( self : Union[str, Any] ): self._test_attention_slicing_forward_pass(expected_max_diff=1E-2 ) def SCREAMING_SNAKE_CASE_ ( self : str ): self._test_save_load_local() def SCREAMING_SNAKE_CASE_ ( self : Tuple ): self._test_inference_batch_single_identical( expected_max_diff=1E-2 , )
21
'''simple docstring''' import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class __magic_name__ ( unittest.TestCase): def __init__( self : Union[str, Any] , lowercase_ : Union[str, Any] , lowercase_ : int=7 , lowercase_ : Optional[Any]=3 , lowercase_ : Optional[Any]=18 , lowercase_ : List[Any]=30 , lowercase_ : int=400 , lowercase_ : Dict=True , lowercase_ : List[Any]=None , lowercase_ : Dict=True , ): lowercase_ : Tuple = size if size is not None else {"""height""": 18, """width""": 18} lowercase_ : List[str] = parent lowercase_ : Any = batch_size lowercase_ : Optional[Any] = num_channels lowercase_ : Tuple = image_size lowercase_ : Optional[Any] = min_resolution lowercase_ : Dict = max_resolution lowercase_ : Optional[int] = do_resize lowercase_ : Optional[Any] = size lowercase_ : Union[str, Any] = do_normalize def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.88_66_44_36_34_03_32_03, 0.66_18_82_93_69_54_49_83, 0.38_91_74_64_01_78_68_04], [-0.60_42_55_91_46_88_11_04, -0.0_22_95_00_88_60_52_84_69, 0.54_23_79_73_69_00_32_96], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class __magic_name__ ( _UpperCAmelCase, unittest.TestCase): UpperCamelCase__ = ImageGPTImageProcessor if is_vision_available() else None def SCREAMING_SNAKE_CASE_ ( self : List[str] ): lowercase_ : Optional[int] = ImageGPTImageProcessingTester(self ) @property def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): return self.image_processor_tester.prepare_image_processor_dict() def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : int = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(lowercase_ , """clusters""" ) ) self.assertTrue(hasattr(lowercase_ , """do_resize""" ) ) self.assertTrue(hasattr(lowercase_ , """size""" ) ) self.assertTrue(hasattr(lowercase_ , """do_normalize""" ) ) def SCREAMING_SNAKE_CASE_ ( self : Tuple ): lowercase_ : Any = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""height""": 18, """width""": 18} ) lowercase_ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"""height""": 42, """width""": 42} ) def SCREAMING_SNAKE_CASE_ ( self : int ): lowercase_ : int = self.image_processing_class(**self.image_processor_dict ) lowercase_ : Union[str, Any] = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(lowercase_ , obj[key] ) ) else: self.assertEqual(obj[key] , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : Optional[int] ): lowercase_ : str = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: lowercase_ : Union[str, Any] = os.path.join(lowercase_ , """image_processor.json""" ) image_processor_first.to_json_file(lowercase_ ) lowercase_ : Optional[Any] = self.image_processing_class.from_json_file(lowercase_ ).to_dict() lowercase_ : Any = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowercase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowercase_ ) def SCREAMING_SNAKE_CASE_ ( self : List[Any] ): lowercase_ : Tuple = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(lowercase_ ) lowercase_ : Any = self.image_processing_class.from_pretrained(lowercase_ ).to_dict() lowercase_ : List[str] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(lowercase_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , lowercase_ ) @unittest.skip("""ImageGPT requires clusters at initialization""" ) def SCREAMING_SNAKE_CASE_ ( self : Any ): pass def lowerCamelCase ( ) -> Any: lowercase_ : Union[str, Any] = load_dataset("""hf-internal-testing/fixtures_image_utils""" , split="""test""" ) lowercase_ : Any = Image.open(dataset[4]["""file"""] ) lowercase_ : Dict = Image.open(dataset[5]["""file"""] ) lowercase_ : int = [imagea, imagea] return images @require_vision @require_torch class __magic_name__ ( unittest.TestCase): @slow def SCREAMING_SNAKE_CASE_ ( self : Any ): lowercase_ : Optional[Any] = ImageGPTImageProcessor.from_pretrained("""openai/imagegpt-small""" ) lowercase_ : Optional[int] = prepare_images() # test non-batched lowercase_ : str = image_processing(images[0] , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 1024) ) lowercase_ : Tuple = [306, 191, 191] self.assertEqual(encoding.input_ids[0, :3].tolist() , lowercase_ ) # test batched lowercase_ : List[str] = image_processing(lowercase_ , return_tensors="""pt""" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 1024) ) lowercase_ : Union[str, Any] = [303, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , lowercase_ )
21
1
from itertools import product from cva import COLOR_BGR2GRAY, cvtColor, imread, imshow, waitKey from numpy import dot, exp, mgrid, pi, ravel, square, uinta, zeros def A ( _UpperCAmelCase : Dict , _UpperCAmelCase : int ) -> int: '''simple docstring''' _UpperCAmelCase = k_size // 2 _UpperCAmelCase , _UpperCAmelCase = mgrid[0 - center : k_size - center, 0 - center : k_size - center] _UpperCAmelCase = 1 / (2 * pi * sigma) * exp(-(square(_UpperCAmelCase ) + square(_UpperCAmelCase )) / (2 * square(_UpperCAmelCase )) ) return g def A ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : Tuple , _UpperCAmelCase : Any ) -> List[str]: '''simple docstring''' _UpperCAmelCase , _UpperCAmelCase = image.shape[0], image.shape[1] # dst image height and width _UpperCAmelCase = height - k_size + 1 _UpperCAmelCase = width - k_size + 1 # im2col, turn the k_size*k_size pixels into a row and np.vstack all rows _UpperCAmelCase = zeros((dst_height * dst_width, k_size * k_size) ) _UpperCAmelCase = 0 for i, j in product(range(_UpperCAmelCase ) , range(_UpperCAmelCase ) ): _UpperCAmelCase = ravel(image[i : i + k_size, j : j + k_size] ) _UpperCAmelCase = window row += 1 # turn the kernel into shape(k*k, 1) _UpperCAmelCase = gen_gaussian_kernel(_UpperCAmelCase , _UpperCAmelCase ) _UpperCAmelCase = ravel(_UpperCAmelCase ) # reshape and get the dst image _UpperCAmelCase = dot(_UpperCAmelCase , _UpperCAmelCase ).reshape(_UpperCAmelCase , _UpperCAmelCase ).astype(_UpperCAmelCase ) return dst if __name__ == "__main__": # read original image UpperCAmelCase__ = imread(r"../image_data/lena.jpg") # turn image in gray scale value UpperCAmelCase__ = cvtColor(img, COLOR_BGR2GRAY) # get values with two different mask size UpperCAmelCase__ = gaussian_filter(gray, 3, sigma=1) UpperCAmelCase__ = gaussian_filter(gray, 5, sigma=0.8) # show result images imshow("gaussian filter with 3x3 mask", gaussianaxa) imshow("gaussian filter with 5x5 mask", gaussianaxa) waitKey()
339
import asyncio import os import shutil import subprocess import sys import tempfile import unittest from distutils.util import strtobool from functools import partial from pathlib import Path from typing import List, Union from unittest import mock import torch from ..state import AcceleratorState, PartialState from ..utils import ( gather, is_bnb_available, is_comet_ml_available, is_datasets_available, is_deepspeed_available, is_mps_available, is_safetensors_available, is_tensorboard_available, is_torch_version, is_tpu_available, is_transformers_available, is_wandb_available, is_xpu_available, ) def A ( _UpperCAmelCase : Tuple , _UpperCAmelCase : Union[str, Any]=False ) -> str: '''simple docstring''' try: _UpperCAmelCase = os.environ[key] except KeyError: # KEY isn't set, default to `default`. _UpperCAmelCase = default else: # KEY is set, convert it to True or False. try: _UpperCAmelCase = strtobool(_UpperCAmelCase ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(F"If set, {key} must be yes or no." ) return _value UpperCAmelCase__ = parse_flag_from_env("RUN_SLOW", default=False) def A ( _UpperCAmelCase : List[str] ) -> List[str]: '''simple docstring''' return unittest.skip('Test was skipped' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Dict ) -> str: '''simple docstring''' return unittest.skipUnless(_run_slow_tests , 'test is slow' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> str: '''simple docstring''' return unittest.skipUnless(not torch.cuda.is_available() , 'test requires only a CPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Dict ) -> Dict: '''simple docstring''' return unittest.skipUnless(torch.cuda.is_available() , 'test requires a GPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless(is_xpu_available() , 'test requires a XPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[int] ) -> List[str]: '''simple docstring''' return unittest.skipUnless(is_mps_available() , 'test requires a `mps` backend support in `torch`' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless( is_transformers_available() and is_datasets_available() , 'test requires the Hugging Face suite' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : str ) -> str: '''simple docstring''' return unittest.skipUnless(is_bnb_available() , 'test requires the bitsandbytes library' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> List[Any]: '''simple docstring''' return unittest.skipUnless(is_tpu_available() , 'test requires TPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[Any] ) -> str: '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() == 1 , 'test requires a GPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Tuple ) -> int: '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() == 1 , 'test requires a XPU' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(torch.cuda.device_count() > 1 , 'test requires multiple GPUs' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Tuple ) -> Any: '''simple docstring''' return unittest.skipUnless(torch.xpu.device_count() > 1 , 'test requires multiple XPUs' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(is_safetensors_available() , 'test requires safetensors' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[Any] ) -> Dict: '''simple docstring''' return unittest.skipUnless(is_deepspeed_available() , 'test requires DeepSpeed' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Optional[int] ) -> str: '''simple docstring''' return unittest.skipUnless(is_torch_version('>=' , '1.12.0' ) , 'test requires torch version >= 1.12.0' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Any=None , _UpperCAmelCase : List[Any]=None ) -> Dict: '''simple docstring''' if test_case is None: return partial(_UpperCAmelCase , version=_UpperCAmelCase ) return unittest.skipUnless(is_torch_version('>=' , _UpperCAmelCase ) , F"test requires torch version >= {version}" )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[str] ) -> int: '''simple docstring''' return unittest.skipUnless(is_tensorboard_available() , 'test requires Tensorboard' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : Union[str, Any] ) -> Union[str, Any]: '''simple docstring''' return unittest.skipUnless(is_wandb_available() , 'test requires wandb' )(_UpperCAmelCase ) def A ( _UpperCAmelCase : List[str] ) -> Optional[int]: '''simple docstring''' return unittest.skipUnless(is_comet_ml_available() , 'test requires comet_ml' )(_UpperCAmelCase ) UpperCAmelCase__ = ( any([is_wandb_available(), is_tensorboard_available()]) and not is_comet_ml_available() ) def A ( _UpperCAmelCase : List[str] ) -> Any: '''simple docstring''' return unittest.skipUnless( _atleast_one_tracker_available , 'test requires at least one tracker to be available and for `comet_ml` to not be installed' , )(_UpperCAmelCase ) class __lowerCAmelCase ( unittest.TestCase ): UpperCamelCase = True @classmethod def _lowerCamelCase ( cls : List[Any]) -> Tuple: """simple docstring""" _UpperCAmelCase = tempfile.mkdtemp() @classmethod def _lowerCamelCase ( cls : Union[str, Any]) -> str: """simple docstring""" if os.path.exists(cls.tmpdir): shutil.rmtree(cls.tmpdir) def _lowerCamelCase ( self : List[str]) -> List[Any]: """simple docstring""" if self.clear_on_setup: for path in Path(self.tmpdir).glob('**/*'): if path.is_file(): path.unlink() elif path.is_dir(): shutil.rmtree(A) class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Dict) -> Tuple: """simple docstring""" super().tearDown() # Reset the state of the AcceleratorState singleton. AcceleratorState._reset_state() PartialState._reset_state() class __lowerCAmelCase ( unittest.TestCase ): def _lowerCamelCase ( self : Optional[int] , A : Union[mock.Mock, List[mock.Mock]]) -> Tuple: """simple docstring""" _UpperCAmelCase = mocks if isinstance(A , (tuple, list)) else [mocks] for m in self.mocks: m.start() self.addCleanup(m.stop) def A ( _UpperCAmelCase : List[Any] ) -> int: '''simple docstring''' _UpperCAmelCase = AcceleratorState() _UpperCAmelCase = tensor[None].clone().to(state.device ) _UpperCAmelCase = gather(_UpperCAmelCase ).cpu() _UpperCAmelCase = tensor[0].cpu() for i in range(tensors.shape[0] ): if not torch.equal(tensors[i] , _UpperCAmelCase ): return False return True class __lowerCAmelCase : def __init__( self : Optional[Any] , A : Union[str, Any] , A : Optional[int] , A : str) -> Optional[int]: """simple docstring""" _UpperCAmelCase = returncode _UpperCAmelCase = stdout _UpperCAmelCase = stderr async def A ( _UpperCAmelCase : str , _UpperCAmelCase : Optional[int] ) -> Optional[Any]: '''simple docstring''' while True: _UpperCAmelCase = await stream.readline() if line: callback(_UpperCAmelCase ) else: break async def A ( _UpperCAmelCase : Optional[int] , _UpperCAmelCase : List[str]=None , _UpperCAmelCase : str=None , _UpperCAmelCase : str=None , _UpperCAmelCase : Dict=False , _UpperCAmelCase : Union[str, Any]=False ) -> _RunOutput: '''simple docstring''' if echo: print('\nRunning: ' , ' '.join(_UpperCAmelCase ) ) _UpperCAmelCase = await asyncio.create_subprocess_exec( cmd[0] , *cmd[1:] , stdin=_UpperCAmelCase , stdout=asyncio.subprocess.PIPE , stderr=asyncio.subprocess.PIPE , env=_UpperCAmelCase , ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) _UpperCAmelCase = [] _UpperCAmelCase = [] def tee(_UpperCAmelCase : Union[str, Any] , _UpperCAmelCase : List[Any] , _UpperCAmelCase : str , _UpperCAmelCase : str="" ): _UpperCAmelCase = line.decode('utf-8' ).rstrip() sink.append(_UpperCAmelCase ) if not quiet: print(_UpperCAmelCase , _UpperCAmelCase , file=_UpperCAmelCase ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ asyncio.create_task(_read_stream(p.stdout , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stdout , label='stdout:' ) ) ), asyncio.create_task(_read_stream(p.stderr , lambda _UpperCAmelCase : tee(_UpperCAmelCase , _UpperCAmelCase , sys.stderr , label='stderr:' ) ) ), ] , timeout=_UpperCAmelCase , ) return _RunOutput(await p.wait() , _UpperCAmelCase , _UpperCAmelCase ) def A ( _UpperCAmelCase : str , _UpperCAmelCase : Dict=None , _UpperCAmelCase : str=None , _UpperCAmelCase : str=180 , _UpperCAmelCase : List[Any]=False , _UpperCAmelCase : List[Any]=True ) -> _RunOutput: '''simple docstring''' _UpperCAmelCase = asyncio.get_event_loop() _UpperCAmelCase = loop.run_until_complete( _stream_subprocess(_UpperCAmelCase , env=_UpperCAmelCase , stdin=_UpperCAmelCase , timeout=_UpperCAmelCase , quiet=_UpperCAmelCase , echo=_UpperCAmelCase ) ) _UpperCAmelCase = ' '.join(_UpperCAmelCase ) if result.returncode > 0: _UpperCAmelCase = '\n'.join(result.stderr ) raise RuntimeError( F"'{cmd_str}' failed with returncode {result.returncode}\n\n" F"The combined stderr from workers follows:\n{stderr}" ) return result class __lowerCAmelCase ( A ): pass def A ( _UpperCAmelCase : List[str] , _UpperCAmelCase : str=False ) -> Tuple: '''simple docstring''' try: _UpperCAmelCase = subprocess.check_output(_UpperCAmelCase , stderr=subprocess.STDOUT ) if return_stdout: if hasattr(_UpperCAmelCase , 'decode' ): _UpperCAmelCase = output.decode('utf-8' ) return output except subprocess.CalledProcessError as e: raise SubprocessCallException( F"Command `{' '.join(_UpperCAmelCase )}` failed with the following error:\n\n{e.output.decode()}" ) from e
339
1
from typing import TYPE_CHECKING from ...utils import _LazyModule lowercase_ = {"tokenization_byt5": ["ByT5Tokenizer"]} if TYPE_CHECKING: from .tokenization_byta import ByTaTokenizer else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
360
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available lowercase_ = { "configuration_blip_2": [ "BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP", "Blip2Config", "Blip2QFormerConfig", "Blip2VisionConfig", ], "processing_blip_2": ["Blip2Processor"], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowercase_ = [ "BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST", "Blip2Model", "Blip2QFormerModel", "Blip2PreTrainedModel", "Blip2ForConditionalGeneration", "Blip2VisionModel", ] if TYPE_CHECKING: from .configuration_blip_a import ( BLIP_2_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipaConfig, BlipaQFormerConfig, BlipaVisionConfig, ) from .processing_blip_a import BlipaProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip_a import ( BLIP_2_PRETRAINED_MODEL_ARCHIVE_LIST, BlipaForConditionalGeneration, BlipaModel, BlipaPreTrainedModel, BlipaQFormerModel, BlipaVisionModel, ) else: import sys lowercase_ = _LazyModule(__name__, globals()["__file__"], _import_structure, module_spec=__spec__)
11
0
import contextlib from multiprocessing import Pool, RLock from tqdm.auto import tqdm from ..utils import experimental, logging lowerCamelCase : Tuple = logging.get_logger(__name__) class __lowercase : """simple docstring""" _snake_case = None @experimental def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) -> List[Any]: if ParallelBackendConfig.backend_name is None: return _map_with_multiprocessing_pool( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) return _map_with_joblib(lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) -> Tuple: snake_case : List[Any] = num_proc if num_proc <= len(lowercase ) else len(lowercase ) snake_case : Optional[Any] = [] # We organize the splits ourselve (contiguous splits) for index in range(lowercase ): snake_case : Any = len(lowercase ) // num_proc snake_case : Tuple = len(lowercase ) % num_proc snake_case : Union[str, Any] = div * index + min(lowercase ,lowercase ) snake_case : int = start + div + (1 if index < mod else 0) split_kwds.append((function, iterable[start:end], types, index, disable_tqdm, desc) ) if len(lowercase ) != sum(len(i[1] ) for i in split_kwds ): raise ValueError( f"""Error dividing inputs iterable among processes. """ f"""Total number of objects {len(lowercase )}, """ f"""length: {sum(len(i[1] ) for i in split_kwds )}""" ) logger.info( f"""Spawning {num_proc} processes for {len(lowercase )} objects in slices of {[len(i[1] ) for i in split_kwds]}""" ) snake_case , snake_case : Tuple = None, None if not disable_tqdm: snake_case , snake_case : Dict = (RLock(),), tqdm.set_lock with Pool(lowercase ,initargs=lowercase ,initializer=lowercase ) as pool: snake_case : int = pool.map(lowercase ,lowercase ) logger.info(f"""Finished {num_proc} processes""" ) snake_case : Union[str, Any] = [obj for proc_res in mapped for obj in proc_res] logger.info(f"""Unpacked {len(lowercase )} objects""" ) return mapped def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ,lowercase ) -> Tuple: # progress bar is not yet supported for _map_with_joblib, because tqdm couldn't accurately be applied to joblib, # and it requires monkey-patching joblib internal classes which is subject to change import joblib with joblib.parallel_backend(ParallelBackendConfig.backend_name ,n_jobs=lowercase ): return joblib.Parallel()( joblib.delayed(lowercase )((function, obj, types, None, True, None) ) for obj in iterable ) @experimental @contextlib.contextmanager def SCREAMING_SNAKE_CASE__ ( lowercase ) -> Optional[Any]: snake_case : List[Any] = backend_name if backend_name == "spark": from joblibspark import register_spark register_spark() # TODO: call create_cache_and_write_probe if "download" in steps # TODO: raise NotImplementedError when Dataset.map etc is called try: yield finally: snake_case : Dict = None
124
from __future__ import annotations import math def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> list: if len(lowercase ) != 2 or len(a[0] ) != 2 or len(lowercase ) != 2 or len(b[0] ) != 2: raise Exception("""Matrices are not 2x2""" ) snake_case : int = [ [a[0][0] * b[0][0] + a[0][1] * b[1][0], a[0][0] * b[0][1] + a[0][1] * b[1][1]], [a[1][0] * b[0][0] + a[1][1] * b[1][0], a[1][0] * b[0][1] + a[1][1] * b[1][1]], ] return new_matrix def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> int: return [ [matrix_a[row][col] + matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(lowercase ) ) ] def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> str: return [ [matrix_a[row][col] - matrix_b[row][col] for col in range(len(matrix_a[row] ) )] for row in range(len(lowercase ) ) ] def SCREAMING_SNAKE_CASE__ ( lowercase ) -> tuple[list, list, list, list]: if len(lowercase ) % 2 != 0 or len(a[0] ) % 2 != 0: raise Exception("""Odd matrices are not supported!""" ) snake_case : Optional[int] = len(lowercase ) snake_case : str = matrix_length // 2 snake_case : int = [[a[i][j] for j in range(lowercase ,lowercase )] for i in range(lowercase )] snake_case : str = [ [a[i][j] for j in range(lowercase ,lowercase )] for i in range(lowercase ,lowercase ) ] snake_case : Optional[Any] = [[a[i][j] for j in range(lowercase )] for i in range(lowercase )] snake_case : str = [[a[i][j] for j in range(lowercase )] for i in range(lowercase ,lowercase )] return top_left, top_right, bot_left, bot_right def SCREAMING_SNAKE_CASE__ ( lowercase ) -> tuple[int, int]: return len(lowercase ), len(matrix[0] ) def SCREAMING_SNAKE_CASE__ ( lowercase ) -> None: print("""\n""".join(str(lowercase ) for line in matrix ) ) def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> list: if matrix_dimensions(lowercase ) == (2, 2): return default_matrix_multiplication(lowercase ,lowercase ) snake_case , snake_case , snake_case , snake_case : Optional[Any] = split_matrix(lowercase ) snake_case , snake_case , snake_case , snake_case : Any = split_matrix(lowercase ) snake_case : List[Any] = actual_strassen(lowercase ,matrix_subtraction(lowercase ,lowercase ) ) snake_case : List[str] = actual_strassen(matrix_addition(lowercase ,lowercase ) ,lowercase ) snake_case : Tuple = actual_strassen(matrix_addition(lowercase ,lowercase ) ,lowercase ) snake_case : str = actual_strassen(lowercase ,matrix_subtraction(lowercase ,lowercase ) ) snake_case : Union[str, Any] = actual_strassen(matrix_addition(lowercase ,lowercase ) ,matrix_addition(lowercase ,lowercase ) ) snake_case : int = actual_strassen(matrix_subtraction(lowercase ,lowercase ) ,matrix_addition(lowercase ,lowercase ) ) snake_case : List[Any] = actual_strassen(matrix_subtraction(lowercase ,lowercase ) ,matrix_addition(lowercase ,lowercase ) ) snake_case : str = matrix_addition(matrix_subtraction(matrix_addition(lowercase ,lowercase ) ,lowercase ) ,lowercase ) snake_case : List[str] = matrix_addition(lowercase ,lowercase ) snake_case : Any = matrix_addition(lowercase ,lowercase ) snake_case : List[str] = matrix_subtraction(matrix_subtraction(matrix_addition(lowercase ,lowercase ) ,lowercase ) ,lowercase ) # construct the new matrix from our 4 quadrants snake_case : Optional[Any] = [] for i in range(len(lowercase ) ): new_matrix.append(top_left[i] + top_right[i] ) for i in range(len(lowercase ) ): new_matrix.append(bot_left[i] + bot_right[i] ) return new_matrix def SCREAMING_SNAKE_CASE__ ( lowercase ,lowercase ) -> list: if matrix_dimensions(lowercase )[1] != matrix_dimensions(lowercase )[0]: snake_case : Optional[Any] = ( """Unable to multiply these matrices, please check the dimensions.\n""" f"""Matrix A: {matrixa}\n""" f"""Matrix B: {matrixa}""" ) raise Exception(lowercase ) snake_case : str = matrix_dimensions(lowercase ) snake_case : Optional[Any] = matrix_dimensions(lowercase ) if dimensiona[0] == dimensiona[1] and dimensiona[0] == dimensiona[1]: return [matrixa, matrixa] snake_case : Dict = max(*lowercase ,*lowercase ) snake_case : Optional[Any] = int(math.pow(2 ,math.ceil(math.loga(lowercase ) ) ) ) snake_case : Any = matrixa snake_case : Optional[Any] = matrixa # Adding zeros to the matrices so that the arrays dimensions are the same and also # power of 2 for i in range(0 ,lowercase ): if i < dimensiona[0]: for _ in range(dimensiona[1] ,lowercase ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) if i < dimensiona[0]: for _ in range(dimensiona[1] ,lowercase ): new_matrixa[i].append(0 ) else: new_matrixa.append([0] * maxim ) snake_case : Optional[int] = actual_strassen(lowercase ,lowercase ) # Removing the additional zeros for i in range(0 ,lowercase ): if i < dimensiona[0]: for _ in range(dimensiona[1] ,lowercase ): final_matrix[i].pop() else: final_matrix.pop() return final_matrix if __name__ == "__main__": lowerCamelCase : Any = [ [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 4, 3, 1], [2, 3, 6, 7], [3, 1, 2, 4], [2, 3, 4, 5], [6, 2, 3, 1], ] lowerCamelCase : int = [[0, 2, 1, 1], [1_6, 2, 3, 3], [2, 2, 7, 7], [1_3, 1_1, 2_2, 4]] print(strassen(matrixa, matrixa))
124
1
'''simple docstring''' # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse import os import platform import numpy as np import psutil import torch from accelerate import __version__ as version from accelerate.commands.config import default_config_file, load_config_from_file from ..utils import is_npu_available, is_xpu_available def a__ ( lowerCAmelCase__=None ) -> int: if subparsers is not None: UpperCAmelCase__ : Union[str, Any] = subparsers.add_parser('''env''' ) else: UpperCAmelCase__ : List[str] = argparse.ArgumentParser('''Accelerate env command''' ) parser.add_argument( '''--config_file''' , default=lowerCAmelCase__ , help='''The config file to use for the default values in the launching script.''' ) if subparsers is not None: parser.set_defaults(func=lowerCAmelCase__ ) return parser def a__ ( lowerCAmelCase__ ) -> List[Any]: UpperCAmelCase__ : int = torch.__version__ UpperCAmelCase__ : List[str] = torch.cuda.is_available() UpperCAmelCase__ : int = is_xpu_available() UpperCAmelCase__ : int = is_npu_available() UpperCAmelCase__ : Union[str, Any] = '''Not found''' # Get the default from the config file. if args.config_file is not None or os.path.isfile(lowerCAmelCase__ ): UpperCAmelCase__ : Optional[int] = load_config_from_file(args.config_file ).to_dict() UpperCAmelCase__ : int = { '''`Accelerate` version''': version, '''Platform''': platform.platform(), '''Python version''': platform.python_version(), '''Numpy version''': np.__version__, '''PyTorch version (GPU?)''': F"""{pt_version} ({pt_cuda_available})""", '''PyTorch XPU available''': str(lowerCAmelCase__ ), '''PyTorch NPU available''': str(lowerCAmelCase__ ), '''System RAM''': F"""{psutil.virtual_memory().total / 10_24 ** 3:.2f} GB""", } if pt_cuda_available: UpperCAmelCase__ : Union[str, Any] = torch.cuda.get_device_name() print('''\nCopy-and-paste the text below in your GitHub issue\n''' ) print('''\n'''.join([F"""- {prop}: {val}""" for prop, val in info.items()] ) ) print('''- `Accelerate` default config:''' if args.config_file is None else '''- `Accelerate` config passed:''' ) UpperCAmelCase__ : Any = ( '''\n'''.join([F"""\t- {prop}: {val}""" for prop, val in accelerate_config.items()] ) if isinstance(lowerCAmelCase__ , lowerCAmelCase__ ) else F"""\t{accelerate_config}""" ) print(lowerCAmelCase__ ) UpperCAmelCase__ : Any = accelerate_config return info def a__ ( ) -> int: UpperCAmelCase__ : Optional[int] = env_command_parser() UpperCAmelCase__ : Tuple = parser.parse_args() env_command(lowerCAmelCase__ ) return 0 if __name__ == "__main__": raise SystemExit(main())
299
'''simple docstring''' from __future__ import annotations import copy import tempfile import unittest from transformers import CONFIG_MAPPING, AutoConfig, BertConfig, GPTaConfig, TaConfig, TapasConfig, is_tf_available from transformers.testing_utils import ( DUMMY_UNKNOWN_IDENTIFIER, SMALL_MODEL_IDENTIFIER, RequestCounter, require_tensorflow_probability, require_tf, slow, ) from ..bert.test_modeling_bert import BertModelTester if is_tf_available(): from transformers import ( TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSeqaSeqLM, TFAutoModelForSequenceClassification, TFAutoModelForTableQuestionAnswering, TFAutoModelForTokenClassification, TFAutoModelWithLMHead, TFBertForMaskedLM, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertModel, TFFunnelBaseModel, TFFunnelModel, TFGPTaLMHeadModel, TFRobertaForMaskedLM, TFTaForConditionalGeneration, TFTapasForQuestionAnswering, ) from transformers.models.auto.modeling_tf_auto import ( TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_MAPPING, ) from transformers.models.bert.modeling_tf_bert import TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.gpta.modeling_tf_gpta import TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.ta.modeling_tf_ta import TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST from transformers.models.tapas.modeling_tf_tapas import TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST class lowerCamelCase_ ( __a ): lowerCAmelCase__ = 'new-model' if is_tf_available(): class lowerCamelCase_ ( __a ): lowerCAmelCase__ = NewModelConfig @require_tf class lowerCamelCase_ ( unittest.TestCase ): @slow def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[str] = '''bert-base-cased''' UpperCAmelCase__ : int = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : Dict = TFAutoModel.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : int ): '''simple docstring''' UpperCAmelCase__ : str = '''bert-base-cased''' UpperCAmelCase__ : Any = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : List[str] = TFAutoModelForPreTraining.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : int ): '''simple docstring''' for model_name in TF_GPT2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : int = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : str = TFAutoModelForCausalLM.from_pretrained(_A ) UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = TFAutoModelForCausalLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : List[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : List[Any] = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : List[Any] = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : Optional[Any] ): '''simple docstring''' for model_name in TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : int = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : List[Any] = TFAutoModelForMaskedLM.from_pretrained(_A ) UpperCAmelCase__ , UpperCAmelCase__ : List[Any] = TFAutoModelForMaskedLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : Optional[int] ): '''simple docstring''' for model_name in TF_T5_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ : Optional[Any] = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : Dict = TFAutoModelForSeqaSeqLM.from_pretrained(_A ) UpperCAmelCase__ , UpperCAmelCase__ : Dict = TFAutoModelForSeqaSeqLM.from_pretrained(_A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : Any ): '''simple docstring''' for model_name in ["bert-base-uncased"]: UpperCAmelCase__ : Any = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : Any = TFAutoModelForSequenceClassification.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow def lowercase_ ( self : Any ): '''simple docstring''' for model_name in ["bert-base-uncased"]: UpperCAmelCase__ : Optional[Any] = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : Dict = TFAutoModelForQuestionAnswering.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) @slow @require_tensorflow_probability def lowercase_ ( self : Optional[int] ): '''simple docstring''' for model_name in TF_TAPAS_PRETRAINED_MODEL_ARCHIVE_LIST[5:6]: UpperCAmelCase__ : List[str] = AutoConfig.from_pretrained(_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : List[str] = TFAutoModelForTableQuestionAnswering.from_pretrained(_A ) UpperCAmelCase__ , UpperCAmelCase__ : Dict = TFAutoModelForTableQuestionAnswering.from_pretrained( _A , output_loading_info=_A ) self.assertIsNotNone(_A ) self.assertIsInstance(_A , _A ) def lowercase_ ( self : Tuple ): '''simple docstring''' UpperCAmelCase__ : List[Any] = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsInstance(_A , _A ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=_A ) , 14_410 ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' UpperCAmelCase__ : List[Any] = TFAutoModelWithLMHead.from_pretrained(_A ) self.assertIsInstance(_A , _A ) self.assertEqual(model.num_parameters() , 14_410 ) self.assertEqual(model.num_parameters(only_trainable=_A ) , 14_410 ) def lowercase_ ( self : Dict ): '''simple docstring''' UpperCAmelCase__ : int = TFAutoModel.from_pretrained('''sgugger/funnel-random-tiny''' ) self.assertIsInstance(_A , _A ) UpperCAmelCase__ : Any = copy.deepcopy(model.config ) UpperCAmelCase__ : Tuple = ['''FunnelBaseModel'''] UpperCAmelCase__ : int = TFAutoModel.from_config(_A ) self.assertIsInstance(_A , _A ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(_A ) UpperCAmelCase__ : str = TFAutoModel.from_pretrained(_A ) self.assertIsInstance(_A , _A ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' try: AutoConfig.register('''new-model''' , _A ) UpperCAmelCase__ : List[Any] = [ TFAutoModel, TFAutoModelForCausalLM, TFAutoModelForMaskedLM, TFAutoModelForPreTraining, TFAutoModelForQuestionAnswering, TFAutoModelForSequenceClassification, TFAutoModelForTokenClassification, ] for auto_class in auto_classes: with self.subTest(auto_class.__name__ ): # Wrong config class will raise an error with self.assertRaises(_A ): auto_class.register(_A , _A ) auto_class.register(_A , _A ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(_A ): auto_class.register(_A , _A ) # Now that the config is registered, it can be used as any other config with the auto-API UpperCAmelCase__ : Tuple = BertModelTester(self ).get_config() UpperCAmelCase__ : str = NewModelConfig(**tiny_config.to_dict() ) UpperCAmelCase__ : str = auto_class.from_config(_A ) self.assertIsInstance(_A , _A ) with tempfile.TemporaryDirectory() as tmp_dir: model.save_pretrained(_A ) UpperCAmelCase__ : str = auto_class.from_pretrained(_A ) self.assertIsInstance(_A , _A ) finally: if "new-model" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["new-model"] for mapping in ( TF_MODEL_MAPPING, TF_MODEL_FOR_PRETRAINING_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, TF_MODEL_FOR_CAUSAL_LM_MAPPING, TF_MODEL_FOR_MASKED_LM_MAPPING, ): if NewModelConfig in mapping._extra_content: del mapping._extra_content[NewModelConfig] def lowercase_ ( self : str ): '''simple docstring''' with self.assertRaisesRegex( _A , '''bert-base is not a local folder and is not a valid model identifier''' ): UpperCAmelCase__ : Dict = TFAutoModel.from_pretrained('''bert-base''' ) def lowercase_ ( self : Tuple ): '''simple docstring''' with self.assertRaisesRegex( _A , R'''aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)''' ): UpperCAmelCase__ : int = TFAutoModel.from_pretrained(_A , revision='''aaaaaa''' ) def lowercase_ ( self : Tuple ): '''simple docstring''' with self.assertRaisesRegex( _A , '''hf-internal-testing/config-no-model does not appear to have a file named pytorch_model.bin''' , ): UpperCAmelCase__ : List[Any] = TFAutoModel.from_pretrained('''hf-internal-testing/config-no-model''' ) def lowercase_ ( self : Optional[int] ): '''simple docstring''' with self.assertRaisesRegex(_A , '''Use `from_pt=True` to load this model''' ): UpperCAmelCase__ : int = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-bert-pt-only''' ) def lowercase_ ( self : List[str] ): '''simple docstring''' UpperCAmelCase__ : List[str] = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) with RequestCounter() as counter: UpperCAmelCase__ : Union[str, Any] = TFAutoModel.from_pretrained('''hf-internal-testing/tiny-random-bert''' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 ) # With a sharded checkpoint UpperCAmelCase__ : Optional[Any] = TFAutoModel.from_pretrained('''ArthurZ/tiny-random-bert-sharded''' ) with RequestCounter() as counter: UpperCAmelCase__ : List[Any] = TFAutoModel.from_pretrained('''ArthurZ/tiny-random-bert-sharded''' ) self.assertEqual(counter.get_request_count , 0 ) self.assertEqual(counter.head_request_count , 1 ) self.assertEqual(counter.other_request_count , 0 )
299
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class lowercase__ ( UpperCamelCase_): UpperCamelCase_ = """philschmid/bart-large-cnn-samsum""" UpperCamelCase_ = ( """This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, """ """and returns a summary of the text.""" ) UpperCamelCase_ = """summarizer""" UpperCamelCase_ = AutoTokenizer UpperCamelCase_ = AutoModelForSeqaSeqLM UpperCamelCase_ = ["""text"""] UpperCamelCase_ = ["""text"""] def __A ( self : Optional[int] , UpperCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor(UpperCamelCase__ , return_tensors='''pt''' , truncation=UpperCamelCase__ ) def __A ( self : Tuple , UpperCamelCase__ : Tuple ): '''simple docstring''' return self.model.generate(**UpperCamelCase__ )[0] def __A ( self : Tuple , UpperCamelCase__ : Optional[int] ): '''simple docstring''' return self.pre_processor.decode(UpperCamelCase__ , skip_special_tokens=UpperCamelCase__ , clean_up_tokenization_spaces=UpperCamelCase__ )
182
import math def A ( _lowercase ): return math.sqrt(_lowercase ) * math.sqrt(_lowercase ) == num def A ( _lowercase ): SCREAMING_SNAKE_CASE : Dict = 0 SCREAMING_SNAKE_CASE : Tuple = n while left <= right: SCREAMING_SNAKE_CASE : Union[str, Any] = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: SCREAMING_SNAKE_CASE : Optional[Any] = mid - 1 else: SCREAMING_SNAKE_CASE : Optional[int] = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
182
1
'''simple docstring''' def _snake_case ( A ) -> list: if len(A ) <= 1: return [tuple(A )] lowerCAmelCase__ = [] def generate(A , A ): if k == 1: res.append(tuple(arr[:] ) ) return generate(k - 1 , A ) for i in range(k - 1 ): if k % 2 == 0: # k is even lowerCAmelCase__ , lowerCAmelCase__ = arr[k - 1], arr[i] else: # k is odd lowerCAmelCase__ , lowerCAmelCase__ = arr[k - 1], arr[0] generate(k - 1 , A ) generate(len(A ) , A ) return res if __name__ == "__main__": __UpperCAmelCase = input('''Enter numbers separated by a comma:\n''').strip() __UpperCAmelCase = [int(item) for item in user_input.split(''',''')] print(heaps(arr))
228
'''simple docstring''' from sklearn.metrics import fa_score, matthews_corrcoef import datasets from .record_evaluation import evaluate as evaluate_record __UpperCAmelCase = '''\ @article{wang2019superglue, title={SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems}, author={Wang, Alex and Pruksachatkun, Yada and Nangia, Nikita and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R}, journal={arXiv preprint arXiv:1905.00537}, year={2019} } ''' __UpperCAmelCase = '''\ SuperGLUE (https://super.gluebenchmark.com/) is a new benchmark styled after GLUE with a new set of more difficult language understanding tasks, improved resources, and a new public leaderboard. ''' __UpperCAmelCase = ''' Compute SuperGLUE evaluation metric associated to each SuperGLUE dataset. Args: predictions: list of predictions to score. Depending on the SuperGlUE subset: - for \'record\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'prediction_text\': the predicted answer text - for \'multirc\': list of question-answer dictionaries with the following keys: - \'idx\': index of the question-answer pair as specified by the dataset - \'prediction\': the predicted answer label - otherwise: list of predicted labels references: list of reference labels. Depending on the SuperGLUE subset: - for \'record\': list of question-answers dictionaries with the following keys: - \'idx\': index of the question as specified by the dataset - \'answers\': list of possible answers - otherwise: list of reference labels Returns: depending on the SuperGLUE subset: - for \'record\': - \'exact_match\': Exact match between answer and gold answer - \'f1\': F1 score - for \'multirc\': - \'exact_match\': Exact match between answer and gold answer - \'f1_m\': Per-question macro-F1 score - \'f1_a\': Average F1 score over all answers - for \'axb\': \'matthews_correlation\': Matthew Correlation - for \'cb\': - \'accuracy\': Accuracy - \'f1\': F1 score - for all others: - \'accuracy\': Accuracy Examples: >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'copa\') # any of ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"] >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'cb\') >>> predictions = [0, 1] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'accuracy\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'record\') >>> predictions = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'prediction_text\': \'answer\'}] >>> references = [{\'idx\': {\'passage\': 0, \'query\': 0}, \'answers\': [\'answer\', \'another_answer\']}] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'multirc\') >>> predictions = [{\'idx\': {\'answer\': 0, \'paragraph\': 0, \'question\': 0}, \'prediction\': 0}, {\'idx\': {\'answer\': 1, \'paragraph\': 2, \'question\': 3}, \'prediction\': 1}] >>> references = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'exact_match\': 1.0, \'f1_m\': 1.0, \'f1_a\': 1.0} >>> super_glue_metric = datasets.load_metric(\'super_glue\', \'axb\') >>> references = [0, 1] >>> predictions = [0, 1] >>> results = super_glue_metric.compute(predictions=predictions, references=references) >>> print(results) {\'matthews_correlation\': 1.0} ''' def _snake_case ( A , A ) -> List[Any]: return float((preds == labels).mean() ) def _snake_case ( A , A , A="binary" ) -> int: lowerCAmelCase__ = simple_accuracy(A , A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=A , average=A ) ) return { "accuracy": acc, "f1": fa, } def _snake_case ( A , A ) -> List[Any]: lowerCAmelCase__ = {} for id_pred, label in zip(A , A ): lowerCAmelCase__ = F"""{id_pred["idx"]["paragraph"]}-{id_pred["idx"]["question"]}""" lowerCAmelCase__ = id_pred['''prediction'''] if question_id in question_map: question_map[question_id].append((pred, label) ) else: lowerCAmelCase__ = [(pred, label)] lowerCAmelCase__ , lowerCAmelCase__ = [], [] for question, preds_labels in question_map.items(): lowerCAmelCase__ , lowerCAmelCase__ = zip(*A ) lowerCAmelCase__ = fa_score(y_true=A , y_pred=A , average='''macro''' ) fas.append(A ) lowerCAmelCase__ = int(sum(pred == label for pred, label in preds_labels ) == len(A ) ) ems.append(A ) lowerCAmelCase__ = float(sum(A ) / len(A ) ) lowerCAmelCase__ = sum(A ) / len(A ) lowerCAmelCase__ = float(fa_score(y_true=A , y_pred=[id_pred['''prediction'''] for id_pred in ids_preds] ) ) return {"exact_match": em, "f1_m": fa_m, "f1_a": fa_a} @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class a__ ( datasets.Metric ): '''simple docstring''' def __SCREAMING_SNAKE_CASE ( self ) -> Dict: if self.config_name not in [ "boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg", ]: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' ) return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , codebase_urls=[] , reference_urls=[] , format='''numpy''' if not self.config_name == '''record''' and not self.config_name == '''multirc''' else None , ) def __SCREAMING_SNAKE_CASE ( self ) -> Optional[int]: if self.config_name == "record": return { "predictions": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "prediction_text": datasets.Value('''string''' ), }, "references": { "idx": { "passage": datasets.Value('''int64''' ), "query": datasets.Value('''int64''' ), }, "answers": datasets.Sequence(datasets.Value('''string''' ) ), }, } elif self.config_name == "multirc": return { "predictions": { "idx": { "answer": datasets.Value('''int64''' ), "paragraph": datasets.Value('''int64''' ), "question": datasets.Value('''int64''' ), }, "prediction": datasets.Value('''int64''' ), }, "references": datasets.Value('''int64''' ), } else: return { "predictions": datasets.Value('''int64''' ), "references": datasets.Value('''int64''' ), } def __SCREAMING_SNAKE_CASE ( self , lowerCamelCase_ , lowerCamelCase_ ) -> Dict: if self.config_name == "axb": return {"matthews_correlation": matthews_corrcoef(lowerCamelCase_ , lowerCamelCase_ )} elif self.config_name == "cb": return acc_and_fa(lowerCamelCase_ , lowerCamelCase_ , fa_avg='''macro''' ) elif self.config_name == "record": lowerCAmelCase__ = [ { '''qas''': [ {'''id''': ref['''idx''']['''query'''], '''answers''': [{'''text''': ans} for ans in ref['''answers''']]} for ref in references ] } ] lowerCAmelCase__ = {pred['''idx''']['''query''']: pred['''prediction_text'''] for pred in predictions} return evaluate_record(lowerCamelCase_ , lowerCamelCase_ )[0] elif self.config_name == "multirc": return evaluate_multirc(lowerCamelCase_ , lowerCamelCase_ ) elif self.config_name in ["copa", "rte", "wic", "wsc", "wsc.fixed", "boolq", "axg"]: return {"accuracy": simple_accuracy(lowerCamelCase_ , lowerCamelCase_ )} else: raise KeyError( '''You should supply a configuration name selected in ''' '''["boolq", "cb", "copa", "multirc", "record", "rte", "wic", "wsc", "wsc.fixed", "axb", "axg",]''' )
228
1
"""simple docstring""" from typing import Union import fire import torch from tqdm import tqdm def lowercase ( _snake_case : str , _snake_case : str = "cpu" , _snake_case : Union[str, None] = None ) ->None: """simple docstring""" __snake_case : Dict = torch.load(_snake_case , map_location=_snake_case ) for k, v in tqdm(state_dict.items() ): if not isinstance(_snake_case , torch.Tensor ): raise TypeError('''FP16 conversion only works on paths that are saved state dicts, like pytorch_model.bin''' ) __snake_case : str = v.half() if save_path is None: # overwrite src_path __snake_case : Any = src_path torch.save(_snake_case , _snake_case ) if __name__ == "__main__": fire.Fire(convert)
102
"""simple docstring""" from math import factorial, radians def lowercase ( _snake_case : float , _snake_case : int = 18 , _snake_case : int = 10 ) ->float: """simple docstring""" __snake_case : Any = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians __snake_case : int = radians(_snake_case ) __snake_case : str = angle_in_radians __snake_case : Optional[int] = 3 __snake_case : List[Any] = -1 for _ in range(_snake_case ): result += (b * (angle_in_radians**a)) / factorial(_snake_case ) __snake_case : int = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(_snake_case , _snake_case ) if __name__ == "__main__": __import__("""doctest""").testmod()
102
1
'''simple docstring''' import inspect import unittest from transformers import MobileNetVaConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation, MobileNetVaModel from transformers.models.mobilenet_va.modeling_mobilenet_va import MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import MobileNetVaImageProcessor class snake_case__ ( _UpperCAmelCase ): def A_ ( self : List[str] ): '''simple docstring''' __snake_case : List[Any] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(lowercase_ , 'tf_padding' ) ) self.parent.assertTrue(hasattr(lowercase_ , 'depth_multiplier' ) ) class snake_case__ : def __init__( self : List[str] , __a : List[Any] , __a : int=13 , __a : List[str]=3 , __a : Tuple=32 , __a : List[Any]=0.2_5 , __a : int=8 , __a : Any=8 , __a : Tuple=6 , __a : Union[str, Any]=32 , __a : Tuple=True , __a : Optional[Any]=True , __a : Optional[Any]=True , __a : Optional[int]="relu6" , __a : Dict=1280 , __a : Any=0.1 , __a : Any=0.0_2 , __a : List[Any]=True , __a : List[str]=True , __a : List[str]=10 , __a : Optional[Any]=None , ): '''simple docstring''' __snake_case : Optional[int] = parent __snake_case : Any = batch_size __snake_case : List[str] = num_channels __snake_case : Union[str, Any] = image_size __snake_case : List[Any] = depth_multiplier __snake_case : Dict = depth_divisible_by __snake_case : Optional[Any] = min_depth __snake_case : Dict = expand_ratio __snake_case : str = tf_padding __snake_case : List[str] = output_stride __snake_case : List[str] = first_layer_is_expansion __snake_case : List[str] = finegrained_output __snake_case : Optional[Any] = hidden_act __snake_case : Any = last_hidden_size if finegrained_output else int(last_hidden_size * depth_multiplier ) __snake_case : List[str] = classifier_dropout_prob __snake_case : List[Any] = use_labels __snake_case : int = is_training __snake_case : Tuple = num_labels __snake_case : Any = initializer_range __snake_case : Dict = scope def A_ ( self : Tuple ): '''simple docstring''' __snake_case : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : int = None __snake_case : List[str] = None if self.use_labels: __snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Dict = ids_tensor([self.batch_size, self.image_size, self.image_size] , self.num_labels ) __snake_case : int = self.get_config() return config, pixel_values, labels, pixel_labels def A_ ( self : Union[str, Any] ): '''simple docstring''' return MobileNetVaConfig( num_channels=self.num_channels , image_size=self.image_size , depth_multiplier=self.depth_multiplier , depth_divisible_by=self.depth_divisible_by , min_depth=self.min_depth , expand_ratio=self.expand_ratio , output_stride=self.output_stride , first_layer_is_expansion=self.first_layer_is_expansion , finegrained_output=self.finegrained_output , hidden_act=self.hidden_act , tf_padding=self.tf_padding , classifier_dropout_prob=self.classifier_dropout_prob , initializer_range=self.initializer_range , ) def A_ ( self : int , __a : List[str] , __a : List[Any] , __a : Union[str, Any] , __a : Tuple ): '''simple docstring''' __snake_case : Union[str, Any] = MobileNetVaModel(config=lowercase_ ) model.to(lowercase_ ) model.eval() __snake_case : Optional[Any] = model(lowercase_ ) self.parent.assertEqual( result.last_hidden_state.shape , ( self.batch_size, self.last_hidden_size, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) self.parent.assertEqual( result.pooler_output.shape , (self.batch_size, self.last_hidden_size) , ) def A_ ( self : Optional[int] , __a : Tuple , __a : List[str] , __a : List[Any] , __a : List[Any] ): '''simple docstring''' __snake_case : Any = self.num_labels __snake_case : int = MobileNetVaForImageClassification(lowercase_ ) model.to(lowercase_ ) model.eval() __snake_case : int = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A_ ( self : List[Any] , __a : int , __a : Optional[int] , __a : Union[str, Any] , __a : int ): '''simple docstring''' __snake_case : Optional[int] = self.num_labels __snake_case : Optional[int] = MobileNetVaForSemanticSegmentation(lowercase_ ) model.to(lowercase_ ) model.eval() __snake_case : str = model(lowercase_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) __snake_case : Dict = model(lowercase_ , labels=lowercase_ ) self.parent.assertEqual( result.logits.shape , ( self.batch_size, self.num_labels, self.image_size // self.output_stride, self.image_size // self.output_stride, ) , ) def A_ ( self : int ): '''simple docstring''' __snake_case : Optional[int] = self.prepare_config_and_inputs() __snake_case : Tuple = config_and_inputs __snake_case : Union[str, Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_torch class snake_case__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A__ = ( (MobileNetVaModel, MobileNetVaForImageClassification, MobileNetVaForSemanticSegmentation) if is_torch_available() else () ) A__ = ( { '''feature-extraction''': MobileNetVaModel, '''image-classification''': MobileNetVaForImageClassification, '''image-segmentation''': MobileNetVaForSemanticSegmentation, } if is_torch_available() else {} ) A__ = False A__ = False A__ = False A__ = False def A_ ( self : Optional[int] ): '''simple docstring''' __snake_case : Tuple = MobileNetVaModelTester(self ) __snake_case : Optional[int] = MobileNetVaConfigTester(self , config_class=lowercase_ , has_text_modality=lowercase_ ) def A_ ( self : Any ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='MobileNetV2 does not use inputs_embeds' ) def A_ ( self : int ): '''simple docstring''' pass @unittest.skip(reason='MobileNetV2 does not support input and output embeddings' ) def A_ ( self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip(reason='MobileNetV2 does not output attentions' ) def A_ ( self : List[str] ): '''simple docstring''' pass def A_ ( self : List[str] ): '''simple docstring''' __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : List[Any] = model_class(lowercase_ ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : Union[str, Any] = [*signature.parameters.keys()] __snake_case : List[Any] = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , lowercase_ ) def A_ ( self : Optional[Any] ): '''simple docstring''' __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*lowercase_ ) def A_ ( self : Optional[int] ): '''simple docstring''' def check_hidden_states_output(__a : List[Any] , __a : List[str] , __a : int ): __snake_case : int = model_class(lowercase_ ) model.to(lowercase_ ) model.eval() with torch.no_grad(): __snake_case : Optional[Any] = model(**self._prepare_for_class(lowercase_ , lowercase_ ) ) __snake_case : Optional[int] = outputs.hidden_states __snake_case : List[Any] = 16 self.assertEqual(len(lowercase_ ) , lowercase_ ) __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : str = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : str = True check_hidden_states_output(lowercase_ , lowercase_ , lowercase_ ) def A_ ( self : Any ): '''simple docstring''' __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*lowercase_ ) def A_ ( self : Optional[int] ): '''simple docstring''' __snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_semantic_segmentation(*lowercase_ ) @slow def A_ ( self : Dict ): '''simple docstring''' for model_name in MOBILENET_V2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : List[str] = MobileNetVaModel.from_pretrained(lowercase_ ) self.assertIsNotNone(lowercase_ ) def a_ ( ) -> Tuple: __snake_case : str = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class snake_case__ ( unittest.TestCase ): @cached_property def A_ ( self : str ): '''simple docstring''' return ( MobileNetVaImageProcessor.from_pretrained('google/mobilenet_v2_1.0_224' ) if is_vision_available() else None ) @slow def A_ ( self : Optional[int] ): '''simple docstring''' __snake_case : str = MobileNetVaForImageClassification.from_pretrained('google/mobilenet_v2_1.0_224' ).to(lowercase_ ) __snake_case : Optional[Any] = self.default_image_processor __snake_case : Optional[Any] = prepare_img() __snake_case : Dict = image_processor(images=lowercase_ , return_tensors='pt' ).to(lowercase_ ) # forward pass with torch.no_grad(): __snake_case : Optional[int] = model(**lowercase_ ) # verify the logits __snake_case : Optional[Any] = torch.Size((1, 1001) ) self.assertEqual(outputs.logits.shape , lowercase_ ) __snake_case : List[str] = torch.tensor([0.2_4_4_5, -1.1_9_9_3, 0.1_9_0_5] ).to(lowercase_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , lowercase_ , atol=1e-4 ) ) @slow def A_ ( self : int ): '''simple docstring''' __snake_case : Any = MobileNetVaForSemanticSegmentation.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) __snake_case : List[str] = model.to(lowercase_ ) __snake_case : Optional[Any] = MobileNetVaImageProcessor.from_pretrained('google/deeplabv3_mobilenet_v2_1.0_513' ) __snake_case : Optional[Any] = prepare_img() __snake_case : str = image_processor(images=lowercase_ , return_tensors='pt' ).to(lowercase_ ) # forward pass with torch.no_grad(): __snake_case : int = model(**lowercase_ ) __snake_case : List[Any] = outputs.logits # verify the logits __snake_case : List[Any] = torch.Size((1, 21, 65, 65) ) self.assertEqual(logits.shape , lowercase_ ) __snake_case : List[Any] = torch.tensor( [ [[1_7.5_7_9_0, 1_7.7_5_8_1, 1_8.3_3_5_5], [1_8.3_2_5_7, 1_8.4_2_3_0, 1_8.8_9_7_3], [1_8.6_1_6_9, 1_8.8_6_5_0, 1_9.2_1_8_7]], [[-2.1_5_9_5, -2.0_9_7_7, -2.3_7_4_1], [-2.4_2_2_6, -2.3_0_2_8, -2.6_8_3_5], [-2.7_8_1_9, -2.5_9_9_1, -2.7_7_0_6]], [[4.2_0_5_8, 4.8_3_1_7, 4.7_6_3_8], [4.4_1_3_6, 5.0_3_6_1, 4.9_3_8_3], [4.5_0_2_8, 4.9_6_4_4, 4.8_7_3_4]], ] , device=lowercase_ , ) self.assertTrue(torch.allclose(logits[0, :3, :3, :3] , lowercase_ , atol=1e-4 ) )
370
'''simple docstring''' from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging A__ : List[Any] = logging.get_logger(__name__) A__ : Tuple = { '''t5-small''': '''https://huggingface.co/t5-small/resolve/main/config.json''', '''t5-base''': '''https://huggingface.co/t5-base/resolve/main/config.json''', '''t5-large''': '''https://huggingface.co/t5-large/resolve/main/config.json''', '''t5-3b''': '''https://huggingface.co/t5-3b/resolve/main/config.json''', '''t5-11b''': '''https://huggingface.co/t5-11b/resolve/main/config.json''', } class snake_case__ ( SCREAMING_SNAKE_CASE_ ): A__ = '''t5''' A__ = ['''past_key_values'''] A__ = {'''hidden_size''': '''d_model''', '''num_attention_heads''': '''num_heads''', '''num_hidden_layers''': '''num_layers'''} def __init__( self : str , __a : Dict=32128 , __a : Dict=512 , __a : Union[str, Any]=64 , __a : str=2048 , __a : Union[str, Any]=6 , __a : Any=None , __a : Any=8 , __a : List[Any]=32 , __a : Any=128 , __a : Tuple=0.1 , __a : str=1e-6 , __a : Dict=1.0 , __a : Tuple="relu" , __a : Dict=True , __a : Union[str, Any]=True , __a : Any=0 , __a : Dict=1 , **__a : Union[str, Any] , ) -> Union[str, Any]: '''simple docstring''' __snake_case : int = vocab_size __snake_case : str = d_model __snake_case : str = d_kv __snake_case : List[Any] = d_ff __snake_case : List[str] = num_layers __snake_case : Tuple = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry __snake_case : Union[str, Any] = num_heads __snake_case : Tuple = relative_attention_num_buckets __snake_case : Optional[int] = relative_attention_max_distance __snake_case : Optional[Any] = dropout_rate __snake_case : str = layer_norm_epsilon __snake_case : List[str] = initializer_factor __snake_case : int = feed_forward_proj __snake_case : Optional[Any] = use_cache __snake_case : Optional[Any] = self.feed_forward_proj.split('-' ) __snake_case : Dict = act_info[-1] __snake_case : List[str] = act_info[0] == 'gated' if len(__a ) > 1 and act_info[0] != "gated" or len(__a ) > 2: raise ValueError( f'''`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.''' 'Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. ' '\'gated-gelu\' or \'relu\'' ) # for backwards compatibility if feed_forward_proj == "gated-gelu": __snake_case : Dict = 'gelu_new' super().__init__( pad_token_id=__a , eos_token_id=__a , is_encoder_decoder=__a , **__a , ) class snake_case__ ( SCREAMING_SNAKE_CASE_ ): @property def A_ ( self : str ) -> Mapping[str, Mapping[int, str]]: '''simple docstring''' __snake_case : Union[str, Any] = { 'input_ids': {0: 'batch', 1: 'encoder_sequence'}, 'attention_mask': {0: 'batch', 1: 'encoder_sequence'}, } if self.use_past: __snake_case : Tuple = 'past_encoder_sequence + sequence' __snake_case : Dict = {0: 'batch'} __snake_case : Dict = {0: 'batch', 1: 'past_decoder_sequence + sequence'} else: __snake_case : Tuple = {0: 'batch', 1: 'decoder_sequence'} __snake_case : int = {0: 'batch', 1: 'decoder_sequence'} if self.use_past: self.fill_with_past_key_values_(__a , direction='inputs' ) return common_inputs @property def A_ ( self : List[Any] ) -> int: '''simple docstring''' return 13
0
0
import json from typing import List, Optional, Tuple from tokenizers import normalizers from tokenizers.pre_tokenizers import BertPreTokenizer, PreTokenizer from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_roformer import RoFormerTokenizer from .tokenization_utils import JiebaPreTokenizer __a : List[str] = logging.get_logger(__name__) __a : str = {"""vocab_file""": """vocab.txt""", """tokenizer_file""": """tokenizer.json"""} __a : str = { """vocab_file""": { """junnyu/roformer_chinese_small""": """https://huggingface.co/junnyu/roformer_chinese_small/resolve/main/vocab.txt""", """junnyu/roformer_chinese_base""": """https://huggingface.co/junnyu/roformer_chinese_base/resolve/main/vocab.txt""", """junnyu/roformer_chinese_char_small""": ( """https://huggingface.co/junnyu/roformer_chinese_char_small/resolve/main/vocab.txt""" ), """junnyu/roformer_chinese_char_base""": ( """https://huggingface.co/junnyu/roformer_chinese_char_base/resolve/main/vocab.txt""" ), """junnyu/roformer_small_discriminator""": ( """https://huggingface.co/junnyu/roformer_small_discriminator/resolve/main/vocab.txt""" ), """junnyu/roformer_small_generator""": ( """https://huggingface.co/junnyu/roformer_small_generator/resolve/main/vocab.txt""" ), } } __a : Union[str, Any] = { """junnyu/roformer_chinese_small""": 1_5_3_6, """junnyu/roformer_chinese_base""": 1_5_3_6, """junnyu/roformer_chinese_char_small""": 5_1_2, """junnyu/roformer_chinese_char_base""": 5_1_2, """junnyu/roformer_small_discriminator""": 1_2_8, """junnyu/roformer_small_generator""": 1_2_8, } __a : Any = { """junnyu/roformer_chinese_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_base""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_small""": {"""do_lower_case""": True}, """junnyu/roformer_chinese_char_base""": {"""do_lower_case""": True}, """junnyu/roformer_small_discriminator""": {"""do_lower_case""": True}, """junnyu/roformer_small_generator""": {"""do_lower_case""": True}, } class _UpperCamelCase ( _UpperCAmelCase ): """simple docstring""" __a : Optional[int] = VOCAB_FILES_NAMES __a : List[str] = PRETRAINED_VOCAB_FILES_MAP __a : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __a : int = PRETRAINED_INIT_CONFIGURATION __a : Union[str, Any] = RoFormerTokenizer def __init__( self , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=True , lowerCAmelCase__="[UNK]" , lowerCAmelCase__="[SEP]" , lowerCAmelCase__="[PAD]" , lowerCAmelCase__="[CLS]" , lowerCAmelCase__="[MASK]" , lowerCAmelCase__=True , lowerCAmelCase__=None , **lowerCAmelCase__ , ) -> List[Any]: '''simple docstring''' super().__init__( lowerCAmelCase__ , tokenizer_file=lowerCAmelCase__ , do_lower_case=lowerCAmelCase__ , unk_token=lowerCAmelCase__ , sep_token=lowerCAmelCase__ , pad_token=lowerCAmelCase__ , cls_token=lowerCAmelCase__ , mask_token=lowerCAmelCase__ , tokenize_chinese_chars=lowerCAmelCase__ , strip_accents=lowerCAmelCase__ , **lowerCAmelCase__ , ) __lowercase = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( pre_tok_state.get('''lowercase''' , lowerCAmelCase__ ) != do_lower_case or pre_tok_state.get('''strip_accents''' , lowerCAmelCase__ ) != strip_accents ): __lowercase = getattr(lowerCAmelCase__ , pre_tok_state.pop('''type''' ) ) __lowercase = do_lower_case __lowercase = strip_accents __lowercase = pre_tok_class(**lowerCAmelCase__ ) __lowercase = do_lower_case def __getstate__( self ) -> Dict: '''simple docstring''' __lowercase = self.__dict__.copy() __lowercase = BertPreTokenizer() return state def __setstate__( self , lowerCAmelCase__ ) -> int: '''simple docstring''' __lowercase = d __lowercase = self.__dict__['''_tokenizer'''].get_vocab() __lowercase = PreTokenizer.custom(JiebaPreTokenizer(lowerCAmelCase__ ) ) def _SCREAMING_SNAKE_CASE ( self , lowerCAmelCase__ , lowerCAmelCase__=None ) -> List[Any]: '''simple docstring''' __lowercase = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def _SCREAMING_SNAKE_CASE ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> List[int]: '''simple docstring''' __lowercase = [self.sep_token_id] __lowercase = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def _SCREAMING_SNAKE_CASE ( self , lowerCAmelCase__ , lowerCAmelCase__ = None ) -> Tuple[str]: '''simple docstring''' __lowercase = self._tokenizer.model.save(lowerCAmelCase__ , name=lowerCAmelCase__ ) return tuple(lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( self , lowerCAmelCase__ , lowerCAmelCase__=None , lowerCAmelCase__=None , lowerCAmelCase__=False , **lowerCAmelCase__ , ) -> Optional[int]: '''simple docstring''' __lowercase = BertPreTokenizer() return super().save_pretrained(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , **lowerCAmelCase__ )
210
from __future__ import annotations import unittest import numpy as np from transformers import OPTConfig, is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import GPTaTokenizer, TFOPTForCausalLM, TFOPTModel def UpperCAmelCase ( lowercase , lowercase , lowercase=None , lowercase=None ): """simple docstring""" if attention_mask is None: __lowercase = tf.cast(tf.math.not_equal(lowercase , config.pad_token_id ) , tf.inta ) return {"input_ids": input_ids, "attention_mask": attention_mask} @require_tf class _UpperCamelCase : """simple docstring""" __a : Tuple = OPTConfig __a : int = {} __a : Dict = '''gelu''' def __init__( self , lowerCAmelCase__ , lowerCAmelCase__=13 , lowerCAmelCase__=7 , lowerCAmelCase__=True , lowerCAmelCase__=False , lowerCAmelCase__=99 , lowerCAmelCase__=16 , lowerCAmelCase__=2 , lowerCAmelCase__=4 , lowerCAmelCase__=4 , lowerCAmelCase__="gelu" , lowerCAmelCase__=0.1 , lowerCAmelCase__=0.1 , lowerCAmelCase__=20 , lowerCAmelCase__=2 , lowerCAmelCase__=1 , lowerCAmelCase__=0 , lowerCAmelCase__=16 , lowerCAmelCase__=16 , ) -> Tuple: '''simple docstring''' __lowercase = parent __lowercase = batch_size __lowercase = seq_length __lowercase = is_training __lowercase = use_labels __lowercase = vocab_size __lowercase = hidden_size __lowercase = num_hidden_layers __lowercase = num_attention_heads __lowercase = intermediate_size __lowercase = hidden_act __lowercase = hidden_dropout_prob __lowercase = attention_probs_dropout_prob __lowercase = max_position_embeddings __lowercase = eos_token_id __lowercase = pad_token_id __lowercase = bos_token_id __lowercase = embed_dim __lowercase = word_embed_proj_dim __lowercase = False def _SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' __lowercase = ids_tensor([self.batch_size, self.seq_length - 1] , self.vocab_size ) __lowercase = tf.expand_dims(tf.constant([self.eos_token_id] * self.batch_size ) , 1 ) __lowercase = tf.concat([input_ids, eos_tensor] , axis=1 ) __lowercase = self.config_cls( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , embed_dim=self.embed_dim , word_embed_proj_dim=self.word_embed_proj_dim , is_encoder_decoder=lowerCAmelCase__ , **self.config_updates , ) __lowercase = prepare_opt_inputs_dict(lowerCAmelCase__ , lowerCAmelCase__ ) return config, inputs_dict def _SCREAMING_SNAKE_CASE ( self , lowerCAmelCase__ , lowerCAmelCase__ ) -> int: '''simple docstring''' __lowercase = TFOPTModel(config=lowerCAmelCase__ ) __lowercase = inputs_dict['''input_ids'''] __lowercase = input_ids[:1, :] __lowercase = inputs_dict['''attention_mask'''][:1, :] __lowercase = 1 # first forward pass __lowercase = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , use_cache=lowerCAmelCase__ ) __lowercase , __lowercase = outputs.to_tuple() # create hypothetical next token and extent to next_input_ids __lowercase = ids_tensor((self.batch_size, 3) , config.vocab_size ) __lowercase = tf.cast(ids_tensor((self.batch_size, 3) , 2 ) , tf.inta ) # append to next input_ids and __lowercase = tf.concat([input_ids, next_tokens] , axis=-1 ) __lowercase = tf.concat([attention_mask, next_attn_mask] , axis=-1 ) __lowercase = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ )[0] __lowercase = model(lowerCAmelCase__ , attention_mask=lowerCAmelCase__ , past_key_values=lowerCAmelCase__ )[0] self.parent.assertEqual(next_tokens.shape[1] , output_from_past.shape[1] ) # select random slice __lowercase = int(ids_tensor((1,) , output_from_past.shape[-1] ) ) __lowercase = output_from_no_past[:, -3:, random_slice_idx] __lowercase = output_from_past[:, :, random_slice_idx] # test that outputs are equal for slice tf.debugging.assert_near(lowerCAmelCase__ , lowerCAmelCase__ , rtol=1E-3 ) @require_tf class _UpperCamelCase ( _UpperCAmelCase ,_UpperCAmelCase ,unittest.TestCase ): """simple docstring""" __a : int = (TFOPTModel, TFOPTForCausalLM) if is_tf_available() else () __a : Optional[Any] = (TFOPTForCausalLM,) if is_tf_available() else () __a : Dict = ( {'''feature-extraction''': TFOPTModel, '''text-generation''': TFOPTForCausalLM} if is_tf_available() else {} ) __a : List[str] = False __a : Optional[Any] = False __a : Union[str, Any] = False __a : List[Any] = 10 def _SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' __lowercase = TFOPTModelTester(self ) __lowercase = ConfigTester(self , config_class=lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' self.config_tester.run_common_tests() def _SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' __lowercase = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_decoder_model_past_large_inputs(*lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( self ) -> Tuple: '''simple docstring''' __lowercase , __lowercase = self.model_tester.prepare_config_and_inputs_for_common() def _get_word_embedding_weight(lowerCAmelCase__ , lowerCAmelCase__ ): if hasattr(lowerCAmelCase__ , '''weight''' ): return embedding_layer.weight else: # Here we build the word embeddings weights if not exists. # And then we retry to get the attribute once built. model.build() if hasattr(lowerCAmelCase__ , '''weight''' ): return embedding_layer.weight else: return None for model_class in self.all_model_classes: for size in [config.vocab_size - 10, config.vocab_size + 10]: # build the embeddings __lowercase = model_class(config=lowerCAmelCase__ ) __lowercase = _get_word_embedding_weight(lowerCAmelCase__ , model.get_input_embeddings() ) __lowercase = _get_word_embedding_weight(lowerCAmelCase__ , model.get_output_embeddings() ) # reshape the embeddings model.resize_token_embeddings(lowerCAmelCase__ ) __lowercase = _get_word_embedding_weight(lowerCAmelCase__ , model.get_input_embeddings() ) __lowercase = _get_word_embedding_weight(lowerCAmelCase__ , model.get_output_embeddings() ) # check that the resized embeddings size matches the desired size. __lowercase = size if size is not None else config.vocab_size self.assertEqual(new_input_embeddings.shape[0] , lowerCAmelCase__ ) # check that weights remain the same after resizing __lowercase = True for pa, pa in zip(old_input_embeddings.value() , new_input_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowercase = False self.assertTrue(lowerCAmelCase__ ) if old_output_embeddings is not None and new_output_embeddings is not None: self.assertEqual(new_output_embeddings.shape[0] , lowerCAmelCase__ ) __lowercase = True for pa, pa in zip(old_output_embeddings.value() , new_output_embeddings.value() ): if tf.math.reduce_sum(tf.math.abs(pa - pa ) ) > 0: __lowercase = False self.assertTrue(lowerCAmelCase__ ) def UpperCAmelCase ( lowercase ): """simple docstring""" return tf.constant(lowercase , dtype=tf.intaa ) @require_tf class _UpperCamelCase ( unittest.TestCase ): """simple docstring""" __a : List[str] = 99 def _SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' __lowercase = tf.ones((4, 1) , dtype=tf.intaa ) * 2 __lowercase = tf.concat([ids_tensor((4, 6) , self.vocab_size - 3 ) + 3, eos_column_vector] , axis=1 ) __lowercase = input_ids.shape[0] __lowercase = OPTConfig( vocab_size=self.vocab_size , hidden_size=24 , num_hidden_layers=2 , num_attention_heads=2 , ffn_dim=32 , max_position_embeddings=48 , eos_token_id=2 , pad_token_id=1 , bos_token_id=0 , ) return config, input_ids, batch_size @require_sentencepiece @require_tf class _UpperCamelCase ( unittest.TestCase ): """simple docstring""" @slow def _SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' __lowercase = TFOPTModel.from_pretrained('''facebook/opt-350m''' ) __lowercase = _long_tensor([[0, 3_14_14, 2_32, 3_28, 7_40, 11_40, 1_26_95, 69, 4_60_78, 15_88, 2]] ) __lowercase = tf.not_equal(lowerCAmelCase__ , model.config.pad_token_id ) with tf.GradientTape(): __lowercase = model(input_ids=lowerCAmelCase__ , attention_mask=lowerCAmelCase__ ).last_hidden_state __lowercase = (1, 11, 5_12) self.assertEqual(output.shape , lowerCAmelCase__ ) __lowercase = tf.constant( [[-0.2873, -1.9218, -0.3033], [-1.2710, -0.1338, -0.1902], [0.4095, 0.1214, -1.3121]] ) self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase__ , atol=4E-3 ) ) __lowercase = tf.function(lowerCAmelCase__ , jit_compile=lowerCAmelCase__ ) __lowercase = xla_generate(lowerCAmelCase__ , lowerCAmelCase__ )[0] self.assertTrue(np.allclose(output[:, :3, :3] , lowerCAmelCase__ , atol=4E-2 ) ) @require_tf @slow class _UpperCamelCase ( unittest.TestCase ): """simple docstring""" def _SCREAMING_SNAKE_CASE ( self ) -> str: '''simple docstring''' super().setUp() __lowercase = '''facebook/opt-350m''' def _SCREAMING_SNAKE_CASE ( self ) -> Any: '''simple docstring''' __lowercase = TFOPTForCausalLM.from_pretrained(self.path_model ) __lowercase = GPTaTokenizer.from_pretrained(self.path_model ) __lowercase = [ '''Today is a beautiful day and I want to''', '''In the city of''', '''Paris is the capital of France and''', '''Computers and mobile phones have taken''', ] # verify that prompt without BOS token is identical to Metaseq -> add_special_tokens=False __lowercase = tokenizer(lowerCAmelCase__ , return_tensors='''tf''' , padding=lowerCAmelCase__ , add_special_tokens=lowerCAmelCase__ ) __lowercase = tf.math.reduce_mean(model(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) __lowercase = tf.constant( [ [1.3851, -13.8923, -10.5229, -10.7533, -0.2309, -10.2384, -0.5365, -9.0947, -5.1670], [-4.7073, -10.6276, -3.9415, -21.5242, -0.2822, -0.2822, -0.2822, -0.2822, -0.2822], [0.6247, -3.4229, -8.9179, -1.4297, -14.1650, 1.4146, -9.0218, -0.2703, -0.2703], [6.4783, -1.9913, -10.7926, -2.3336, 1.5092, -0.9974, -6.8213, 1.3477, 1.3477], ] ) self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-4 ) ) __lowercase = tf.function(lowerCAmelCase__ , jit_compile=lowerCAmelCase__ ) __lowercase = tf.math.reduce_mean(xla_generate(inputs.input_ids , attention_mask=inputs.attention_mask )[0] , axis=-1 ) self.assertTrue(np.allclose(lowerCAmelCase__ , lowerCAmelCase__ , atol=1E-4 ) ) @require_tf @slow class _UpperCamelCase ( unittest.TestCase ): """simple docstring""" @property def _SCREAMING_SNAKE_CASE ( self ) -> Dict: '''simple docstring''' return [ "Today is a beautiful day and I want", "In the city of", "Paris is the capital of France and", "Computers and mobile phones have taken", ] def _SCREAMING_SNAKE_CASE ( self ) -> Optional[Any]: '''simple docstring''' __lowercase = '''facebook/opt-125m''' __lowercase = [ '''Today is a beautiful day and I want to''', '''In the city of New York, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowercase = [] __lowercase = GPTaTokenizer.from_pretrained(lowerCAmelCase__ ) __lowercase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase__ ) for prompt in self.prompts: __lowercase = tokenizer(lowerCAmelCase__ , return_tensors='''tf''' ).input_ids __lowercase = model.generate(lowerCAmelCase__ , max_length=10 ) __lowercase = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) def _SCREAMING_SNAKE_CASE ( self ) -> int: '''simple docstring''' __lowercase = '''facebook/opt-350m''' __lowercase = GPTaTokenizer.from_pretrained(lowerCAmelCase__ ) __lowercase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase__ ) __lowercase = '''left''' # use different length sentences to test batching __lowercase = [ '''Hello, my dog is a little''', '''Today, I''', ] __lowercase = tokenizer(lowerCAmelCase__ , return_tensors='''tf''' , padding=lowerCAmelCase__ ) __lowercase = inputs['''input_ids'''] __lowercase = model.generate(input_ids=lowerCAmelCase__ , attention_mask=inputs['''attention_mask'''] ) __lowercase = tokenizer(sentences[0] , return_tensors='''tf''' ).input_ids __lowercase = model.generate(input_ids=lowerCAmelCase__ ) __lowercase = inputs_non_padded.shape[-1] - tf.math.reduce_sum( tf.cast(inputs['''attention_mask'''][-1] , tf.intaa ) ) __lowercase = tokenizer(sentences[1] , return_tensors='''tf''' ).input_ids __lowercase = model.generate(input_ids=lowerCAmelCase__ , max_length=model.config.max_length - num_paddings ) __lowercase = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) __lowercase = tokenizer.decode(output_non_padded[0] , skip_special_tokens=lowerCAmelCase__ ) __lowercase = tokenizer.decode(output_padded[0] , skip_special_tokens=lowerCAmelCase__ ) __lowercase = [ '''Hello, my dog is a little bit of a dork.\nI\'m a little bit''', '''Today, I was in the middle of a conversation with a friend about the''', ] self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ ) self.assertListEqual(lowerCAmelCase__ , [non_padded_sentence, padded_sentence] ) def _SCREAMING_SNAKE_CASE ( self ) -> List[str]: '''simple docstring''' __lowercase = '''facebook/opt-350m''' __lowercase = [ '''Today is a beautiful day and I want to''', '''In the city of San Francisco, the city''', '''Paris is the capital of France and the capital''', '''Computers and mobile phones have taken over the''', ] __lowercase = [] __lowercase = GPTaTokenizer.from_pretrained(lowerCAmelCase__ ) __lowercase = TFOPTForCausalLM.from_pretrained(lowerCAmelCase__ ) for prompt in self.prompts: __lowercase = tokenizer(lowerCAmelCase__ , return_tensors='''tf''' ).input_ids __lowercase = model.generate(lowerCAmelCase__ , max_length=10 ) __lowercase = tokenizer.batch_decode(lowerCAmelCase__ , skip_special_tokens=lowerCAmelCase__ ) predicted_outputs += generated_string self.assertListEqual(lowerCAmelCase__ , lowerCAmelCase__ )
210
1
import unittest from transformers import MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING, AutoTokenizer, is_vision_available from transformers.pipelines import pipeline from transformers.pipelines.document_question_answering import apply_tesseract from transformers.testing_utils import ( is_pipeline_test, nested_simplify, require_detectrona, require_pytesseract, require_tf, require_torch, require_vision, slow, ) from .test_pipelines_common import ANY if is_vision_available(): from PIL import Image from transformers.image_utils import load_image else: class a : @staticmethod def _UpperCAmelCase ( *A_ , **A_ ): '''simple docstring''' pass def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: Dict ) -> Optional[Any]: return None # This is a pinned image from a specific revision of a document question answering space, hosted by HuggingFace, # so we can expect it to be available. SCREAMING_SNAKE_CASE_ = ( 'https://huggingface.co/spaces/impira/docquery/resolve/2f6c96314dc84dfda62d40de9da55f2f5165d403/invoice.png' ) @is_pipeline_test @require_torch @require_vision class a ( unittest.TestCase ): _lowercase = MODEL_FOR_DOCUMENT_QUESTION_ANSWERING_MAPPING @require_pytesseract @require_vision def _UpperCAmelCase ( self , A_ , A_ , A_ ): '''simple docstring''' _UpperCAmelCase : Optional[int] = pipeline( "document-question-answering" , model=A_ , tokenizer=A_ , image_processor=A_ ) _UpperCAmelCase : Optional[int] = INVOICE_URL _UpperCAmelCase : Tuple = list(zip(*apply_tesseract(load_image(A_ ) , A_ , "" ) ) ) _UpperCAmelCase : int = "What is the placebo?" _UpperCAmelCase : Optional[Any] = [ { "image": load_image(A_ ), "question": question, }, { "image": image, "question": question, }, { "image": image, "question": question, "word_boxes": word_boxes, }, ] return dqa_pipeline, examples def _UpperCAmelCase ( self , A_ , A_ ): '''simple docstring''' _UpperCAmelCase : Any = dqa_pipeline(A_ , top_k=2 ) self.assertEqual( A_ , [ [ {"score": ANY(A_ ), "answer": ANY(A_ ), "start": ANY(A_ ), "end": ANY(A_ )}, {"score": ANY(A_ ), "answer": ANY(A_ ), "start": ANY(A_ ), "end": ANY(A_ )}, ] ] * 3 , ) @require_torch @require_detectrona @require_pytesseract def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : List[Any] = pipeline("document-question-answering" , model="hf-internal-testing/tiny-random-layoutlmv2" ) _UpperCAmelCase : Any = INVOICE_URL _UpperCAmelCase : List[str] = "How many cats are there?" _UpperCAmelCase : Union[str, Any] = [ {"score": 0.00_01, "answer": "oy 2312/2019", "start": 38, "end": 39}, {"score": 0.00_01, "answer": "oy 2312/2019 DUE", "start": 38, "end": 40}, ] _UpperCAmelCase : Optional[int] = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual(nested_simplify(A_ , decimals=4 ) , A_ ) _UpperCAmelCase : Dict = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual(nested_simplify(A_ , decimals=4 ) , A_ ) # This image does not detect ANY text in it, meaning layoutlmv2 should fail. # Empty answer probably _UpperCAmelCase : Tuple = "./tests/fixtures/tests_samples/COCO/000000039769.png" _UpperCAmelCase : Union[str, Any] = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual(A_ , [] ) # We can optionnally pass directly the words and bounding boxes _UpperCAmelCase : Dict = "./tests/fixtures/tests_samples/COCO/000000039769.png" _UpperCAmelCase : str = [] _UpperCAmelCase : Union[str, Any] = [] _UpperCAmelCase : int = dqa_pipeline(image=A_ , question=A_ , words=A_ , boxes=A_ , top_k=2 ) self.assertEqual(A_ , [] ) @slow @require_torch @require_detectrona @require_pytesseract def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : str = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , ) _UpperCAmelCase : Any = INVOICE_URL _UpperCAmelCase : Dict = "What is the invoice number?" _UpperCAmelCase : Any = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_44, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.00_09, "answer": "us-001", "start": 16, "end": 16}, ] , ) _UpperCAmelCase : Optional[Any] = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_44, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.00_09, "answer": "us-001", "start": 16, "end": 16}, ] , ) _UpperCAmelCase : Optional[int] = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ [ {"score": 0.99_44, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.00_09, "answer": "us-001", "start": 16, "end": 16}, ], ] * 2 , ) @slow @require_torch @require_detectrona @require_pytesseract def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Optional[int] = pipeline( "document-question-answering" , model="tiennvcs/layoutlmv2-base-uncased-finetuned-docvqa" , revision="9977165" , max_seq_len=50 , ) _UpperCAmelCase : int = INVOICE_URL _UpperCAmelCase : List[Any] = "What is the invoice number?" _UpperCAmelCase : Union[str, Any] = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_74, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.99_48, "answer": "us-001", "start": 16, "end": 16}, ] , ) _UpperCAmelCase : Optional[int] = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_74, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.99_48, "answer": "us-001", "start": 16, "end": 16}, ] , ) _UpperCAmelCase : Union[str, Any] = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ [ {"score": 0.99_74, "answer": "1110212019", "start": 23, "end": 23}, {"score": 0.99_48, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) @slow @require_torch @require_pytesseract @require_vision def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : str = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=A_ ) _UpperCAmelCase : Any = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=A_ , revision="3dc6de3" , ) _UpperCAmelCase : Optional[int] = INVOICE_URL _UpperCAmelCase : Any = "What is the invoice number?" _UpperCAmelCase : List[Any] = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.42_51, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.08_19, "answer": "1110212019", "start": 23, "end": 23}, ] , ) _UpperCAmelCase : Dict = dqa_pipeline({"image": image, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.42_51, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.08_19, "answer": "1110212019", "start": 23, "end": 23}, ] , ) _UpperCAmelCase : Any = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ [ {"score": 0.42_51, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.08_19, "answer": "1110212019", "start": 23, "end": 23}, ] ] * 2 , ) _UpperCAmelCase : Any = list(zip(*apply_tesseract(load_image(A_ ) , A_ , "" ) ) ) # This model should also work if `image` is set to None _UpperCAmelCase : Tuple = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.42_51, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.08_19, "answer": "1110212019", "start": 23, "end": 23}, ] , ) @slow @require_torch @require_pytesseract @require_vision def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = AutoTokenizer.from_pretrained( "impira/layoutlm-document-qa" , revision="3dc6de3" , add_prefix_space=A_ ) _UpperCAmelCase : Optional[Any] = pipeline( "document-question-answering" , model="impira/layoutlm-document-qa" , tokenizer=A_ , revision="3dc6de3" , max_seq_len=50 , ) _UpperCAmelCase : Any = INVOICE_URL _UpperCAmelCase : Tuple = "What is the invoice number?" _UpperCAmelCase : Union[str, Any] = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_99, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.99_98, "answer": "us-001", "start": 16, "end": 16}, ] , ) _UpperCAmelCase : Any = dqa_pipeline( [{"image": image, "question": question}, {"image": image, "question": question}] , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ [ {"score": 0.99_99, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.99_98, "answer": "us-001", "start": 16, "end": 16}, ] ] * 2 , ) _UpperCAmelCase : Optional[int] = list(zip(*apply_tesseract(load_image(A_ ) , A_ , "" ) ) ) # This model should also work if `image` is set to None _UpperCAmelCase : Optional[Any] = dqa_pipeline({"image": None, "word_boxes": word_boxes, "question": question} , top_k=2 ) self.assertEqual( nested_simplify(A_ , decimals=4 ) , [ {"score": 0.99_99, "answer": "us-001", "start": 16, "end": 16}, {"score": 0.99_98, "answer": "us-001", "start": 16, "end": 16}, ] , ) @slow @require_torch def _UpperCAmelCase ( self ): '''simple docstring''' _UpperCAmelCase : Union[str, Any] = pipeline( "document-question-answering" , model="naver-clova-ix/donut-base-finetuned-docvqa" , tokenizer=AutoTokenizer.from_pretrained("naver-clova-ix/donut-base-finetuned-docvqa" ) , feature_extractor="naver-clova-ix/donut-base-finetuned-docvqa" , ) _UpperCAmelCase : Any = INVOICE_URL _UpperCAmelCase : List[str] = "What is the invoice number?" _UpperCAmelCase : int = dqa_pipeline(image=A_ , question=A_ , top_k=2 ) self.assertEqual(nested_simplify(A_ , decimals=4 ) , [{"answer": "us-001"}] ) @require_tf @unittest.skip("Document question answering not implemented in TF" ) def _UpperCAmelCase ( self ): '''simple docstring''' pass
365
import logging import os import random import sys from dataclasses import dataclass, field from typing import Optional import datasets import numpy as np import pandas as pd from datasets import load_dataset import transformers from transformers import ( AutoConfig, BartForSequenceClassification, DataCollatorWithPadding, EvalPrediction, HfArgumentParser, TapexTokenizer, Trainer, TrainingArguments, default_data_collator, set_seed, ) from transformers.trainer_utils import get_last_checkpoint from transformers.utils import check_min_version from transformers.utils.versions import require_version # Will error if the minimal version of Transformers is not installed. Remove at your own risks. check_min_version('4.17.0.dev0') require_version('datasets>=1.8.0', 'To fix: pip install -r examples/pytorch/text-classification/requirements.txt') SCREAMING_SNAKE_CASE_ = logging.getLogger(__name__) @dataclass class a : _lowercase = field( default="tab_fact" , metadata={"help": "The name of the dataset to use (via the datasets library)."} ) _lowercase = field( default="tab_fact" , metadata={"help": "The configuration name of the dataset to use (via the datasets library)."} , ) _lowercase = field( default=1_0_2_4 , metadata={ "help": ( "The maximum total input sequence length after tokenization. Sequences longer " "than this will be truncated, sequences shorter will be padded." ) } , ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "Overwrite the cached preprocessed datasets or not."} ) _lowercase = field( default=UpperCAmelCase , metadata={ "help": ( "Whether to pad all samples to `max_seq_length`. " "If False, will pad the samples dynamically when batching to the maximum length in the batch." ) } , ) _lowercase = field( default=UpperCAmelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of training examples to this " "value if set." ) } , ) _lowercase = field( default=UpperCAmelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of evaluation examples to this " "value if set." ) } , ) _lowercase = field( default=UpperCAmelCase , metadata={ "help": ( "For debugging purposes or quicker training, truncate the number of prediction examples to this " "value if set." ) } , ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "A csv or a json file containing the training data."} ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "A csv or a json file containing the validation data."} ) _lowercase = field(default=UpperCAmelCase , metadata={"help": "A csv or a json file containing the test data."} ) def _UpperCAmelCase ( self ): '''simple docstring''' if self.dataset_name is not None: pass elif self.train_file is None or self.validation_file is None: raise ValueError("Need either a GLUE task, a training/validation file or a dataset name." ) else: _UpperCAmelCase : int = self.train_file.split("." )[-1] assert train_extension in ["csv", "json"], "`train_file` should be a csv or a json file." _UpperCAmelCase : Optional[int] = self.validation_file.split("." )[-1] assert ( validation_extension == train_extension ), "`validation_file` should have the same extension (csv or json) as `train_file`." @dataclass class a : _lowercase = field( default=UpperCAmelCase , metadata={"help": "Path to pretrained model or model identifier from huggingface.co/models"} ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "Pretrained config name or path if not the same as model_name"} ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "Pretrained tokenizer name or path if not the same as model_name"} ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "Where do you want to store the pretrained models downloaded from huggingface.co"} , ) _lowercase = field( default=UpperCAmelCase , metadata={"help": "Whether to use one of the fast tokenizer (backed by the tokenizers library) or not."} , ) _lowercase = field( default="main" , metadata={"help": "The specific model version to use (can be a branch name, tag name or commit id)."} , ) _lowercase = field( default=UpperCAmelCase , metadata={ "help": ( "Will use the token generated when running `huggingface-cli login` (necessary to use this script " "with private models)." ) } , ) def __SCREAMING_SNAKE_CASE ( ) -> List[str]: # See all possible arguments in src/transformers/training_args.py # or by passing the --help flag to this script. # We now keep distinct sets of args, for a cleaner separation of concerns. _UpperCAmelCase : List[Any] = HfArgumentParser((ModelArguments, DataTrainingArguments, TrainingArguments) ) if len(sys.argv ) == 2 and sys.argv[1].endswith(".json" ): # If we pass only one argument to the script and it's the path to a json file, # let's parse it to get our arguments. _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = parser.parse_json_file(json_file=os.path.abspath(sys.argv[1] ) ) else: _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase : Optional[Any] = parser.parse_args_into_dataclasses() # Setup logging logging.basicConfig( format="%(asctime)s - %(levelname)s - %(name)s - %(message)s" , datefmt="%m/%d/%Y %H:%M:%S" , handlers=[logging.StreamHandler(sys.stdout )] , ) _UpperCAmelCase : Tuple = training_args.get_process_log_level() logger.setLevel(lowerCAmelCase ) datasets.utils.logging.set_verbosity(lowerCAmelCase ) transformers.utils.logging.set_verbosity(lowerCAmelCase ) transformers.utils.logging.enable_default_handler() transformers.utils.logging.enable_explicit_format() # Log on each process the small summary: logger.warning( F'Process rank: {training_args.local_rank}, device: {training_args.device}, n_gpu: {training_args.n_gpu}' + F'distributed training: {bool(training_args.local_rank != -1 )}, 16-bits training: {training_args.fpaa}' ) logger.info(F'Training/evaluation parameters {training_args}' ) # Detecting last checkpoint. _UpperCAmelCase : List[str] = None if os.path.isdir(training_args.output_dir ) and training_args.do_train and not training_args.overwrite_output_dir: _UpperCAmelCase : Tuple = get_last_checkpoint(training_args.output_dir ) if last_checkpoint is None and len(os.listdir(training_args.output_dir ) ) > 0: raise ValueError( F'Output directory ({training_args.output_dir}) already exists and is not empty. ' "Use --overwrite_output_dir to overcome." ) elif last_checkpoint is not None and training_args.resume_from_checkpoint is None: logger.info( F'Checkpoint detected, resuming training at {last_checkpoint}. To avoid this behavior, change ' "the `--output_dir` or add `--overwrite_output_dir` to train from scratch." ) # Set seed before initializing model. set_seed(training_args.seed ) # Get the datasets: you can either provide your own CSV/JSON training and evaluation files (see below) # or specify a GLUE benchmark task (the dataset will be downloaded automatically from the datasets Hub). # # For JSON files, this script will use the `question` column for the input question and `table` column for the corresponding table. # # If the CSVs/JSONs contain only one non-label column, the script does single sentence classification on this # single column. You can easily tweak this behavior (see below) # # In distributed training, the load_dataset function guarantee that only one local process can concurrently # download the dataset. if data_args.dataset_name is not None: # Downloading and loading a dataset from the hub. _UpperCAmelCase : Tuple = load_dataset( data_args.dataset_name , data_args.dataset_config_name , cache_dir=model_args.cache_dir ) else: # Loading a dataset from your local files. # CSV/JSON training and evaluation files are needed. _UpperCAmelCase : int = {"train": data_args.train_file, "validation": data_args.validation_file} # Get the test dataset: you can provide your own CSV/JSON test file (see below) # when you use `do_predict` without specifying a GLUE benchmark task. if training_args.do_predict: if data_args.test_file is not None: _UpperCAmelCase : Tuple = data_args.train_file.split("." )[-1] _UpperCAmelCase : Optional[Any] = data_args.test_file.split("." )[-1] assert ( test_extension == train_extension ), "`test_file` should have the same extension (csv or json) as `train_file`." _UpperCAmelCase : Union[str, Any] = data_args.test_file else: raise ValueError("Need either a GLUE task or a test file for `do_predict`." ) for key in data_files.keys(): logger.info(F'load a local file for {key}: {data_files[key]}' ) if data_args.train_file.endswith(".csv" ): # Loading a dataset from local csv files _UpperCAmelCase : List[str] = load_dataset("csv" , data_files=lowerCAmelCase , cache_dir=model_args.cache_dir ) else: # Loading a dataset from local json files _UpperCAmelCase : Dict = load_dataset("json" , data_files=lowerCAmelCase , cache_dir=model_args.cache_dir ) # See more about loading any type of standard or custom dataset at # https://huggingface.co/docs/datasets/loading_datasets.html. # Labels _UpperCAmelCase : List[str] = raw_datasets["train"].features["label"].names _UpperCAmelCase : Tuple = len(lowerCAmelCase ) # Load pretrained model and tokenizer # # In distributed training, the .from_pretrained methods guarantee that only one local process can concurrently # download model & vocab. _UpperCAmelCase : Union[str, Any] = AutoConfig.from_pretrained( model_args.config_name if model_args.config_name else model_args.model_name_or_path , num_labels=lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # load tapex tokenizer _UpperCAmelCase : Tuple = TapexTokenizer.from_pretrained( model_args.tokenizer_name if model_args.tokenizer_name else model_args.model_name_or_path , cache_dir=model_args.cache_dir , use_fast=model_args.use_fast_tokenizer , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , add_prefix_space=lowerCAmelCase , ) _UpperCAmelCase : Union[str, Any] = BartForSequenceClassification.from_pretrained( model_args.model_name_or_path , from_tf=bool(".ckpt" in model_args.model_name_or_path ) , config=lowerCAmelCase , cache_dir=model_args.cache_dir , revision=model_args.model_revision , use_auth_token=True if model_args.use_auth_token else None , ) # Padding strategy if data_args.pad_to_max_length: _UpperCAmelCase : Tuple = "max_length" else: # We will pad later, dynamically at batch creation, to the max sequence length in each batch _UpperCAmelCase : List[Any] = False # Some models have set the order of the labels to use, so let's make sure we do use it. _UpperCAmelCase : List[Any] = {"Refused": 0, "Entailed": 1} _UpperCAmelCase : Optional[int] = {0: "Refused", 1: "Entailed"} if data_args.max_seq_length > tokenizer.model_max_length: logger.warning( F'The max_seq_length passed ({data_args.max_seq_length}) is larger than the maximum length for the' F'model ({tokenizer.model_max_length}). Using max_seq_length={tokenizer.model_max_length}.' ) _UpperCAmelCase : int = min(data_args.max_seq_length , tokenizer.model_max_length ) def preprocess_tabfact_function(lowerCAmelCase: str ): # Tokenize the texts def _convert_table_text_to_pandas(lowerCAmelCase: List[Any] ): _UpperCAmelCase : Any = [_table_row.split("#" ) for _table_row in _table_text.strip("\n" ).split("\n" )] _UpperCAmelCase : List[str] = pd.DataFrame.from_records(_table_content[1:] , columns=_table_content[0] ) return _table_pd _UpperCAmelCase : Tuple = examples["statement"] _UpperCAmelCase : List[Any] = list(map(_convert_table_text_to_pandas , examples["table_text"] ) ) _UpperCAmelCase : Dict = tokenizer(lowerCAmelCase , lowerCAmelCase , padding=lowerCAmelCase , max_length=lowerCAmelCase , truncation=lowerCAmelCase ) _UpperCAmelCase : List[str] = examples["label"] return result with training_args.main_process_first(desc="dataset map pre-processing" ): _UpperCAmelCase : List[Any] = raw_datasets.map( lowerCAmelCase , batched=lowerCAmelCase , load_from_cache_file=not data_args.overwrite_cache , desc="Running tokenizer on dataset" , ) if training_args.do_train: if "train" not in raw_datasets: raise ValueError("--do_train requires a train dataset" ) _UpperCAmelCase : Dict = raw_datasets["train"] if data_args.max_train_samples is not None: _UpperCAmelCase : List[Any] = train_dataset.select(range(data_args.max_train_samples ) ) if training_args.do_eval: if "validation" not in raw_datasets and "validation_matched" not in raw_datasets: raise ValueError("--do_eval requires a validation dataset" ) _UpperCAmelCase : Union[str, Any] = raw_datasets["validation"] if data_args.max_eval_samples is not None: _UpperCAmelCase : Any = eval_dataset.select(range(data_args.max_eval_samples ) ) if training_args.do_predict or data_args.test_file is not None: if "test" not in raw_datasets and "test_matched" not in raw_datasets: raise ValueError("--do_predict requires a test dataset" ) _UpperCAmelCase : Dict = raw_datasets["test"] if data_args.max_predict_samples is not None: _UpperCAmelCase : Any = predict_dataset.select(range(data_args.max_predict_samples ) ) # Log a few random samples from the training set: if training_args.do_train: for index in random.sample(range(len(lowerCAmelCase ) ) , 3 ): logger.info(F'Sample {index} of the training set: {train_dataset[index]}.' ) # You can define your custom compute_metrics function. It takes an `EvalPrediction` object (a namedtuple with a # predictions and label_ids field) and has to return a dictionary string to float. def compute_metrics(lowerCAmelCase: EvalPrediction ): _UpperCAmelCase : Optional[int] = p.predictions[0] if isinstance(p.predictions , lowerCAmelCase ) else p.predictions _UpperCAmelCase : Optional[Any] = np.argmax(lowerCAmelCase , axis=1 ) return {"accuracy": (preds == p.label_ids).astype(np.floataa ).mean().item()} # Data collator will default to DataCollatorWithPadding, so we change it if we already did the padding. if data_args.pad_to_max_length: _UpperCAmelCase : str = default_data_collator elif training_args.fpaa: _UpperCAmelCase : int = DataCollatorWithPadding(lowerCAmelCase , pad_to_multiple_of=8 ) else: _UpperCAmelCase : List[str] = None # Initialize our Trainer _UpperCAmelCase : List[Any] = Trainer( model=lowerCAmelCase , args=lowerCAmelCase , train_dataset=train_dataset if training_args.do_train else None , eval_dataset=eval_dataset if training_args.do_eval else None , compute_metrics=lowerCAmelCase , tokenizer=lowerCAmelCase , data_collator=lowerCAmelCase , ) # Training if training_args.do_train: _UpperCAmelCase : List[Any] = None if training_args.resume_from_checkpoint is not None: _UpperCAmelCase : List[str] = training_args.resume_from_checkpoint elif last_checkpoint is not None: _UpperCAmelCase : Dict = last_checkpoint _UpperCAmelCase : str = trainer.train(resume_from_checkpoint=lowerCAmelCase ) _UpperCAmelCase : Tuple = train_result.metrics _UpperCAmelCase : Optional[int] = ( data_args.max_train_samples if data_args.max_train_samples is not None else len(lowerCAmelCase ) ) _UpperCAmelCase : Any = min(lowerCAmelCase , len(lowerCAmelCase ) ) trainer.save_model() # Saves the tokenizer too for easy upload trainer.log_metrics("train" , lowerCAmelCase ) trainer.save_metrics("train" , lowerCAmelCase ) trainer.save_state() # Evaluation if training_args.do_eval: logger.info("*** Evaluate ***" ) _UpperCAmelCase : Optional[int] = trainer.evaluate(eval_dataset=lowerCAmelCase ) _UpperCAmelCase : Optional[Any] = data_args.max_eval_samples if data_args.max_eval_samples is not None else len(lowerCAmelCase ) _UpperCAmelCase : Any = min(lowerCAmelCase , len(lowerCAmelCase ) ) trainer.log_metrics("eval" , lowerCAmelCase ) trainer.save_metrics("eval" , lowerCAmelCase ) if training_args.do_predict: logger.info("*** Predict ***" ) # Removing the `label` columns because it contains -1 and Trainer won't like that. _UpperCAmelCase : int = predict_dataset.remove_columns("label" ) _UpperCAmelCase : Any = trainer.predict(lowerCAmelCase , metric_key_prefix="predict" ).predictions _UpperCAmelCase : List[str] = np.argmax(lowerCAmelCase , axis=1 ) _UpperCAmelCase : int = os.path.join(training_args.output_dir , "predict_results_tabfact.txt" ) if trainer.is_world_process_zero(): with open(lowerCAmelCase , "w" ) as writer: logger.info("***** Predict Results *****" ) writer.write("index\tprediction\n" ) for index, item in enumerate(lowerCAmelCase ): _UpperCAmelCase : List[Any] = label_list[item] writer.write(F'{index}\t{item}\n' ) _UpperCAmelCase : int = {"finetuned_from": model_args.model_name_or_path, "tasks": "text-classification"} if training_args.push_to_hub: trainer.push_to_hub(**lowerCAmelCase ) else: trainer.create_model_card(**lowerCAmelCase ) def __SCREAMING_SNAKE_CASE ( lowerCAmelCase: Dict ) -> Union[str, Any]: # For xla_spawn (TPUs) main() if __name__ == "__main__": main()
189
0
'''simple docstring''' import math import time from typing import Dict, List, Optional from torch.utils.data import Dataset from transformers import SeqaSeqTrainer, is_torch_tpu_available from transformers.trainer_utils import PredictionOutput, speed_metrics if is_torch_tpu_available(check_device=False): import torch_xla.core.xla_model as xm import torch_xla.debug.metrics as met class lowerCAmelCase__ ( lowerCamelCase_ ): def __init__( self , *__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE=None , **__SCREAMING_SNAKE_CASE ): """simple docstring""" super().__init__(*__SCREAMING_SNAKE_CASE , **__SCREAMING_SNAKE_CASE ) lowercase_ : Tuple = eval_examples lowercase_ : Tuple = post_process_function def _snake_case ( self , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE = None , __SCREAMING_SNAKE_CASE = "eval" , **__SCREAMING_SNAKE_CASE , ): """simple docstring""" lowercase_ : Optional[int] = gen_kwargs.copy() lowercase_ : List[str] = ( gen_kwargs['''max_length'''] if gen_kwargs.get('''max_length''' ) is not None else self.args.generation_max_length ) lowercase_ : str = ( gen_kwargs['''num_beams'''] if gen_kwargs.get('''num_beams''' ) is not None else self.args.generation_num_beams ) lowercase_ : Dict = gen_kwargs lowercase_ : List[Any] = self.eval_dataset if eval_dataset is None else eval_dataset lowercase_ : List[str] = self.get_eval_dataloader(__SCREAMING_SNAKE_CASE ) lowercase_ : Optional[Any] = self.eval_examples if eval_examples is None else eval_examples # Temporarily disable metric computation, we will do it in the loop here. lowercase_ : Union[str, Any] = self.compute_metrics lowercase_ : Optional[int] = None lowercase_ : Tuple = time.time() lowercase_ : Tuple = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase_ : str = eval_loop( __SCREAMING_SNAKE_CASE , description='''Evaluation''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=__SCREAMING_SNAKE_CASE , metric_key_prefix=__SCREAMING_SNAKE_CASE , ) finally: lowercase_ : Any = compute_metrics lowercase_ : Any = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is not None and self.compute_metrics is not None and self.args.should_save: # Only the main node write the results by default lowercase_ : Optional[Any] = self.post_process_function(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE ) lowercase_ : Optional[int] = self.compute_metrics(__SCREAMING_SNAKE_CASE ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): lowercase_ : List[Any] = metrics.pop(__SCREAMING_SNAKE_CASE ) metrics.update(output.metrics ) else: lowercase_ : List[Any] = output.metrics if self.args.should_log: # Only the main node log the results by default self.log(__SCREAMING_SNAKE_CASE ) if self.args.tpu_metrics_debug or self.args.debug: # tpu-comment: Logging debug metrics for PyTorch/XLA (compile, execute times, ops, etc.) xm.master_print(met.metrics_report() ) lowercase_ : List[Any] = self.callback_handler.on_evaluate(self.args , self.state , self.control , __SCREAMING_SNAKE_CASE ) return metrics def _snake_case ( self , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE=None , __SCREAMING_SNAKE_CASE = "test" , **__SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase_ : Union[str, Any] = gen_kwargs.copy() lowercase_ : Tuple = self.get_test_dataloader(__SCREAMING_SNAKE_CASE ) # Temporarily disable metric computation, we will do it in the loop here. lowercase_ : Optional[Any] = self.compute_metrics lowercase_ : Optional[int] = None lowercase_ : List[Any] = time.time() lowercase_ : int = self.prediction_loop if self.args.use_legacy_prediction_loop else self.evaluation_loop try: lowercase_ : Tuple = eval_loop( __SCREAMING_SNAKE_CASE , description='''Prediction''' , prediction_loss_only=True if compute_metrics is None else None , ignore_keys=__SCREAMING_SNAKE_CASE , metric_key_prefix=__SCREAMING_SNAKE_CASE , ) finally: lowercase_ : Any = compute_metrics lowercase_ : Tuple = self.args.eval_batch_size * self.args.world_size if F'''{metric_key_prefix}_jit_compilation_time''' in output.metrics: start_time += output.metrics[F'''{metric_key_prefix}_jit_compilation_time'''] output.metrics.update( speed_metrics( __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , num_samples=output.num_samples , num_steps=math.ceil(output.num_samples / total_batch_size ) , ) ) if self.post_process_function is None or self.compute_metrics is None: return output lowercase_ : Any = self.post_process_function(__SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , __SCREAMING_SNAKE_CASE , '''predict''' ) lowercase_ : str = self.compute_metrics(__SCREAMING_SNAKE_CASE ) # Prefix all keys with metric_key_prefix + '_' for key in list(metrics.keys() ): if not key.startswith(F'''{metric_key_prefix}_''' ): lowercase_ : Optional[int] = metrics.pop(__SCREAMING_SNAKE_CASE ) metrics.update(output.metrics ) return PredictionOutput(predictions=predictions.predictions , label_ids=predictions.label_ids , metrics=__SCREAMING_SNAKE_CASE )
93
import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinvaConfig, SwinvaForImageClassification def UpperCamelCase( __UpperCamelCase : List[str] ): lowerCAmelCase_ : List[str] = SwinvaConfig() lowerCAmelCase_ : List[str] = swinva_name.split('''_''' ) lowerCAmelCase_ : str = name_split[1] if "to" in name_split[3]: lowerCAmelCase_ : List[Any] = int(name_split[3][-3:] ) else: lowerCAmelCase_ : List[Any] = int(name_split[3] ) if "to" in name_split[2]: lowerCAmelCase_ : List[str] = int(name_split[2][-2:] ) else: lowerCAmelCase_ : int = int(name_split[2][6:] ) if model_size == "tiny": lowerCAmelCase_ : Any = 96 lowerCAmelCase_ : List[str] = (2, 2, 6, 2) lowerCAmelCase_ : Union[str, Any] = (3, 6, 12, 24) elif model_size == "small": lowerCAmelCase_ : List[str] = 96 lowerCAmelCase_ : Any = (2, 2, 18, 2) lowerCAmelCase_ : Dict = (3, 6, 12, 24) elif model_size == "base": lowerCAmelCase_ : Union[str, Any] = 128 lowerCAmelCase_ : List[Any] = (2, 2, 18, 2) lowerCAmelCase_ : Tuple = (4, 8, 16, 32) else: lowerCAmelCase_ : Optional[Any] = 192 lowerCAmelCase_ : List[Any] = (2, 2, 18, 2) lowerCAmelCase_ : List[Any] = (6, 12, 24, 48) if "to" in swinva_name: lowerCAmelCase_ : Union[str, Any] = (12, 12, 12, 6) if ("22k" in swinva_name) and ("to" not in swinva_name): lowerCAmelCase_ : Optional[int] = 21841 lowerCAmelCase_ : Any = '''huggingface/label-files''' lowerCAmelCase_ : Tuple = '''imagenet-22k-id2label.json''' lowerCAmelCase_ : Any = json.load(open(hf_hub_download(__UpperCamelCase ,__UpperCamelCase ,repo_type='''dataset''' ) ,'''r''' ) ) lowerCAmelCase_ : Optional[Any] = {int(__UpperCamelCase ): v for k, v in idalabel.items()} lowerCAmelCase_ : str = idalabel lowerCAmelCase_ : List[str] = {v: k for k, v in idalabel.items()} else: lowerCAmelCase_ : Optional[int] = 1000 lowerCAmelCase_ : Tuple = '''huggingface/label-files''' lowerCAmelCase_ : Union[str, Any] = '''imagenet-1k-id2label.json''' lowerCAmelCase_ : Dict = json.load(open(hf_hub_download(__UpperCamelCase ,__UpperCamelCase ,repo_type='''dataset''' ) ,'''r''' ) ) lowerCAmelCase_ : int = {int(__UpperCamelCase ): v for k, v in idalabel.items()} lowerCAmelCase_ : List[str] = idalabel lowerCAmelCase_ : Optional[int] = {v: k for k, v in idalabel.items()} lowerCAmelCase_ : Optional[int] = img_size lowerCAmelCase_ : Dict = num_classes lowerCAmelCase_ : Dict = embed_dim lowerCAmelCase_ : Optional[Any] = depths lowerCAmelCase_ : Optional[int] = num_heads lowerCAmelCase_ : Dict = window_size return config def UpperCamelCase( __UpperCamelCase : List[str] ): if "patch_embed.proj" in name: lowerCAmelCase_ : Dict = name.replace('''patch_embed.proj''' ,'''embeddings.patch_embeddings.projection''' ) if "patch_embed.norm" in name: lowerCAmelCase_ : List[Any] = name.replace('''patch_embed.norm''' ,'''embeddings.norm''' ) if "layers" in name: lowerCAmelCase_ : int = '''encoder.''' + name if "attn.proj" in name: lowerCAmelCase_ : Union[str, Any] = name.replace('''attn.proj''' ,'''attention.output.dense''' ) if "attn" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''attn''' ,'''attention.self''' ) if "norm1" in name: lowerCAmelCase_ : Union[str, Any] = name.replace('''norm1''' ,'''layernorm_before''' ) if "norm2" in name: lowerCAmelCase_ : Tuple = name.replace('''norm2''' ,'''layernorm_after''' ) if "mlp.fc1" in name: lowerCAmelCase_ : Optional[Any] = name.replace('''mlp.fc1''' ,'''intermediate.dense''' ) if "mlp.fc2" in name: lowerCAmelCase_ : Tuple = name.replace('''mlp.fc2''' ,'''output.dense''' ) if "q_bias" in name: lowerCAmelCase_ : Tuple = name.replace('''q_bias''' ,'''query.bias''' ) if "k_bias" in name: lowerCAmelCase_ : Tuple = name.replace('''k_bias''' ,'''key.bias''' ) if "v_bias" in name: lowerCAmelCase_ : int = name.replace('''v_bias''' ,'''value.bias''' ) if "cpb_mlp" in name: lowerCAmelCase_ : Any = name.replace('''cpb_mlp''' ,'''continuous_position_bias_mlp''' ) if name == "norm.weight": lowerCAmelCase_ : Dict = '''layernorm.weight''' if name == "norm.bias": lowerCAmelCase_ : Any = '''layernorm.bias''' if "head" in name: lowerCAmelCase_ : int = name.replace('''head''' ,'''classifier''' ) else: lowerCAmelCase_ : Union[str, Any] = '''swinv2.''' + name return name def UpperCamelCase( __UpperCamelCase : Any ,__UpperCamelCase : Optional[int] ): for key in orig_state_dict.copy().keys(): lowerCAmelCase_ : Optional[int] = orig_state_dict.pop(__UpperCamelCase ) if "mask" in key: continue elif "qkv" in key: lowerCAmelCase_ : Dict = key.split('''.''' ) lowerCAmelCase_ : Any = int(key_split[1] ) lowerCAmelCase_ : Optional[int] = int(key_split[3] ) lowerCAmelCase_ : Dict = model.swinva.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: lowerCAmelCase_ : Optional[Any] = val[:dim, :] lowerCAmelCase_ : Any = val[dim : dim * 2, :] lowerCAmelCase_ : List[Any] = val[-dim:, :] else: lowerCAmelCase_ : Dict = val[:dim] lowerCAmelCase_ : Union[str, Any] = val[ dim : dim * 2 ] lowerCAmelCase_ : Dict = val[-dim:] else: lowerCAmelCase_ : Optional[Any] = val return orig_state_dict def UpperCamelCase( __UpperCamelCase : int ,__UpperCamelCase : Dict ): lowerCAmelCase_ : Optional[Any] = timm.create_model(__UpperCamelCase ,pretrained=__UpperCamelCase ) timm_model.eval() lowerCAmelCase_ : List[str] = get_swinva_config(__UpperCamelCase ) lowerCAmelCase_ : Union[str, Any] = SwinvaForImageClassification(__UpperCamelCase ) model.eval() lowerCAmelCase_ : str = convert_state_dict(timm_model.state_dict() ,__UpperCamelCase ) model.load_state_dict(__UpperCamelCase ) lowerCAmelCase_ : List[Any] = '''http://images.cocodataset.org/val2017/000000039769.jpg''' lowerCAmelCase_ : Tuple = AutoImageProcessor.from_pretrained('''microsoft/{}'''.format(swinva_name.replace('''_''' ,'''-''' ) ) ) lowerCAmelCase_ : Union[str, Any] = Image.open(requests.get(__UpperCamelCase ,stream=__UpperCamelCase ).raw ) lowerCAmelCase_ : Optional[Any] = image_processor(images=__UpperCamelCase ,return_tensors='''pt''' ) lowerCAmelCase_ : List[str] = timm_model(inputs['''pixel_values'''] ) lowerCAmelCase_ : Union[str, Any] = model(**__UpperCamelCase ).logits assert torch.allclose(__UpperCamelCase ,__UpperCamelCase ,atol=1e-3 ) print(f"""Saving model {swinva_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__UpperCamelCase ) print(f"""Saving image processor to {pytorch_dump_folder_path}""" ) image_processor.save_pretrained(__UpperCamelCase ) model.push_to_hub( repo_path_or_name=Path(__UpperCamelCase ,__UpperCamelCase ) ,organization='''nandwalritik''' ,commit_message='''Add model''' ,) if __name__ == "__main__": A__ : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swinv2_name''', default='''swinv2_tiny_patch4_window8_256''', type=str, help='''Name of the Swinv2 timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A__ : Optional[Any] = parser.parse_args() convert_swinva_checkpoint(args.swinva_name, args.pytorch_dump_folder_path)
103
0
'''simple docstring''' import gc import math import unittest import torch from diffusers import UNetaDModel from diffusers.utils import floats_tensor, logging, slow, torch_all_close, torch_device from diffusers.utils.testing_utils import enable_full_determinism from .test_modeling_common import ModelTesterMixin, UNetTesterMixin UpperCAmelCase = logging.get_logger(__name__) enable_full_determinism() class __snake_case( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : int = UNetaDModel UpperCAmelCase : Optional[int] = "sample" @property def __snake_case ( self ) -> List[str]: lowerCAmelCase = 4 lowerCAmelCase = 3 lowerCAmelCase = (32, 32) lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) lowerCAmelCase = torch.tensor([10] ).to(A_ ) return {"sample": noise, "timestep": time_step} @property def __snake_case ( self ) -> List[str]: return (3, 32, 32) @property def __snake_case ( self ) -> Dict: return (3, 32, 32) def __snake_case ( self ) -> str: lowerCAmelCase = { """block_out_channels""": (32, 64), """down_block_types""": ("""DownBlock2D""", """AttnDownBlock2D"""), """up_block_types""": ("""AttnUpBlock2D""", """UpBlock2D"""), """attention_head_dim""": 3, """out_channels""": 3, """in_channels""": 3, """layers_per_block""": 2, """sample_size""": 32, } lowerCAmelCase = self.dummy_input return init_dict, inputs_dict class __snake_case( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : str = UNetaDModel UpperCAmelCase : Optional[int] = "sample" @property def __snake_case ( self ) -> Union[str, Any]: lowerCAmelCase = 4 lowerCAmelCase = 4 lowerCAmelCase = (32, 32) lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) lowerCAmelCase = torch.tensor([10] ).to(A_ ) return {"sample": noise, "timestep": time_step} @property def __snake_case ( self ) -> int: return (4, 32, 32) @property def __snake_case ( self ) -> List[str]: return (4, 32, 32) def __snake_case ( self ) -> List[Any]: lowerCAmelCase = { """sample_size""": 32, """in_channels""": 4, """out_channels""": 4, """layers_per_block""": 2, """block_out_channels""": (32, 64), """attention_head_dim""": 32, """down_block_types""": ("""DownBlock2D""", """DownBlock2D"""), """up_block_types""": ("""UpBlock2D""", """UpBlock2D"""), } lowerCAmelCase = self.dummy_input return init_dict, inputs_dict def __snake_case ( self ) -> List[str]: lowerCAmelCase, lowerCAmelCase = UNetaDModel.from_pretrained("""fusing/unet-ldm-dummy-update""" , output_loading_info=A_ ) self.assertIsNotNone(A_ ) self.assertEqual(len(loading_info["""missing_keys"""] ) , 0 ) model.to(A_ ) lowerCAmelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != """cuda""" , """This test is supposed to run on GPU""" ) def __snake_case ( self ) -> Dict: lowerCAmelCase, lowerCAmelCase = UNetaDModel.from_pretrained("""fusing/unet-ldm-dummy-update""" , output_loading_info=A_ ) model.to(A_ ) lowerCAmelCase = model(**self.dummy_input ).sample assert image is not None, "Make sure output is not None" @unittest.skipIf(torch_device != """cuda""" , """This test is supposed to run on GPU""" ) def __snake_case ( self ) -> Any: # by defautl model loading will use accelerate as `low_cpu_mem_usage=True` lowerCAmelCase, lowerCAmelCase = UNetaDModel.from_pretrained("""fusing/unet-ldm-dummy-update""" , output_loading_info=A_ ) model_accelerate.to(A_ ) model_accelerate.eval() lowerCAmelCase = torch.randn( 1 , model_accelerate.config.in_channels , model_accelerate.config.sample_size , model_accelerate.config.sample_size , generator=torch.manual_seed(0 ) , ) lowerCAmelCase = noise.to(A_ ) lowerCAmelCase = torch.tensor([10] * noise.shape[0] ).to(A_ ) lowerCAmelCase = model_accelerate(A_ , A_ )["""sample"""] # two models don't need to stay in the device at the same time del model_accelerate torch.cuda.empty_cache() gc.collect() lowerCAmelCase, lowerCAmelCase = UNetaDModel.from_pretrained( """fusing/unet-ldm-dummy-update""" , output_loading_info=A_ , low_cpu_mem_usage=A_ ) model_normal_load.to(A_ ) model_normal_load.eval() lowerCAmelCase = model_normal_load(A_ , A_ )["""sample"""] assert torch_all_close(A_ , A_ , rtol=1e-3 ) def __snake_case ( self ) -> Any: lowerCAmelCase = UNetaDModel.from_pretrained("""fusing/unet-ldm-dummy-update""" ) model.eval() model.to(A_ ) lowerCAmelCase = torch.randn( 1 , model.config.in_channels , model.config.sample_size , model.config.sample_size , generator=torch.manual_seed(0 ) , ) lowerCAmelCase = noise.to(A_ ) lowerCAmelCase = torch.tensor([10] * noise.shape[0] ).to(A_ ) with torch.no_grad(): lowerCAmelCase = model(A_ , A_ ).sample lowerCAmelCase = output[0, -1, -3:, -3:].flatten().cpu() # fmt: off lowerCAmelCase = torch.tensor([-13.3258, -20.1100, -15.9873, -17.6617, -23.0596, -17.9419, -13.3675, -16.1889, -12.3800] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1e-3 ) ) class __snake_case( _lowerCAmelCase , _lowerCAmelCase , unittest.TestCase ): '''simple docstring''' UpperCAmelCase : List[Any] = UNetaDModel UpperCAmelCase : Optional[Any] = "sample" @property def __snake_case ( self , A_=(32, 32) ) -> int: lowerCAmelCase = 4 lowerCAmelCase = 3 lowerCAmelCase = floats_tensor((batch_size, num_channels) + sizes ).to(A_ ) lowerCAmelCase = torch.tensor(batch_size * [10] ).to(dtype=torch.intaa , device=A_ ) return {"sample": noise, "timestep": time_step} @property def __snake_case ( self ) -> Union[str, Any]: return (3, 32, 32) @property def __snake_case ( self ) -> str: return (3, 32, 32) def __snake_case ( self ) -> Dict: lowerCAmelCase = { """block_out_channels""": [32, 64, 64, 64], """in_channels""": 3, """layers_per_block""": 1, """out_channels""": 3, """time_embedding_type""": """fourier""", """norm_eps""": 1e-6, """mid_block_scale_factor""": math.sqrt(2.0 ), """norm_num_groups""": None, """down_block_types""": [ """SkipDownBlock2D""", """AttnSkipDownBlock2D""", """SkipDownBlock2D""", """SkipDownBlock2D""", ], """up_block_types""": [ """SkipUpBlock2D""", """SkipUpBlock2D""", """AttnSkipUpBlock2D""", """SkipUpBlock2D""", ], } lowerCAmelCase = self.dummy_input return init_dict, inputs_dict @slow def __snake_case ( self ) -> Optional[Any]: lowerCAmelCase, lowerCAmelCase = UNetaDModel.from_pretrained("""google/ncsnpp-celebahq-256""" , output_loading_info=A_ ) self.assertIsNotNone(A_ ) self.assertEqual(len(loading_info["""missing_keys"""] ) , 0 ) model.to(A_ ) lowerCAmelCase = self.dummy_input lowerCAmelCase = floats_tensor((4, 3) + (256, 256) ).to(A_ ) lowerCAmelCase = noise lowerCAmelCase = model(**A_ ) assert image is not None, "Make sure output is not None" @slow def __snake_case ( self ) -> Optional[int]: lowerCAmelCase = UNetaDModel.from_pretrained("""google/ncsnpp-celebahq-256""" ) model.to(A_ ) lowerCAmelCase = 4 lowerCAmelCase = 3 lowerCAmelCase = (256, 256) lowerCAmelCase = torch.ones((batch_size, num_channels) + sizes ).to(A_ ) lowerCAmelCase = torch.tensor(batch_size * [1e-4] ).to(A_ ) with torch.no_grad(): lowerCAmelCase = model(A_ , A_ ).sample lowerCAmelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off lowerCAmelCase = torch.tensor([-4842.8691, -6499.6631, -3800.1953, -7978.2686, -1_0980.7129, -2_0028.8535, 8148.2822, 2342.2905, 567.7608] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1e-2 ) ) def __snake_case ( self ) -> List[Any]: lowerCAmelCase = UNetaDModel.from_pretrained("""fusing/ncsnpp-ffhq-ve-dummy-update""" ) model.to(A_ ) lowerCAmelCase = 4 lowerCAmelCase = 3 lowerCAmelCase = (32, 32) lowerCAmelCase = torch.ones((batch_size, num_channels) + sizes ).to(A_ ) lowerCAmelCase = torch.tensor(batch_size * [1e-4] ).to(A_ ) with torch.no_grad(): lowerCAmelCase = model(A_ , A_ ).sample lowerCAmelCase = output[0, -3:, -3:, -1].flatten().cpu() # fmt: off lowerCAmelCase = torch.tensor([-0.0_3_2_5, -0.0_9_0_0, -0.0_8_6_9, -0.0_3_3_2, -0.0_7_2_5, -0.0_2_7_0, -0.0_1_0_1, 0.0_2_2_7, 0.0_2_5_6] ) # fmt: on self.assertTrue(torch_all_close(A_ , A_ , rtol=1e-2 ) ) def __snake_case ( self ) -> int: # not required for this model pass
358
'''simple docstring''' from torch import nn def _snake_case ( _SCREAMING_SNAKE_CASE : Optional[int] ) -> Union[str, Any]: """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'Unsupported activation function: {act_fn}' )
187
0
import inspect from typing import Optional, Union import numpy as np import PIL import torch from torch.nn import functional as F from torchvision import transforms from transformers import CLIPFeatureExtractor, CLIPModel, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DiffusionPipeline, DPMSolverMultistepScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.pipelines.stable_diffusion.pipeline_stable_diffusion import StableDiffusionPipelineOutput from diffusers.utils import ( PIL_INTERPOLATION, randn_tensor, ) def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase ) -> Dict: """simple docstring""" if isinstance(_lowerCAmelCase , torch.Tensor ): return image elif isinstance(_lowerCAmelCase , PIL.Image.Image ): A : Optional[Any] = [image] if isinstance(image[0] , PIL.Image.Image ): A : Any = [np.array(i.resize((w, h) , resample=PIL_INTERPOLATION["""lanczos"""] ) )[None, :] for i in image] A : Tuple = np.concatenate(_lowerCAmelCase , axis=0 ) A : int = np.array(_lowerCAmelCase ).astype(np.floataa ) / 255.0 A : Union[str, Any] = image.transpose(0 , 3 , 1 , 2 ) A : Optional[int] = 2.0 * image - 1.0 A : Tuple = torch.from_numpy(_lowerCAmelCase ) elif isinstance(image[0] , torch.Tensor ): A : Optional[int] = torch.cat(_lowerCAmelCase , dim=0 ) return image def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase , _lowerCAmelCase=0.9_995 ) -> Optional[int]: """simple docstring""" if not isinstance(_lowerCAmelCase , np.ndarray ): A : Tuple = True A : Optional[int] = va.device A : Tuple = va.cpu().numpy() A : List[str] = va.cpu().numpy() A : Union[str, Any] = np.sum(va * va / (np.linalg.norm(_lowerCAmelCase ) * np.linalg.norm(_lowerCAmelCase )) ) if np.abs(_lowerCAmelCase ) > DOT_THRESHOLD: A : List[Any] = (1 - t) * va + t * va else: A : Optional[Any] = np.arccos(_lowerCAmelCase ) A : Dict = np.sin(_lowerCAmelCase ) A : Union[str, Any] = theta_a * t A : List[Any] = np.sin(_lowerCAmelCase ) A : Optional[Any] = np.sin(theta_a - theta_t ) / sin_theta_a A : int = sin_theta_t / sin_theta_a A : Optional[Any] = sa * va + sa * va if inputs_are_torch: A : List[Any] = torch.from_numpy(_lowerCAmelCase ).to(_lowerCAmelCase ) return va def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> Optional[Any]: """simple docstring""" A : List[Any] = F.normalize(_lowerCAmelCase , dim=-1 ) A : Optional[int] = F.normalize(_lowerCAmelCase , dim=-1 ) return (x - y).norm(dim=-1 ).div(2 ).arcsin().pow(2 ).mul(2 ) def __UpperCamelCase ( _lowerCAmelCase , _lowerCAmelCase ) -> Dict: """simple docstring""" for param in model.parameters(): A : Any = value class SCREAMING_SNAKE_CASE__ ( SCREAMING_SNAKE_CASE__ ): '''simple docstring''' def __init__( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__=None, lowerCamelCase__=None, lowerCamelCase__=None, ): super().__init__() self.register_modules( vae=lowerCamelCase__, text_encoder=lowerCamelCase__, clip_model=lowerCamelCase__, tokenizer=lowerCamelCase__, unet=lowerCamelCase__, scheduler=lowerCamelCase__, feature_extractor=lowerCamelCase__, coca_model=lowerCamelCase__, coca_tokenizer=lowerCamelCase__, coca_transform=lowerCamelCase__, ) A : Union[str, Any] = ( feature_extractor.size if isinstance(feature_extractor.size, lowerCamelCase__ ) else feature_extractor.size["""shortest_edge"""] ) A : Optional[Any] = transforms.Normalize(mean=feature_extractor.image_mean, std=feature_extractor.image_std ) set_requires_grad(self.text_encoder, lowerCamelCase__ ) set_requires_grad(self.clip_model, lowerCamelCase__ ) def _lowerCAmelCase ( self, lowerCamelCase__ = "auto" ): if slice_size == "auto": # half the attention head size is usually a good trade-off between # speed and memory A : Optional[Any] = self.unet.config.attention_head_dim // 2 self.unet.set_attention_slice(lowerCamelCase__ ) def _lowerCAmelCase ( self ): self.enable_attention_slicing(lowerCamelCase__ ) def _lowerCAmelCase ( self ): set_requires_grad(self.vae, lowerCamelCase__ ) def _lowerCAmelCase ( self ): set_requires_grad(self.vae, lowerCamelCase__ ) def _lowerCAmelCase ( self ): set_requires_grad(self.unet, lowerCamelCase__ ) def _lowerCAmelCase ( self ): set_requires_grad(self.unet, lowerCamelCase__ ) def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ): # get the original timestep using init_timestep A : Dict = min(int(num_inference_steps * strength ), lowerCamelCase__ ) A : Union[str, Any] = max(num_inference_steps - init_timestep, 0 ) A : List[Any] = self.scheduler.timesteps[t_start:] return timesteps, num_inference_steps - t_start def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__=None ): if not isinstance(lowerCamelCase__, torch.Tensor ): raise ValueError(f'''`image` has to be of type `torch.Tensor` but is {type(lowerCamelCase__ )}''' ) A : List[Any] = image.to(device=lowerCamelCase__, dtype=lowerCamelCase__ ) if isinstance(lowerCamelCase__, lowerCamelCase__ ): A : List[Any] = [ self.vae.encode(image[i : i + 1] ).latent_dist.sample(generator[i] ) for i in range(lowerCamelCase__ ) ] A : int = torch.cat(lowerCamelCase__, dim=0 ) else: A : Union[str, Any] = self.vae.encode(lowerCamelCase__ ).latent_dist.sample(lowerCamelCase__ ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor A : Tuple = 0.1_8215 * init_latents A : Tuple = init_latents.repeat_interleave(lowerCamelCase__, dim=0 ) A : Union[str, Any] = randn_tensor(init_latents.shape, generator=lowerCamelCase__, device=lowerCamelCase__, dtype=lowerCamelCase__ ) # get latents A : List[Any] = self.scheduler.add_noise(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) A : Dict = init_latents return latents def _lowerCAmelCase ( self, lowerCamelCase__ ): A : Union[str, Any] = self.coca_transform(lowerCamelCase__ ).unsqueeze(0 ) with torch.no_grad(), torch.cuda.amp.autocast(): A : Union[str, Any] = self.coca_model.generate(transformed_image.to(device=self.device, dtype=self.coca_model.dtype ) ) A : Tuple = self.coca_tokenizer.decode(generated[0].cpu().numpy() ) return generated.split("""<end_of_text>""" )[0].replace("""<start_of_text>""", """""" ).rstrip(""" .,""" ) def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__ ): A : List[str] = self.feature_extractor.preprocess(lowerCamelCase__ ) A : int = torch.from_numpy(clip_image_input["""pixel_values"""][0] ).unsqueeze(0 ).to(self.device ).half() A : Any = self.clip_model.get_image_features(lowerCamelCase__ ) A : Union[str, Any] = image_embeddings_clip / image_embeddings_clip.norm(p=2, dim=-1, keepdim=lowerCamelCase__ ) A : Optional[int] = image_embeddings_clip.repeat_interleave(lowerCamelCase__, dim=0 ) return image_embeddings_clip @torch.enable_grad() def _lowerCAmelCase ( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, ): A : int = latents.detach().requires_grad_() A : Any = self.scheduler.scale_model_input(lowerCamelCase__, lowerCamelCase__ ) # predict the noise residual A : Union[str, Any] = self.unet(lowerCamelCase__, lowerCamelCase__, encoder_hidden_states=lowerCamelCase__ ).sample if isinstance(self.scheduler, (PNDMScheduler, DDIMScheduler, DPMSolverMultistepScheduler) ): A : Tuple = self.scheduler.alphas_cumprod[timestep] A : Any = 1 - alpha_prod_t # compute predicted original sample from predicted noise also called # "predicted x_0" of formula (12) from https://arxiv.org/pdf/2010.02502.pdf A : Union[str, Any] = (latents - beta_prod_t ** 0.5 * noise_pred) / alpha_prod_t ** 0.5 A : Tuple = torch.sqrt(lowerCamelCase__ ) A : Tuple = pred_original_sample * (fac) + latents * (1 - fac) elif isinstance(self.scheduler, lowerCamelCase__ ): A : Optional[int] = self.scheduler.sigmas[index] A : Any = latents - sigma * noise_pred else: raise ValueError(f'''scheduler type {type(self.scheduler )} not supported''' ) # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor A : List[str] = 1 / 0.1_8215 * sample A : Dict = self.vae.decode(lowerCamelCase__ ).sample A : List[str] = (image / 2 + 0.5).clamp(0, 1 ) A : List[Any] = transforms.Resize(self.feature_extractor_size )(lowerCamelCase__ ) A : Optional[Any] = self.normalize(lowerCamelCase__ ).to(latents.dtype ) A : int = self.clip_model.get_image_features(lowerCamelCase__ ) A : str = image_embeddings_clip / image_embeddings_clip.norm(p=2, dim=-1, keepdim=lowerCamelCase__ ) A : List[Any] = spherical_dist_loss(lowerCamelCase__, lowerCamelCase__ ).mean() * clip_guidance_scale A : int = -torch.autograd.grad(lowerCamelCase__, lowerCamelCase__ )[0] if isinstance(self.scheduler, lowerCamelCase__ ): A : int = latents.detach() + grads * (sigma**2) A : Tuple = noise_pred_original else: A : Optional[Any] = noise_pred_original - torch.sqrt(lowerCamelCase__ ) * grads return noise_pred, latents @torch.no_grad() def __call__( self, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ = None, lowerCamelCase__ = None, lowerCamelCase__ = 512, lowerCamelCase__ = 512, lowerCamelCase__ = 0.6, lowerCamelCase__ = 50, lowerCamelCase__ = 7.5, lowerCamelCase__ = 1, lowerCamelCase__ = 0.0, lowerCamelCase__ = 100, lowerCamelCase__ = None, lowerCamelCase__ = "pil", lowerCamelCase__ = True, lowerCamelCase__ = 0.8, lowerCamelCase__ = 0.1, lowerCamelCase__ = 0.1, ): if isinstance(lowerCamelCase__, lowerCamelCase__ ) and len(lowerCamelCase__ ) != batch_size: raise ValueError(f'''You have passed {batch_size} batch_size, but only {len(lowerCamelCase__ )} generators.''' ) if height % 8 != 0 or width % 8 != 0: raise ValueError(f'''`height` and `width` have to be divisible by 8 but are {height} and {width}.''' ) if isinstance(lowerCamelCase__, torch.Generator ) and batch_size > 1: A : Union[str, Any] = [generator] + [None] * (batch_size - 1) A : List[Any] = [ ("""model""", self.coca_model is None), ("""tokenizer""", self.coca_tokenizer is None), ("""transform""", self.coca_transform is None), ] A : List[Any] = [x[0] for x in coca_is_none if x[1]] A : List[Any] = """, """.join(lowerCamelCase__ ) # generate prompts with coca model if prompt is None if content_prompt is None: if len(lowerCamelCase__ ): raise ValueError( f'''Content prompt is None and CoCa [{coca_is_none_str}] is None.''' f'''Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) A : Any = self.get_image_description(lowerCamelCase__ ) if style_prompt is None: if len(lowerCamelCase__ ): raise ValueError( f'''Style prompt is None and CoCa [{coca_is_none_str}] is None.''' f''' Set prompt or pass Coca [{coca_is_none_str}] to DiffusionPipeline.''' ) A : Optional[Any] = self.get_image_description(lowerCamelCase__ ) # get prompt text embeddings for content and style A : Any = self.tokenizer( lowerCamelCase__, padding="""max_length""", max_length=self.tokenizer.model_max_length, truncation=lowerCamelCase__, return_tensors="""pt""", ) A : List[Any] = self.text_encoder(content_text_input.input_ids.to(self.device ) )[0] A : Union[str, Any] = self.tokenizer( lowerCamelCase__, padding="""max_length""", max_length=self.tokenizer.model_max_length, truncation=lowerCamelCase__, return_tensors="""pt""", ) A : str = self.text_encoder(style_text_input.input_ids.to(self.device ) )[0] A : List[Any] = slerp(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) # duplicate text embeddings for each generation per prompt A : Optional[Any] = text_embeddings.repeat_interleave(lowerCamelCase__, dim=0 ) # set timesteps A : Any = """offset""" in set(inspect.signature(self.scheduler.set_timesteps ).parameters.keys() ) A : Optional[int] = {} if accepts_offset: A : int = 1 self.scheduler.set_timesteps(lowerCamelCase__, **lowerCamelCase__ ) # Some schedulers like PNDM have timesteps as arrays # It's more optimized to move all timesteps to correct device beforehand self.scheduler.timesteps.to(self.device ) A , A : int = self.get_timesteps(lowerCamelCase__, lowerCamelCase__, self.device ) A : Union[str, Any] = timesteps[:1].repeat(lowerCamelCase__ ) # Preprocess image A : int = preprocess(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) A : Optional[Any] = self.prepare_latents( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, text_embeddings.dtype, self.device, lowerCamelCase__ ) A : List[Any] = preprocess(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) A : List[str] = self.prepare_latents( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, text_embeddings.dtype, self.device, lowerCamelCase__ ) A : str = slerp(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) if clip_guidance_scale > 0: A : str = self.get_clip_image_embeddings(lowerCamelCase__, lowerCamelCase__ ) A : Dict = self.get_clip_image_embeddings(lowerCamelCase__, lowerCamelCase__ ) A : List[str] = slerp( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__ ) # here `guidance_scale` is defined analog to the guidance weight `w` of equation (2) # of the Imagen paper: https://arxiv.org/pdf/2205.11487.pdf . `guidance_scale = 1` # corresponds to doing no classifier free guidance. A : Tuple = guidance_scale > 1.0 # get unconditional embeddings for classifier free guidance if do_classifier_free_guidance: A : Optional[Any] = content_text_input.input_ids.shape[-1] A : List[Any] = self.tokenizer([""""""], padding="""max_length""", max_length=lowerCamelCase__, return_tensors="""pt""" ) A : Tuple = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # duplicate unconditional embeddings for each generation per prompt A : Optional[Any] = uncond_embeddings.repeat_interleave(lowerCamelCase__, dim=0 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes A : Tuple = torch.cat([uncond_embeddings, text_embeddings] ) # get the initial random noise unless the user supplied it # Unlike in other pipelines, latents need to be generated in the target device # for 1-to-1 results reproducibility with the CompVis implementation. # However this currently doesn't work in `mps`. A : str = (batch_size, self.unet.config.in_channels, height // 8, width // 8) A : int = text_embeddings.dtype if latents is None: if self.device.type == "mps": # randn does not work reproducibly on mps A : Union[str, Any] = torch.randn(lowerCamelCase__, generator=lowerCamelCase__, device="""cpu""", dtype=lowerCamelCase__ ).to( self.device ) else: A : str = torch.randn(lowerCamelCase__, generator=lowerCamelCase__, device=self.device, dtype=lowerCamelCase__ ) else: if latents.shape != latents_shape: raise ValueError(f'''Unexpected latents shape, got {latents.shape}, expected {latents_shape}''' ) A : int = latents.to(self.device ) # scale the initial noise by the standard deviation required by the scheduler A : Dict = latents * self.scheduler.init_noise_sigma # prepare extra kwargs for the scheduler step, since not all schedulers have the same signature # eta (η) is only used with the DDIMScheduler, it will be ignored for other schedulers. # eta corresponds to η in DDIM paper: https://arxiv.org/abs/2010.02502 # and should be between [0, 1] A : Optional[Any] = """eta""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) A : int = {} if accepts_eta: A : Optional[Any] = eta # check if the scheduler accepts generator A : List[str] = """generator""" in set(inspect.signature(self.scheduler.step ).parameters.keys() ) if accepts_generator: A : List[str] = generator with self.progress_bar(total=lowerCamelCase__ ): for i, t in enumerate(lowerCamelCase__ ): # expand the latents if we are doing classifier free guidance A : Union[str, Any] = torch.cat([latents] * 2 ) if do_classifier_free_guidance else latents A : int = self.scheduler.scale_model_input(lowerCamelCase__, lowerCamelCase__ ) # predict the noise residual A : List[Any] = self.unet(lowerCamelCase__, lowerCamelCase__, encoder_hidden_states=lowerCamelCase__ ).sample # perform classifier free guidance if do_classifier_free_guidance: A , A : Dict = noise_pred.chunk(2 ) A : int = noise_pred_uncond + guidance_scale * (noise_pred_text - noise_pred_uncond) # perform clip guidance if clip_guidance_scale > 0: A : Union[str, Any] = ( text_embeddings.chunk(2 )[1] if do_classifier_free_guidance else text_embeddings ) A , A : List[Any] = self.cond_fn( lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, ) # compute the previous noisy sample x_t -> x_t-1 A : Union[str, Any] = self.scheduler.step(lowerCamelCase__, lowerCamelCase__, lowerCamelCase__, **lowerCamelCase__ ).prev_sample # Hardcode 0.18215 because stable-diffusion-2-base has not self.vae.config.scaling_factor A : List[str] = 1 / 0.1_8215 * latents A : Optional[int] = self.vae.decode(lowerCamelCase__ ).sample A : Tuple = (image / 2 + 0.5).clamp(0, 1 ) A : Union[str, Any] = image.cpu().permute(0, 2, 3, 1 ).numpy() if output_type == "pil": A : Union[str, Any] = self.numpy_to_pil(lowerCamelCase__ ) if not return_dict: return (image, None) return StableDiffusionPipelineOutput(images=lowerCamelCase__, nsfw_content_detected=lowerCamelCase__ )
116
from __future__ import annotations from math import ceil, floor, sqrt def __UpperCamelCase ( _lowerCAmelCase = 200_0000 ) -> int: """simple docstring""" A : list[int] = [0] A : int for idx in range(1 , ceil(sqrt(target * 2 ) * 1.1 ) ): triangle_numbers.append(triangle_numbers[-1] + idx ) # we want this to be as close as possible to target A : int = 0 # the area corresponding to the grid that gives the product closest to target A : int = 0 # an estimate of b, using the quadratic formula A : float # the largest integer less than b_estimate A : int # the largest integer less than b_estimate A : int # the triangle number corresponding to b_floor A : int # the triangle number corresponding to b_ceil A : int for idx_a, triangle_a in enumerate(triangle_numbers[1:] , 1 ): A : Union[str, Any] = (-1 + sqrt(1 + 8 * target / triangle_a )) / 2 A : List[Any] = floor(_lowerCAmelCase ) A : Tuple = ceil(_lowerCAmelCase ) A : int = triangle_numbers[b_floor] A : Dict = triangle_numbers[b_ceil] if abs(target - triangle_b_first_guess * triangle_a ) < abs( target - best_product ): A : Optional[int] = triangle_b_first_guess * triangle_a A : Optional[int] = idx_a * b_floor if abs(target - triangle_b_second_guess * triangle_a ) < abs( target - best_product ): A : Tuple = triangle_b_second_guess * triangle_a A : Tuple = idx_a * b_ceil return area if __name__ == "__main__": print(F"""{solution() = }""")
116
1
def lowerCAmelCase__ ( _a : float , _a : float ): if density <= 0: raise ValueError("Impossible fluid density" ) if bulk_modulus <= 0: raise ValueError("Impossible bulk modulus" ) return (bulk_modulus / density) ** 0.5 if __name__ == "__main__": import doctest doctest.testmod()
36
from __future__ import annotations def lowerCAmelCase__ ( _a : str , _a : str ): snake_case_ : Optional[Any] = get_failure_array(_a ) # 2) Step through text searching for pattern snake_case_ , snake_case_ : Dict = 0, 0 # index into text, pattern while i < len(_a ): if pattern[j] == text[i]: if j == (len(_a ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: snake_case_ : Dict = failure[j - 1] continue i += 1 return False def lowerCAmelCase__ ( _a : str ): snake_case_ : str = [0] snake_case_ : Optional[int] = 0 snake_case_ : Optional[Any] = 1 while j < len(_a ): if pattern[i] == pattern[j]: i += 1 elif i > 0: snake_case_ : Union[str, Any] = failure[i - 1] continue j += 1 failure.append(_a ) return failure if __name__ == "__main__": # Test 1) lowercase : Union[str, Any] = '''abc1abc12''' lowercase : Optional[int] = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' lowercase : Tuple = '''alskfjaldsk23adsfabcabc''' assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) lowercase : Union[str, Any] = '''ABABX''' lowercase : Any = '''ABABZABABYABABX''' assert kmp(pattern, text) # Test 3) lowercase : str = '''AAAB''' lowercase : str = '''ABAAAAAB''' assert kmp(pattern, text) # Test 4) lowercase : Optional[int] = '''abcdabcy''' lowercase : List[Any] = '''abcxabcdabxabcdabcdabcy''' assert kmp(pattern, text) # Test 5) lowercase : Dict = '''aabaabaaa''' assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
36
1
from ....configuration_utils import PretrainedConfig from ....utils import logging lowercase__ =logging.get_logger(__name__) lowercase__ ={ "Visual-Attention-Network/van-base": ( "https://huggingface.co/Visual-Attention-Network/van-base/blob/main/config.json" ), } class UpperCamelCase__ ( __SCREAMING_SNAKE_CASE ): _SCREAMING_SNAKE_CASE : Dict = "van" def __init__(self : Dict , snake_case_ : List[Any]=2_2_4 , snake_case_ : Optional[Any]=3 , snake_case_ : Optional[Any]=[7, 3, 3, 3] , snake_case_ : str=[4, 2, 2, 2] , snake_case_ : int=[6_4, 1_2_8, 3_2_0, 5_1_2] , snake_case_ : Dict=[3, 3, 1_2, 3] , snake_case_ : List[str]=[8, 8, 4, 4] , snake_case_ : int="gelu" , snake_case_ : int=0.02 , snake_case_ : List[str]=1E-6 , snake_case_ : Optional[Any]=1E-2 , snake_case_ : List[Any]=0.0 , snake_case_ : Tuple=0.0 , **snake_case_ : List[str] , ): super().__init__(**__UpperCAmelCase ) __a : List[str] = image_size __a : Any = num_channels __a : Tuple = patch_sizes __a : Union[str, Any] = strides __a : str = hidden_sizes __a : str = depths __a : str = mlp_ratios __a : Tuple = hidden_act __a : Optional[int] = initializer_range __a : Dict = layer_norm_eps __a : List[Any] = layer_scale_init_value __a : List[Any] = drop_path_rate __a : Union[str, Any] = dropout_rate
216
"""simple docstring""" from __future__ import annotations import string from itertools import cycle, product from pathlib import Path UpperCamelCase : str = ( string.ascii_letters + string.digits + string.punctuation + string.whitespace ) UpperCamelCase : list[int] = [ord(letter) for letter in string.ascii_lowercase] UpperCamelCase : set[int] = {ord(char) for char in VALID_CHARS} UpperCamelCase : list[str] = ["the", "be", "to", "of", "and", "in", "that", "have"] def A ( snake_case :list[int] , snake_case :tuple[int, ...] ) -> str | None: __UpperCamelCase = "" __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 for keychar, cipherchar in zip(cycle(snake_case ) , snake_case ): __UpperCamelCase = cipherchar ^ keychar if decodedchar not in VALID_INTS: return None decoded += chr(snake_case ) return decoded def A ( snake_case :list[int] ) -> list[str]: __UpperCamelCase = [] for key in product(snake_case , repeat=3 ): __UpperCamelCase = try_key(snake_case , snake_case ) if encoded is not None: possibles.append(snake_case ) return possibles def A ( snake_case :list[str] , snake_case :str ) -> list[str]: return [possible for possible in possibles if common_word in possible.lower()] def A ( snake_case :str = "p059_cipher.txt" ) -> int: __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = 42 __UpperCamelCase = Path(snake_case ).parent.joinpath(snake_case ).read_text(encoding='utf-8' ) __UpperCamelCase = [int(snake_case ) for number in data.strip().split(',' )] __UpperCamelCase = filter_valid_chars(snake_case ) for common_word in COMMON_WORDS: __UpperCamelCase = filter_common_word(snake_case , snake_case ) if len(snake_case ) == 1: break __UpperCamelCase = possibles[0] return sum(ord(snake_case ) for char in decoded_text ) if __name__ == "__main__": print(f'''{solution() = }''')
316
0
import argparse import shutil from pathlib import Path from tqdm import tqdm from transformers import AutoTokenizer def a_ ( lowerCAmelCase_ : int, lowerCAmelCase_ : Any, lowerCAmelCase_ : List[str], lowerCAmelCase_ : int=1024 ): __lowerCAmelCase , __lowerCAmelCase = [], [] __lowerCAmelCase = list(zip(lowerCAmelCase_, lowerCAmelCase_ ) ) __lowerCAmelCase , __lowerCAmelCase = sorted_examples[0] def is_too_big(lowerCAmelCase_ : Dict ): return tok(lowerCAmelCase_, return_tensors='pt' ).input_ids.shape[1] > max_tokens for src, tgt in tqdm(sorted_examples[1:] ): __lowerCAmelCase = new_src + ' ' + src __lowerCAmelCase = new_tgt + ' ' + tgt if is_too_big(lowerCAmelCase_ ) or is_too_big(lowerCAmelCase_ ): # cant fit, finalize example finished_src.append(lowerCAmelCase_ ) finished_tgt.append(lowerCAmelCase_ ) __lowerCAmelCase , __lowerCAmelCase = src, tgt else: # can fit, keep adding __lowerCAmelCase , __lowerCAmelCase = cand_src, cand_tgt # cleanup if new_src: assert new_tgt finished_src.append(lowerCAmelCase_ ) finished_tgt.append(lowerCAmelCase_ ) return finished_src, finished_tgt def a_ ( lowerCAmelCase_ : Optional[int], lowerCAmelCase_ : Path, lowerCAmelCase_ : int, lowerCAmelCase_ : List[str] ): __lowerCAmelCase = Path(lowerCAmelCase_ ) save_path.mkdir(exist_ok=lowerCAmelCase_ ) for split in ["train"]: __lowerCAmelCase , __lowerCAmelCase = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" __lowerCAmelCase = [x.rstrip() for x in Path(lowerCAmelCase_ ).open().readlines()] __lowerCAmelCase = [x.rstrip() for x in Path(lowerCAmelCase_ ).open().readlines()] __lowerCAmelCase , __lowerCAmelCase = pack_examples(lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_, lowerCAmelCase_ ) print(F"""packed {split} split from {len(lowerCAmelCase_ )} examples -> {len(lowerCAmelCase_ )}.""" ) Path(save_path / F"""{split}.source""" ).open('w' ).write('\n'.join(lowerCAmelCase_ ) ) Path(save_path / F"""{split}.target""" ).open('w' ).write('\n'.join(lowerCAmelCase_ ) ) for split in ["val", "test"]: __lowerCAmelCase , __lowerCAmelCase = data_dir / F"""{split}.source""", data_dir / F"""{split}.target""" shutil.copyfile(lowerCAmelCase_, save_path / F"""{split}.source""" ) shutil.copyfile(lowerCAmelCase_, save_path / F"""{split}.target""" ) def a_ ( ): __lowerCAmelCase = argparse.ArgumentParser() parser.add_argument('--tok_name', type=lowerCAmelCase_, help='like facebook/bart-large-cnn,t5-base, etc.' ) parser.add_argument('--max_seq_len', type=lowerCAmelCase_, default=128 ) parser.add_argument('--data_dir', type=lowerCAmelCase_ ) parser.add_argument('--save_path', type=lowerCAmelCase_ ) __lowerCAmelCase = parser.parse_args() __lowerCAmelCase = AutoTokenizer.from_pretrained(args.tok_name ) return pack_data_dir(lowerCAmelCase_, Path(args.data_dir ), args.max_seq_len, args.save_path ) if __name__ == "__main__": packer_cli()
207
import warnings from typing import List, Optional, Tuple, Union import numpy as np import PIL import torch from ...models import UNetaDModel from ...schedulers import RePaintScheduler from ...utils import PIL_INTERPOLATION, logging, randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput _snake_case : Union[str, Any] = logging.get_logger(__name__) # pylint: disable=invalid-name def a_ ( lowerCAmelCase_ : Union[List, PIL.Image.Image, torch.Tensor] ): warnings.warn( 'The preprocess method is deprecated and will be removed in a future version. Please' ' use VaeImageProcessor.preprocess instead', lowerCAmelCase_, ) if isinstance(lowerCAmelCase_, torch.Tensor ): return image elif isinstance(lowerCAmelCase_, PIL.Image.Image ): __lowerCAmelCase = [image] if isinstance(image[0], PIL.Image.Image ): __lowerCAmelCase , __lowerCAmelCase = image[0].size __lowerCAmelCase , __lowerCAmelCase = (x - x % 8 for x in (w, h)) # resize to integer multiple of 8 __lowerCAmelCase = [np.array(i.resize((w, h), resample=PIL_INTERPOLATION['lanczos'] ) )[None, :] for i in image] __lowerCAmelCase = np.concatenate(lowerCAmelCase_, axis=0 ) __lowerCAmelCase = np.array(lowerCAmelCase_ ).astype(np.floataa ) / 255.0 __lowerCAmelCase = image.transpose(0, 3, 1, 2 ) __lowerCAmelCase = 2.0 * image - 1.0 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) elif isinstance(image[0], torch.Tensor ): __lowerCAmelCase = torch.cat(lowerCAmelCase_, dim=0 ) return image def a_ ( lowerCAmelCase_ : Union[List, PIL.Image.Image, torch.Tensor] ): if isinstance(lowerCAmelCase_, torch.Tensor ): return mask elif isinstance(lowerCAmelCase_, PIL.Image.Image ): __lowerCAmelCase = [mask] if isinstance(mask[0], PIL.Image.Image ): __lowerCAmelCase , __lowerCAmelCase = mask[0].size __lowerCAmelCase , __lowerCAmelCase = (x - x % 32 for x in (w, h)) # resize to integer multiple of 32 __lowerCAmelCase = [np.array(m.convert('L' ).resize((w, h), resample=PIL_INTERPOLATION['nearest'] ) )[None, :] for m in mask] __lowerCAmelCase = np.concatenate(lowerCAmelCase_, axis=0 ) __lowerCAmelCase = mask.astype(np.floataa ) / 255.0 __lowerCAmelCase = 0 __lowerCAmelCase = 1 __lowerCAmelCase = torch.from_numpy(lowerCAmelCase_ ) elif isinstance(mask[0], torch.Tensor ): __lowerCAmelCase = torch.cat(lowerCAmelCase_, dim=0 ) return mask class _UpperCAmelCase ( _UpperCamelCase ): """simple docstring""" a_ = 42 a_ = 42 def __init__( self : Optional[Any] , lowerCAmelCase_ : List[str] , lowerCAmelCase_ : List[Any] ) -> Optional[int]: super().__init__() self.register_modules(unet=lowerCAmelCase_ , scheduler=lowerCAmelCase_ ) @torch.no_grad() def __call__( self : Dict , lowerCAmelCase_ : Union[torch.Tensor, PIL.Image.Image] , lowerCAmelCase_ : Union[torch.Tensor, PIL.Image.Image] , lowerCAmelCase_ : int = 2_5_0 , lowerCAmelCase_ : float = 0.0 , lowerCAmelCase_ : int = 1_0 , lowerCAmelCase_ : int = 1_0 , lowerCAmelCase_ : Optional[Union[torch.Generator, List[torch.Generator]]] = None , lowerCAmelCase_ : Optional[str] = "pil" , lowerCAmelCase_ : bool = True , ) -> Union[ImagePipelineOutput, Tuple]: __lowerCAmelCase = image __lowerCAmelCase = _preprocess_image(lowerCAmelCase_ ) __lowerCAmelCase = original_image.to(device=self.device , dtype=self.unet.dtype ) __lowerCAmelCase = _preprocess_mask(lowerCAmelCase_ ) __lowerCAmelCase = mask_image.to(device=self.device , dtype=self.unet.dtype ) __lowerCAmelCase = original_image.shape[0] # sample gaussian noise to begin the loop if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) and len(lowerCAmelCase_ ) != batch_size: raise ValueError( f"""You have passed a list of generators of length {len(lowerCAmelCase_ )}, but requested an effective batch""" f""" size of {batch_size}. Make sure the batch size matches the length of the generators.""" ) __lowerCAmelCase = original_image.shape __lowerCAmelCase = randn_tensor(lowerCAmelCase_ , generator=lowerCAmelCase_ , device=self.device , dtype=self.unet.dtype ) # set step values self.scheduler.set_timesteps(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , self.device ) __lowerCAmelCase = eta __lowerCAmelCase = self.scheduler.timesteps[0] + 1 __lowerCAmelCase = generator[0] if isinstance(lowerCAmelCase_ , lowerCAmelCase_ ) else generator for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): if t < t_last: # predict the noise residual __lowerCAmelCase = self.unet(lowerCAmelCase_ , lowerCAmelCase_ ).sample # compute previous image: x_t -> x_t-1 __lowerCAmelCase = self.scheduler.step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ).prev_sample else: # compute the reverse: x_t-1 -> x_t __lowerCAmelCase = self.scheduler.undo_step(lowerCAmelCase_ , lowerCAmelCase_ , lowerCAmelCase_ ) __lowerCAmelCase = t __lowerCAmelCase = (image / 2 + 0.5).clamp(0 , 1 ) __lowerCAmelCase = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": __lowerCAmelCase = self.numpy_to_pil(lowerCAmelCase_ ) if not return_dict: return (image,) return ImagePipelineOutput(images=lowerCAmelCase_ )
207
1
from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A__ : List[str] = { 'configuration_informer': [ 'INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP', 'InformerConfig', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A__ : List[Any] = [ 'INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST', 'InformerForPrediction', 'InformerModel', 'InformerPreTrainedModel', ] if TYPE_CHECKING: from .configuration_informer import INFORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, InformerConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_informer import ( INFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, InformerForPrediction, InformerModel, InformerPreTrainedModel, ) else: import sys A__ : Optional[int] = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
103
'''simple docstring''' from __future__ import annotations import inspect import unittest from typing import List, Tuple from transformers import RegNetConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFRegNetForImageClassification, TFRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _lowerCAmelCase : """simple docstring""" def __init__( self , _lowerCamelCase , _lowerCamelCase=3 , _lowerCamelCase=32 , _lowerCamelCase=3 , _lowerCamelCase=10 , _lowerCamelCase=[10, 20, 30, 40] , _lowerCamelCase=[1, 1, 2, 1] , _lowerCamelCase=True , _lowerCamelCase=True , _lowerCamelCase="relu" , _lowerCamelCase=3 , _lowerCamelCase=None , ) -> List[str]: A_ : Any = parent A_ : List[Any] = batch_size A_ : List[Any] = image_size A_ : Optional[int] = num_channels A_ : Tuple = embeddings_size A_ : str = hidden_sizes A_ : Optional[Any] = depths A_ : Any = is_training A_ : int = use_labels A_ : int = hidden_act A_ : Optional[Any] = num_labels A_ : str = scope A_ : Optional[int] = len(_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A_ : Dict = None if self.use_labels: A_ : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) A_ : Union[str, Any] = self.get_config() return config, pixel_values, labels def UpperCAmelCase_ ( self ) -> Optional[Any]: return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> List[str]: A_ : Dict = TFRegNetModel(config=_lowerCamelCase ) A_ : Optional[int] = model(_lowerCamelCase , training=_lowerCamelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def UpperCAmelCase_ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) -> Optional[int]: A_ : Optional[Any] = self.num_labels A_ : int = TFRegNetForImageClassification(_lowerCamelCase ) A_ : Tuple = model(_lowerCamelCase , labels=_lowerCamelCase , training=_lowerCamelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def UpperCAmelCase_ ( self ) -> str: A_ : Any = self.prepare_config_and_inputs() A_ , A_ , A_ : str = config_and_inputs A_ : Optional[int] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class _lowerCAmelCase ( __A, __A, unittest.TestCase ): """simple docstring""" lowerCamelCase = (TFRegNetModel, TFRegNetForImageClassification) if is_tf_available() else () lowerCamelCase = ( {'''feature-extraction''': TFRegNetModel, '''image-classification''': TFRegNetForImageClassification} if is_tf_available() else {} ) lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False lowerCamelCase = False def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : Dict = TFRegNetModelTester(self ) A_ : Optional[int] = ConfigTester(self , config_class=_lowerCamelCase , has_text_modality=_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> str: return @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def UpperCAmelCase_ ( self ) -> Dict: pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def UpperCAmelCase_ ( self ) -> int: super().test_keras_fit() @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def UpperCAmelCase_ ( self ) -> Optional[Any]: pass def UpperCAmelCase_ ( self ) -> int: A_ , A_ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[int] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A_ : int = [*signature.parameters.keys()] A_ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Tuple: A_ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*_lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: def check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ): A_ : Optional[int] = model_class(_lowerCamelCase ) A_ : List[Any] = model(**self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) , training=_lowerCamelCase ) A_ : List[str] = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A_ : Optional[int] = self.model_tester.num_stages self.assertEqual(len(_lowerCamelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) A_ , A_ : Any = self.model_tester.prepare_config_and_inputs_for_common() A_ : List[str] = ["""basic""", """bottleneck"""] for model_class in self.all_model_classes: for layer_type in layers_type: A_ : Dict = layer_type A_ : List[Any] = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A_ : str = True check_hidden_states_output(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) def UpperCAmelCase_ ( self ) -> Dict: A_ , A_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() def check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase={} ): A_ : Dict = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ) A_ : Optional[Any] = model(_lowerCamelCase , return_dict=_lowerCamelCase , **_lowerCamelCase ).to_tuple() def recursive_check(_lowerCamelCase , _lowerCamelCase ): if isinstance(_lowerCamelCase , (List, Tuple) ): for tuple_iterable_value, dict_iterable_value in zip(_lowerCamelCase , _lowerCamelCase ): recursive_check(_lowerCamelCase , _lowerCamelCase ) elif tuple_object is None: return else: self.assertTrue( all(tf.equal(_lowerCamelCase , _lowerCamelCase ) ) , msg=( """Tuple and dict output are not equal. Difference:""" F" {tf.math.reduce_max(tf.abs(tuple_object - dict_object ) )}" ) , ) recursive_check(_lowerCamelCase , _lowerCamelCase ) for model_class in self.all_model_classes: A_ : Optional[Any] = model_class(_lowerCamelCase ) A_ : Optional[Any] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : Optional[int] = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Any = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase ) A_ : Dict = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) A_ : int = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) A_ : Tuple = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) A_ : str = self._prepare_for_class(_lowerCamelCase , _lowerCamelCase , return_labels=_lowerCamelCase ) check_equivalence(_lowerCamelCase , _lowerCamelCase , _lowerCamelCase , {"""output_hidden_states""": True} ) def UpperCAmelCase_ ( self ) -> str: A_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*_lowerCamelCase ) @slow def UpperCAmelCase_ ( self ) -> Tuple: for model_name in TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A_ : Dict = TFRegNetModel.from_pretrained(_lowerCamelCase ) self.assertIsNotNone(_lowerCamelCase ) def UpperCAmelCase ( ) -> Tuple: """simple docstring""" A_ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" @cached_property def UpperCAmelCase_ ( self ) -> int: return ( AutoImageProcessor.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def UpperCAmelCase_ ( self ) -> Optional[Any]: A_ : str = TFRegNetForImageClassification.from_pretrained(TF_REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A_ : Tuple = self.default_image_processor A_ : Optional[int] = prepare_img() A_ : Any = image_processor(images=_lowerCamelCase , return_tensors="""tf""" ) # forward pass A_ : List[Any] = model(**_lowerCamelCase , training=_lowerCamelCase ) # verify the logits A_ : Optional[Any] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , _lowerCamelCase ) A_ : Optional[Any] = tf.constant([-0.4180, -1.5051, -3.4836] ) tf.debugging.assert_near(outputs.logits[0, :3] , _lowerCamelCase , atol=1e-4 )
344
0
from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import ShapEPipeline else: from .camera import create_pan_cameras from .pipeline_shap_e import ShapEPipeline from .pipeline_shap_e_img2img import ShapEImgaImgPipeline from .renderer import ( BoundingBoxVolume, ImportanceRaySampler, MLPNeRFModelOutput, MLPNeRSTFModel, ShapEParamsProjModel, ShapERenderer, StratifiedRaySampler, VoidNeRFModel, )
211
import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionPipeline from diffusers.utils.testing_utils import load_image, nightly, require_torch_gpu, torch_device lowerCamelCase = False class _a ( unittest.TestCase): pass @nightly @require_torch_gpu class _a ( unittest.TestCase): def UpperCAmelCase__( self : int )-> int: # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def UpperCAmelCase__( self : Any )-> Dict: lowerCAmelCase__ : int = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) lowerCAmelCase__ : List[str] = torch.manual_seed(0 ) lowerCAmelCase__ : List[str] = pipe.dual_guided( prompt='''first prompt''' , image=_SCREAMING_SNAKE_CASE , text_to_image_strength=0.75 , generator=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[Any] = VersatileDiffusionPipeline.from_pretrained(_SCREAMING_SNAKE_CASE , torch_dtype=torch.floataa ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : List[str] = generator.manual_seed(0 ) lowerCAmelCase__ : str = pipe.dual_guided( prompt='''first prompt''' , image=_SCREAMING_SNAKE_CASE , text_to_image_strength=0.75 , generator=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=2 , output_type='''numpy''' , ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def UpperCAmelCase__( self : Tuple )-> List[Any]: lowerCAmelCase__ : Optional[int] = VersatileDiffusionPipeline.from_pretrained('''shi-labs/versatile-diffusion''' , torch_dtype=torch.floataa ) pipe.to(_SCREAMING_SNAKE_CASE ) pipe.set_progress_bar_config(disable=_SCREAMING_SNAKE_CASE ) lowerCAmelCase__ : int = '''cyberpunk 2077''' lowerCAmelCase__ : List[str] = load_image( '''https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/versatile_diffusion/benz.jpg''' ) lowerCAmelCase__ : List[Any] = torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = pipe.dual_guided( prompt=_SCREAMING_SNAKE_CASE , image=_SCREAMING_SNAKE_CASE , text_to_image_strength=0.75 , generator=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' , ).images lowerCAmelCase__ : List[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase__ : List[str] = np.array([0.1448, 0.1619, 0.1741, 0.1086, 0.1147, 0.1128, 0.1199, 0.1165, 0.1001] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 lowerCAmelCase__ : int = '''A painting of a squirrel eating a burger ''' lowerCAmelCase__ : Optional[Any] = torch.manual_seed(0 ) lowerCAmelCase__ : Union[str, Any] = pipe.text_to_image( prompt=_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , guidance_scale=7.5 , num_inference_steps=50 , output_type='''numpy''' ).images lowerCAmelCase__ : List[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase__ : Optional[Any] = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 lowerCAmelCase__ : Optional[Any] = pipe.image_variation(_SCREAMING_SNAKE_CASE , generator=_SCREAMING_SNAKE_CASE , output_type='''numpy''' ).images lowerCAmelCase__ : List[Any] = image[0, 253:256, 253:256, -1] assert image.shape == (1, 512, 512, 3) lowerCAmelCase__ : Any = np.array([0.3076, 0.3123, 0.3284, 0.3782, 0.3770, 0.3894, 0.4297, 0.4331, 0.4456] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1
211
1
"""simple docstring""" import unittest from transformers import BarthezTokenizer, BarthezTokenizerFast, BatchEncoding from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers @require_sentencepiece @slow # see https://github.com/huggingface/transformers/issues/11457 class lowerCamelCase (UpperCAmelCase__ ,unittest.TestCase ): lowerCamelCase__ : Optional[Any] = BarthezTokenizer lowerCamelCase__ : Tuple = BarthezTokenizerFast lowerCamelCase__ : Optional[int] = True lowerCamelCase__ : Optional[Any] = True def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Dict: super().setUp() SCREAMING_SNAKE_CASE__ = BarthezTokenizerFast.from_pretrained("""moussaKam/mbarthez""" ) tokenizer.save_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname , legacy_format=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = tokenizer def SCREAMING_SNAKE_CASE ( self : Optional[Any] ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__ = """<pad>""" SCREAMING_SNAKE_CASE__ = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(UpperCAmelCase__ ) , UpperCAmelCase__ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(UpperCAmelCase__ ) , UpperCAmelCase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Tuple: SCREAMING_SNAKE_CASE__ = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , """<s>""" ) self.assertEqual(vocab_keys[1] , """<pad>""" ) self.assertEqual(vocab_keys[-1] , """<mask>""" ) self.assertEqual(len(UpperCAmelCase__ ) , 1_0_1_1_2_2 ) def SCREAMING_SNAKE_CASE ( self : Optional[int] ) -> Any: self.assertEqual(self.get_tokenizer().vocab_size , 1_0_1_1_2_2 ) @require_torch def SCREAMING_SNAKE_CASE ( self : List[str] ) -> str: SCREAMING_SNAKE_CASE__ = ["""A long paragraph for summarization.""", """Another paragraph for summarization."""] SCREAMING_SNAKE_CASE__ = [0, 5_7, 3_0_1_8, 7_0_3_0_7, 9_1, 2] SCREAMING_SNAKE_CASE__ = self.tokenizer( UpperCAmelCase__ , max_length=len(UpperCAmelCase__ ) , padding=UpperCAmelCase__ , truncation=UpperCAmelCase__ , return_tensors="""pt""" ) self.assertIsInstance(UpperCAmelCase__ , UpperCAmelCase__ ) self.assertEqual((2, 6) , batch.input_ids.shape ) self.assertEqual((2, 6) , batch.attention_mask.shape ) SCREAMING_SNAKE_CASE__ = batch.input_ids.tolist()[0] self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) def SCREAMING_SNAKE_CASE ( self : Tuple ) -> Optional[Any]: if not self.test_rust_tokenizer: return SCREAMING_SNAKE_CASE__ = self.get_tokenizer() SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = """I was born in 92000, and this is falsé.""" SCREAMING_SNAKE_CASE__ = tokenizer.tokenize(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.tokenize(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase__ , add_special_tokens=UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = self.get_rust_tokenizer() SCREAMING_SNAKE_CASE__ = tokenizer.encode(UpperCAmelCase__ ) SCREAMING_SNAKE_CASE__ = rust_tokenizer.encode(UpperCAmelCase__ ) self.assertListEqual(UpperCAmelCase__ , UpperCAmelCase__ ) @slow def SCREAMING_SNAKE_CASE ( self : Union[str, Any] ) -> int: SCREAMING_SNAKE_CASE__ = {"""input_ids""": [[0, 4_9_0, 1_4_3_2_8, 4_5_0_7, 3_5_4, 4_7, 4_3_6_6_9, 9_5, 2_5, 7_8_1_1_7, 2_0_2_1_5, 1_9_7_7_9, 1_9_0, 2_2, 4_0_0, 4, 3_5_3_4_3, 8_0_3_1_0, 6_0_3, 8_6, 2_4_9_3_7, 1_0_5, 3_3_4_3_8, 9_4_7_6_2, 1_9_6, 3_9_6_4_2, 7, 1_5, 1_5_9_3_3, 1_7_3, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [0, 1_0_5_3_4, 8_7, 2_5, 6_6, 3_3_5_8, 1_9_6, 5_5_2_8_9, 8, 8_2_9_6_1, 8_1, 2_2_0_4, 7_5_2_0_3, 7, 1_5, 7_6_3, 1_2_9_5_6, 2_1_6, 1_7_8, 1_4_3_2_8, 9_5_9_5, 1_3_7_7, 6_9_6_9_3, 7, 4_4_8, 7_1_0_2_1, 1_9_6, 1_8_1_0_6, 1_4_3_7, 1_3_9_7_4, 1_0_8, 9_0_8_3, 4, 4_9_3_1_5, 7, 3_9, 8_6, 1_3_2_6, 2_7_9_3, 4_6_3_3_3, 4, 4_4_8, 1_9_6, 7_4_5_8_8, 7, 4_9_3_1_5, 7, 3_9, 2_1, 8_2_2, 3_8_4_7_0, 7_4, 2_1, 6_6_7_2_3, 6_2_4_8_0, 8, 2_2_0_5_0, 5, 2]], """attention_mask""": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]]} # noqa: E501 # fmt: on # moussaKam/mbarthez is a french model. So we also use french texts. SCREAMING_SNAKE_CASE__ = [ """Le transformeur est un modèle d\'apprentissage profond introduit en 2017, """ """utilisé principalement dans le domaine du traitement automatique des langues (TAL).""", """À l\'instar des réseaux de neurones récurrents (RNN), les transformeurs sont conçus """ """pour gérer des données séquentielles, telles que le langage naturel, pour des tâches """ """telles que la traduction et la synthèse de texte.""", ] self.tokenizer_integration_test_util( expected_encoding=UpperCAmelCase__ , model_name="""moussaKam/mbarthez""" , revision="""c2e4ecbca5e3cd2c37fe1ac285ca4fbdf1366fb6""" , sequences=UpperCAmelCase__ , )
165
def SCREAMING_SNAKE_CASE ( lowercase_ ) -> Any: """simple docstring""" A__ = [0] * len(lowercase_ ) A__ = [] A__ = [1] * len(lowercase_ ) for values in graph.values(): for i in values: indegree[i] += 1 for i in range(len(lowercase_ ) ): if indegree[i] == 0: queue.append(lowercase_ ) while queue: A__ = queue.pop(0 ) for x in graph[vertex]: indegree[x] -= 1 if long_dist[vertex] + 1 > long_dist[x]: A__ = long_dist[vertex] + 1 if indegree[x] == 0: queue.append(lowercase_ ) print(max(lowercase_ ) ) # Adjacency list of Graph _lowerCamelCase : Optional[int] = {0: [2, 3, 4], 1: [2, 7], 2: [5], 3: [5, 7], 4: [7], 5: [6], 6: [7], 7: []} longest_distance(graph)
14
0
class __a: """simple docstring""" def __init__( self ,_SCREAMING_SNAKE_CASE ) -> Tuple: UpperCAmelCase_ : Union[str, Any] = n UpperCAmelCase_ : Union[str, Any] = [None] * self.n UpperCAmelCase_ : Any = 0 # index of the first element UpperCAmelCase_ : Optional[int] = 0 UpperCAmelCase_ : List[Any] = 0 def __len__( self ) -> int: return self.size def a__ ( self ) -> bool: return self.size == 0 def a__ ( self ) -> str: return False if self.is_empty() else self.array[self.front] def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> Optional[int]: if self.size >= self.n: raise Exception('''QUEUE IS FULL''' ) UpperCAmelCase_ : Optional[Any] = data UpperCAmelCase_ : List[str] = (self.rear + 1) % self.n self.size += 1 return self def a__ ( self ) -> Union[str, Any]: if self.size == 0: raise Exception('''UNDERFLOW''' ) UpperCAmelCase_ : Union[str, Any] = self.array[self.front] UpperCAmelCase_ : str = None UpperCAmelCase_ : Optional[int] = (self.front + 1) % self.n self.size -= 1 return temp
364
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_pegasus import PegasusTokenizer else: __a = None __a = logging.get_logger(__name__) __a = '▁' __a = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} __a = { 'vocab_file': {'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/spiece.model'}, 'tokenizer_file': { 'google/pegasus-xsum': 'https://huggingface.co/google/pegasus-xsum/resolve/main/tokenizer.json' }, } __a = { 'google/pegasus-xsum': 512, } class __a( _a ): """simple docstring""" lowerCAmelCase = VOCAB_FILES_NAMES lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP lowerCAmelCase = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES lowerCAmelCase = PegasusTokenizer lowerCAmelCase = ['''input_ids''', '''attention_mask'''] def __init__( self ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE="<pad>" ,_SCREAMING_SNAKE_CASE="</s>" ,_SCREAMING_SNAKE_CASE="<unk>" ,_SCREAMING_SNAKE_CASE="<mask_2>" ,_SCREAMING_SNAKE_CASE="<mask_1>" ,_SCREAMING_SNAKE_CASE=None ,_SCREAMING_SNAKE_CASE=103 ,**_SCREAMING_SNAKE_CASE ,) -> Optional[Any]: UpperCAmelCase_ : Dict = offset if additional_special_tokens is not None: if not isinstance(_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE ): raise TypeError( f'''additional_special_tokens should be of type {type(_SCREAMING_SNAKE_CASE )}, but is''' f''' {type(_SCREAMING_SNAKE_CASE )}''' ) UpperCAmelCase_ : str = ( ([mask_token_sent] + additional_special_tokens) if mask_token_sent not in additional_special_tokens and mask_token_sent is not None else additional_special_tokens ) # fill additional tokens with ..., <unk_token_102> in case not all additional tokens are already taken additional_special_tokens_extended += [ f'''<unk_{i}>''' for i in range(len(_SCREAMING_SNAKE_CASE ) ,self.offset - 1 ) ] if len(set(_SCREAMING_SNAKE_CASE ) ) != len(_SCREAMING_SNAKE_CASE ): raise ValueError( '''Please make sure that the provided additional_special_tokens do not contain an incorrectly''' f''' shifted list of <unk_x> tokens. Found {additional_special_tokens_extended}.''' ) UpperCAmelCase_ : int = additional_special_tokens_extended else: UpperCAmelCase_ : Any = [mask_token_sent] if mask_token_sent is not None else [] additional_special_tokens += [f'''<unk_{i}>''' for i in range(2 ,self.offset )] super().__init__( _SCREAMING_SNAKE_CASE ,tokenizer_file=_SCREAMING_SNAKE_CASE ,pad_token=_SCREAMING_SNAKE_CASE ,eos_token=_SCREAMING_SNAKE_CASE ,unk_token=_SCREAMING_SNAKE_CASE ,mask_token=_SCREAMING_SNAKE_CASE ,mask_token_sent=_SCREAMING_SNAKE_CASE ,offset=_SCREAMING_SNAKE_CASE ,additional_special_tokens=_SCREAMING_SNAKE_CASE ,**_SCREAMING_SNAKE_CASE ,) UpperCAmelCase_ : str = vocab_file UpperCAmelCase_ : Dict = False if not self.vocab_file else True def a__ ( self ,_SCREAMING_SNAKE_CASE ) -> Any: UpperCAmelCase_ : Dict = set(self.all_special_ids ) # call it once instead of inside list comp all_special_ids.remove(self.unk_token_id ) # <unk> is only sometimes special if all_special_ids != set(range(len(self.additional_special_tokens ) + 3 ) ): raise ValueError( '''There should be 3 special tokens: mask_token, pad_token, and eos_token +''' f''' {len(self.additional_special_tokens )} additional_special_tokens, but got {all_special_ids}''' ) return [1 if x in all_special_ids else 0 for x in seq] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ,_SCREAMING_SNAKE_CASE = False ) -> List[int]: if already_has_special_tokens: return self._special_token_mask(_SCREAMING_SNAKE_CASE ) elif token_ids_a is None: return self._special_token_mask(_SCREAMING_SNAKE_CASE ) + [1] else: return self._special_token_mask(token_ids_a + token_ids_a ) + [1] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE=None ) -> List[int]: if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def a__ ( self ,_SCREAMING_SNAKE_CASE ,_SCREAMING_SNAKE_CASE = None ) -> Tuple[str]: if not self.can_save_slow_tokenizer: raise ValueError( '''Your fast tokenizer does not have the necessary information to save the vocabulary for a slow ''' '''tokenizer.''' ) if not os.path.isdir(_SCREAMING_SNAKE_CASE ): logger.error(f'''Vocabulary path ({save_directory}) should be a directory''' ) return UpperCAmelCase_ : Optional[int] = os.path.join( _SCREAMING_SNAKE_CASE ,(filename_prefix + '''-''' if filename_prefix else '''''') + VOCAB_FILES_NAMES['''vocab_file'''] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(_SCREAMING_SNAKE_CASE ): copyfile(self.vocab_file ,_SCREAMING_SNAKE_CASE ) return (out_vocab_file,)
235
0
from typing import Callable, List, Optional, Tuple, Union import torch from transformers import CLIPTextModel, CLIPTokenizer from ...configuration_utils import ConfigMixin, register_to_config from ...models import ModelMixin, TransformeraDModel, VQModel from ...schedulers import VQDiffusionScheduler from ...utils import logging from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput __UpperCAmelCase = logging.get_logger(__name__) # pylint: disable=invalid-name class lowerCamelCase (_snake_case , _snake_case ): '''simple docstring''' @register_to_config def __init__( self , _UpperCamelCase , _UpperCamelCase = None , _UpperCamelCase = None ) -> Optional[int]: super().__init__() UpperCAmelCase_ : str = learnable if self.learnable: assert hidden_size is not None, "learnable=True requires `hidden_size` to be set" assert length is not None, "learnable=True requires `length` to be set" UpperCAmelCase_ : List[str] = torch.zeros(_UpperCamelCase , _UpperCamelCase ) else: UpperCAmelCase_ : Optional[Any] = None UpperCAmelCase_ : int = torch.nn.Parameter(_UpperCamelCase ) class lowerCamelCase (_snake_case ): '''simple docstring''' _snake_case : VQModel _snake_case : CLIPTextModel _snake_case : CLIPTokenizer _snake_case : TransformeraDModel _snake_case : LearnedClassifierFreeSamplingEmbeddings _snake_case : VQDiffusionScheduler def __init__( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase , ) -> Optional[int]: super().__init__() self.register_modules( vqvae=_UpperCamelCase , transformer=_UpperCamelCase , text_encoder=_UpperCamelCase , tokenizer=_UpperCamelCase , scheduler=_UpperCamelCase , learned_classifier_free_sampling_embeddings=_UpperCamelCase , ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) -> List[Any]: UpperCAmelCase_ : Tuple = len(_UpperCamelCase ) if isinstance(_UpperCamelCase , _UpperCamelCase ) else 1 # get prompt text embeddings UpperCAmelCase_ : Optional[Any] = self.tokenizer( _UpperCamelCase , padding='max_length' , max_length=self.tokenizer.model_max_length , return_tensors='pt' , ) UpperCAmelCase_ : Optional[Any] = text_inputs.input_ids if text_input_ids.shape[-1] > self.tokenizer.model_max_length: UpperCAmelCase_ : List[str] = self.tokenizer.batch_decode(text_input_ids[:, self.tokenizer.model_max_length :] ) logger.warning( 'The following part of your input was truncated because CLIP can only handle sequences up to' f" {self.tokenizer.model_max_length} tokens: {removed_text}" ) UpperCAmelCase_ : Tuple = text_input_ids[:, : self.tokenizer.model_max_length] UpperCAmelCase_ : str = self.text_encoder(text_input_ids.to(self.device ) )[0] # NOTE: This additional step of normalizing the text embeddings is from VQ-Diffusion. # While CLIP does normalize the pooled output of the text transformer when combining # the image and text embeddings, CLIP does not directly normalize the last hidden state. # # CLIP normalizing the pooled output. # https://github.com/huggingface/transformers/blob/d92e22d1f28324f513f3080e5c47c071a3916721/src/transformers/models/clip/modeling_clip.py#L1052-L1053 UpperCAmelCase_ : Any = prompt_embeds / prompt_embeds.norm(dim=-1 , keepdim=_UpperCamelCase ) # duplicate text embeddings for each generation per prompt UpperCAmelCase_ : Dict = prompt_embeds.repeat_interleave(_UpperCamelCase , dim=0 ) if do_classifier_free_guidance: if self.learned_classifier_free_sampling_embeddings.learnable: UpperCAmelCase_ : Optional[Any] = self.learned_classifier_free_sampling_embeddings.embeddings UpperCAmelCase_ : List[str] = negative_prompt_embeds.unsqueeze(0 ).repeat(_UpperCamelCase , 1 , 1 ) else: UpperCAmelCase_ : Tuple = [''] * batch_size UpperCAmelCase_ : Union[str, Any] = text_input_ids.shape[-1] UpperCAmelCase_ : List[Any] = self.tokenizer( _UpperCamelCase , padding='max_length' , max_length=_UpperCamelCase , truncation=_UpperCamelCase , return_tensors='pt' , ) UpperCAmelCase_ : List[Any] = self.text_encoder(uncond_input.input_ids.to(self.device ) )[0] # See comment for normalizing text embeddings UpperCAmelCase_ : List[Any] = negative_prompt_embeds / negative_prompt_embeds.norm(dim=-1 , keepdim=_UpperCamelCase ) # duplicate unconditional embeddings for each generation per prompt, using mps friendly method UpperCAmelCase_ : Dict = negative_prompt_embeds.shape[1] UpperCAmelCase_ : Optional[Any] = negative_prompt_embeds.repeat(1 , _UpperCamelCase , 1 ) UpperCAmelCase_ : Union[str, Any] = negative_prompt_embeds.view(batch_size * num_images_per_prompt , _UpperCamelCase , -1 ) # For classifier free guidance, we need to do two forward passes. # Here we concatenate the unconditional and text embeddings into a single batch # to avoid doing two forward passes UpperCAmelCase_ : str = torch.cat([negative_prompt_embeds, prompt_embeds] ) return prompt_embeds @torch.no_grad() def __call__( self , _UpperCamelCase , _UpperCamelCase = 1_0_0 , _UpperCamelCase = 5.0 , _UpperCamelCase = 1.0 , _UpperCamelCase = 1 , _UpperCamelCase = None , _UpperCamelCase = None , _UpperCamelCase = "pil" , _UpperCamelCase = True , _UpperCamelCase = None , _UpperCamelCase = 1 , ) -> Union[ImagePipelineOutput, Tuple]: if isinstance(_UpperCamelCase , _UpperCamelCase ): UpperCAmelCase_ : Any = 1 elif isinstance(_UpperCamelCase , _UpperCamelCase ): UpperCAmelCase_ : Optional[int] = len(_UpperCamelCase ) else: raise ValueError(f"`prompt` has to be of type `str` or `list` but is {type(_UpperCamelCase )}" ) UpperCAmelCase_ : List[str] = batch_size * num_images_per_prompt UpperCAmelCase_ : int = guidance_scale > 1.0 UpperCAmelCase_ : Tuple = self._encode_prompt(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) if (callback_steps is None) or ( callback_steps is not None and (not isinstance(_UpperCamelCase , _UpperCamelCase ) or callback_steps <= 0) ): raise ValueError( f"`callback_steps` has to be a positive integer but is {callback_steps} of type" f" {type(_UpperCamelCase )}." ) # get the initial completely masked latents unless the user supplied it UpperCAmelCase_ : Tuple = (batch_size, self.transformer.num_latent_pixels) if latents is None: UpperCAmelCase_ : Optional[Any] = self.transformer.num_vector_embeds - 1 UpperCAmelCase_ : Any = torch.full(_UpperCamelCase , _UpperCamelCase ).to(self.device ) else: if latents.shape != latents_shape: raise ValueError(f"Unexpected latents shape, got {latents.shape}, expected {latents_shape}" ) if (latents < 0).any() or (latents >= self.transformer.num_vector_embeds).any(): raise ValueError( 'Unexpected latents value(s). All latents be valid embedding indices i.e. in the range 0,' f" {self.transformer.num_vector_embeds - 1} (inclusive)." ) UpperCAmelCase_ : Dict = latents.to(self.device ) # set timesteps self.scheduler.set_timesteps(_UpperCamelCase , device=self.device ) UpperCAmelCase_ : List[Any] = self.scheduler.timesteps.to(self.device ) UpperCAmelCase_ : Optional[int] = latents for i, t in enumerate(self.progress_bar(_UpperCamelCase ) ): # expand the sample if we are doing classifier free guidance UpperCAmelCase_ : Any = torch.cat([sample] * 2 ) if do_classifier_free_guidance else sample # predict the un-noised image # model_output == `log_p_x_0` UpperCAmelCase_ : Optional[int] = self.transformer(_UpperCamelCase , encoder_hidden_states=_UpperCamelCase , timestep=_UpperCamelCase ).sample if do_classifier_free_guidance: UpperCAmelCase_ , UpperCAmelCase_ : Optional[Any] = model_output.chunk(2 ) UpperCAmelCase_ : Optional[int] = model_output_uncond + guidance_scale * (model_output_text - model_output_uncond) model_output -= torch.logsumexp(_UpperCamelCase , dim=1 , keepdim=_UpperCamelCase ) UpperCAmelCase_ : Optional[Any] = self.truncate(_UpperCamelCase , _UpperCamelCase ) # remove `log(0)`'s (`-inf`s) UpperCAmelCase_ : Optional[Any] = model_output.clamp(-7_0 ) # compute the previous noisy sample x_t -> x_t-1 UpperCAmelCase_ : Optional[Any] = self.scheduler.step(_UpperCamelCase , timestep=_UpperCamelCase , sample=_UpperCamelCase , generator=_UpperCamelCase ).prev_sample # call the callback, if provided if callback is not None and i % callback_steps == 0: callback(_UpperCamelCase , _UpperCamelCase , _UpperCamelCase ) UpperCAmelCase_ : List[str] = self.vqvae.config.vq_embed_dim UpperCAmelCase_ : Union[str, Any] = (batch_size, self.transformer.height, self.transformer.width, embedding_channels) UpperCAmelCase_ : List[Any] = self.vqvae.quantize.get_codebook_entry(_UpperCamelCase , shape=_UpperCamelCase ) UpperCAmelCase_ : Dict = self.vqvae.decode(_UpperCamelCase , force_not_quantize=_UpperCamelCase ).sample UpperCAmelCase_ : str = (image / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase_ : Tuple = image.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase_ : List[str] = self.numpy_to_pil(_UpperCamelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=_UpperCamelCase ) def __UpperCAmelCase ( self , _UpperCamelCase , _UpperCamelCase ) -> torch.FloatTensor: UpperCAmelCase_ , UpperCAmelCase_ : Optional[int] = torch.sort(_UpperCamelCase , 1 , descending=_UpperCamelCase ) UpperCAmelCase_ : Tuple = torch.exp(_UpperCamelCase ) UpperCAmelCase_ : Any = sorted_p_x_0.cumsum(dim=1 ) < truncation_rate # Ensure that at least the largest probability is not zeroed out UpperCAmelCase_ : str = torch.full_like(keep_mask[:, 0:1, :] , _UpperCamelCase ) UpperCAmelCase_ : Optional[int] = torch.cat((all_true, keep_mask) , dim=1 ) UpperCAmelCase_ : Tuple = keep_mask[:, :-1, :] UpperCAmelCase_ : List[str] = keep_mask.gather(1 , indices.argsort(1 ) ) UpperCAmelCase_ : str = log_p_x_0.clone() UpperCAmelCase_ : Optional[int] = -torch.inf # -inf = log(0) return rv
29
from __future__ import annotations def lowercase__ ( __snake_case : tuple[int, int] , __snake_case : int ): '''simple docstring''' UpperCAmelCase_ , UpperCAmelCase_ : Tuple = position UpperCAmelCase_ : str = [ (y + 1, x + 2), (y - 1, x + 2), (y + 1, x - 2), (y - 1, x - 2), (y + 2, x + 1), (y + 2, x - 1), (y - 2, x + 1), (y - 2, x - 1), ] UpperCAmelCase_ : Optional[Any] = [] for position in positions: UpperCAmelCase_ , UpperCAmelCase_ : Tuple = position if 0 <= y_test < n and 0 <= x_test < n: permissible_positions.append(__snake_case ) return permissible_positions def lowercase__ ( __snake_case : list[list[int]] ): '''simple docstring''' return not any(elem == 0 for row in board for elem in row ) def lowercase__ ( __snake_case : list[list[int]] , __snake_case : tuple[int, int] , __snake_case : int ): '''simple docstring''' if is_complete(__snake_case ): return True for position in get_valid_pos(__snake_case , len(__snake_case ) ): UpperCAmelCase_ , UpperCAmelCase_ : Any = position if board[y][x] == 0: UpperCAmelCase_ : Optional[Any] = curr + 1 if open_knight_tour_helper(__snake_case , __snake_case , curr + 1 ): return True UpperCAmelCase_ : List[Any] = 0 return False def lowercase__ ( __snake_case : int ): '''simple docstring''' UpperCAmelCase_ : str = [[0 for i in range(__snake_case )] for j in range(__snake_case )] for i in range(__snake_case ): for j in range(__snake_case ): UpperCAmelCase_ : Optional[Any] = 1 if open_knight_tour_helper(__snake_case , (i, j) , 1 ): return board UpperCAmelCase_ : List[Any] = 0 UpperCAmelCase_ : List[str] = F"Open Kight Tour cannot be performed on a board of size {n}" raise ValueError(__snake_case ) if __name__ == "__main__": import doctest doctest.testmod()
29
1
'''simple docstring''' import copy from ...configuration_utils import PretrainedConfig from ...utils import logging from ..bit import BitConfig snake_case__ = logging.get_logger(__name__) snake_case__ = { """Intel/dpt-large""": """https://huggingface.co/Intel/dpt-large/resolve/main/config.json""", # See all DPT models at https://huggingface.co/models?filter=dpt } class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = 'dpt' def __init__( self : int , _lowerCamelCase : Union[str, Any]=768 , _lowerCamelCase : Optional[Any]=12 , _lowerCamelCase : Optional[int]=12 , _lowerCamelCase : Tuple=3072 , _lowerCamelCase : Optional[Any]="gelu" , _lowerCamelCase : Tuple=0.0 , _lowerCamelCase : Union[str, Any]=0.0 , _lowerCamelCase : Any=0.02 , _lowerCamelCase : str=1E-12 , _lowerCamelCase : Dict=384 , _lowerCamelCase : int=16 , _lowerCamelCase : str=3 , _lowerCamelCase : Optional[int]=False , _lowerCamelCase : List[Any]=True , _lowerCamelCase : Any=[2, 5, 8, 11] , _lowerCamelCase : Dict="project" , _lowerCamelCase : Dict=[4, 2, 1, 0.5] , _lowerCamelCase : Tuple=[96, 192, 384, 768] , _lowerCamelCase : Dict=256 , _lowerCamelCase : Tuple=-1 , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : Optional[int]=True , _lowerCamelCase : Tuple=0.4 , _lowerCamelCase : str=255 , _lowerCamelCase : int=0.1 , _lowerCamelCase : Dict=[1, 1024, 24, 24] , _lowerCamelCase : List[Any]=[0, 1] , _lowerCamelCase : str=None , **_lowerCamelCase : Optional[int] , ): """simple docstring""" super().__init__(**_lowerCamelCase ) A_ : Union[str, Any] = hidden_size A_ : Tuple = is_hybrid if self.is_hybrid: if backbone_config is None: logger.info('''Initializing the config with a `BiT` backbone.''' ) A_ : Optional[Any] = { '''global_padding''': '''same''', '''layer_type''': '''bottleneck''', '''depths''': [3, 4, 9], '''out_features''': ['''stage1''', '''stage2''', '''stage3'''], '''embedding_dynamic_padding''': True, } A_ : List[str] = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): logger.info('''Initializing the config with a `BiT` backbone.''' ) A_ : Optional[Any] = BitConfig(**_lowerCamelCase ) elif isinstance(_lowerCamelCase , _lowerCamelCase ): A_ : List[str] = backbone_config else: raise ValueError( f'backbone_config must be a dictionary or a `PretrainedConfig`, got {backbone_config.__class__}.' ) A_ : Any = backbone_featmap_shape A_ : int = neck_ignore_stages if readout_type != "project": raise ValueError('''Readout type must be \'project\' when using `DPT-hybrid` mode.''' ) else: A_ : Tuple = None A_ : Dict = None A_ : Optional[Any] = [] A_ : Optional[int] = num_hidden_layers A_ : Optional[int] = num_attention_heads A_ : Any = intermediate_size A_ : Tuple = hidden_act A_ : Union[str, Any] = hidden_dropout_prob A_ : str = attention_probs_dropout_prob A_ : Union[str, Any] = initializer_range A_ : int = layer_norm_eps A_ : Tuple = image_size A_ : Optional[Any] = patch_size A_ : Dict = num_channels A_ : str = qkv_bias A_ : Tuple = backbone_out_indices if readout_type not in ["ignore", "add", "project"]: raise ValueError('''Readout_type must be one of [\'ignore\', \'add\', \'project\']''' ) A_ : Any = readout_type A_ : Any = reassemble_factors A_ : Any = neck_hidden_sizes A_ : str = fusion_hidden_size A_ : List[str] = head_in_index A_ : Optional[int] = use_batch_norm_in_fusion_residual # auxiliary head attributes (semantic segmentation) A_ : int = use_auxiliary_head A_ : Optional[int] = auxiliary_loss_weight A_ : Tuple = semantic_loss_ignore_index A_ : List[Any] = semantic_classifier_dropout def _a ( self : str ): """simple docstring""" A_ : int = copy.deepcopy(self.__dict__ ) if output["backbone_config"] is not None: A_ : Any = self.backbone_config.to_dict() A_ : Any = self.__class__.model_type return output
358
'''simple docstring''' import json import os from pathlib import Path from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple, Union import sentencepiece from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging snake_case__ = logging.get_logger(__name__) snake_case__ = """▁""" snake_case__ = { """vocab_file""": """vocab.json""", """spm_file""": """sentencepiece.bpe.model""", } snake_case__ = { """vocab_file""": { """facebook/s2t-small-librispeech-asr""": ( """https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/vocab.json""" ), }, """spm_file""": { """facebook/s2t-small-librispeech-asr""": ( """https://huggingface.co/facebook/s2t-small-librispeech-asr/resolve/main/sentencepiece.bpe.model""" ) }, } snake_case__ = { """facebook/s2t-small-librispeech-asr""": 10_24, } snake_case__ = ["""pt""", """fr""", """ru""", """nl""", """ro""", """it""", """es""", """de"""] snake_case__ = {"""mustc""": MUSTC_LANGS} class UpperCamelCase_ (a__ ): """simple docstring""" _lowerCAmelCase = VOCAB_FILES_NAMES _lowerCAmelCase = PRETRAINED_VOCAB_FILES_MAP _lowerCAmelCase = MAX_MODEL_INPUT_SIZES _lowerCAmelCase = ['input_ids', 'attention_mask'] _lowerCAmelCase = [] def __init__( self : Optional[int] , _lowerCamelCase : List[str] , _lowerCamelCase : List[str] , _lowerCamelCase : str="<s>" , _lowerCamelCase : Union[str, Any]="</s>" , _lowerCamelCase : Dict="<pad>" , _lowerCamelCase : str="<unk>" , _lowerCamelCase : Union[str, Any]=False , _lowerCamelCase : int=False , _lowerCamelCase : Any=None , _lowerCamelCase : Any=None , _lowerCamelCase : Optional[Dict[str, Any]] = None , **_lowerCamelCase : Optional[int] , ): """simple docstring""" A_ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=_lowerCamelCase , eos_token=_lowerCamelCase , unk_token=_lowerCamelCase , pad_token=_lowerCamelCase , do_upper_case=_lowerCamelCase , do_lower_case=_lowerCamelCase , tgt_lang=_lowerCamelCase , lang_codes=_lowerCamelCase , sp_model_kwargs=self.sp_model_kwargs , **_lowerCamelCase , ) A_ : Optional[int] = do_upper_case A_ : Tuple = do_lower_case A_ : Tuple = load_json(_lowerCamelCase ) A_ : Tuple = {v: k for k, v in self.encoder.items()} A_ : List[Any] = spm_file A_ : List[str] = load_spm(_lowerCamelCase , self.sp_model_kwargs ) if lang_codes is not None: A_ : Any = lang_codes A_ : Optional[Any] = LANGUAGES[lang_codes] A_ : Optional[Any] = [f'<lang:{lang}>' for lang in self.langs] A_ : Union[str, Any] = {lang: self.sp_model.PieceToId(f'<lang:{lang}>' ) for lang in self.langs} A_ : Optional[int] = self.lang_tokens A_ : int = tgt_lang if tgt_lang is not None else self.langs[0] self.set_tgt_lang_special_tokens(self._tgt_lang ) else: A_ : Dict = {} @property def _a ( self : Tuple ): """simple docstring""" return len(self.encoder ) @property def _a ( self : int ): """simple docstring""" return self._tgt_lang @tgt_lang.setter def _a ( self : List[str] , _lowerCamelCase : Any ): """simple docstring""" A_ : int = new_tgt_lang self.set_tgt_lang_special_tokens(_lowerCamelCase ) def _a ( self : Tuple , _lowerCamelCase : str ): """simple docstring""" A_ : List[str] = self.lang_code_to_id[tgt_lang] A_ : Optional[Any] = [lang_code_id] def _a ( self : Optional[Any] , _lowerCamelCase : str ): """simple docstring""" return self.sp_model.encode(_lowerCamelCase , out_type=_lowerCamelCase ) def _a ( self : List[Any] , _lowerCamelCase : int ): """simple docstring""" return self.encoder.get(_lowerCamelCase , self.encoder[self.unk_token] ) def _a ( self : int , _lowerCamelCase : int ): """simple docstring""" return self.decoder.get(_lowerCamelCase , self.unk_token ) def _a ( self : int , _lowerCamelCase : List[str] ): """simple docstring""" A_ : List[Any] = [] A_ : Any = '''''' for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: A_ : Union[str, Any] = self.sp_model.decode(_lowerCamelCase ) out_string += (decoded.upper() if self.do_upper_case else decoded) + token + " " A_ : Optional[Any] = [] else: current_sub_tokens.append(_lowerCamelCase ) A_ : Tuple = self.sp_model.decode(_lowerCamelCase ) out_string += decoded.upper() if self.do_upper_case else decoded return out_string.strip() def _a ( self : int , _lowerCamelCase : Dict , _lowerCamelCase : Any=None ): """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + [self.eos_token_id] def _a ( self : List[Any] , _lowerCamelCase : List[int] , _lowerCamelCase : Optional[List[int]] = None , _lowerCamelCase : bool = False ): """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=_lowerCamelCase , token_ids_a=_lowerCamelCase , already_has_special_tokens=_lowerCamelCase ) A_ : Tuple = [1] * len(self.prefix_tokens ) A_ : Tuple = [1] if token_ids_a is None: return prefix_ones + ([0] * len(_lowerCamelCase )) + suffix_ones return prefix_ones + ([0] * len(_lowerCamelCase )) + ([0] * len(_lowerCamelCase )) + suffix_ones def _a ( self : Dict ): """simple docstring""" A_ : Union[str, Any] = self.encoder.copy() vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : Union[str, Any] ): """simple docstring""" A_ : Dict = self.__dict__.copy() A_ : List[Any] = None return state def __setstate__( self : List[str] , _lowerCamelCase : Dict ): """simple docstring""" A_ : Dict = d # for backward compatibility if not hasattr(self , '''sp_model_kwargs''' ): A_ : Optional[int] = {} A_ : int = load_spm(self.spm_file , self.sp_model_kwargs ) def _a ( self : Optional[Any] , _lowerCamelCase : str , _lowerCamelCase : Optional[str] = None ): """simple docstring""" A_ : Dict = Path(_lowerCamelCase ) assert save_dir.is_dir(), f'{save_directory} should be a directory' A_ : Optional[int] = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''vocab_file'''] ) A_ : Optional[int] = save_dir / ( (filename_prefix + '''-''' if filename_prefix else '''''') + self.vocab_files_names['''spm_file'''] ) save_json(self.encoder , _lowerCamelCase ) if os.path.abspath(self.spm_file ) != os.path.abspath(_lowerCamelCase ) and os.path.isfile(self.spm_file ): copyfile(self.spm_file , _lowerCamelCase ) elif not os.path.isfile(self.spm_file ): with open(_lowerCamelCase , '''wb''' ) as fi: A_ : List[str] = self.sp_model.serialized_model_proto() fi.write(_lowerCamelCase ) return (str(_lowerCamelCase ), str(_lowerCamelCase )) def snake_case__ ( lowerCamelCase__ : str , lowerCamelCase__ : Dict[str, Any] ) -> sentencepiece.SentencePieceProcessor: A_ : Tuple = sentencepiece.SentencePieceProcessor(**lowerCamelCase__ ) spm.Load(str(lowerCamelCase__ ) ) return spm def snake_case__ ( lowerCamelCase__ : str ) -> Union[Dict, List]: with open(lowerCamelCase__ , '''r''' ) as f: return json.load(lowerCamelCase__ ) def snake_case__ ( lowerCamelCase__ : Any , lowerCamelCase__ : str ) -> None: with open(lowerCamelCase__ , '''w''' ) as f: json.dump(lowerCamelCase__ , lowerCamelCase__ , indent=2 )
4
0
"""simple docstring""" import copy import inspect import unittest from transformers import PretrainedConfig, SwiftFormerConfig from transformers.testing_utils import ( require_torch, require_vision, slow, torch_device, ) from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import SwiftFormerForImageClassification, SwiftFormerModel from transformers.models.swiftformer.modeling_swiftformer import SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase : '''simple docstring''' def __init__(self , a_ , a_=13 , a_=3 , a_=True , a_=True , a_=0.1 , a_=0.1 , a_=2_24 , a_=10_00 , a_=[3, 3, 6, 4] , a_=[48, 56, 1_12, 2_20] , ): '''simple docstring''' __snake_case : Any = parent __snake_case : Tuple = batch_size __snake_case : Dict = num_channels __snake_case : Dict = is_training __snake_case : List[Any] = use_labels __snake_case : Dict = hidden_dropout_prob __snake_case : Union[str, Any] = attention_probs_dropout_prob __snake_case : Optional[int] = num_labels __snake_case : str = image_size __snake_case : List[str] = layer_depths __snake_case : Dict = embed_dims def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : Optional[int] = None if self.use_labels: __snake_case : List[str] = ids_tensor([self.batch_size] , self.num_labels ) __snake_case : Dict = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return SwiftFormerConfig( depths=self.layer_depths , embed_dims=self.embed_dims , mlp_ratio=4 , downsamples=[True, True, True, True] , hidden_act='''gelu''' , num_labels=self.num_labels , down_patch_size=3 , down_stride=2 , down_pad=1 , drop_rate=0.0 , drop_path_rate=0.0 , use_layer_scale=a_ , layer_scale_init_value=1E-5 , ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ ): '''simple docstring''' __snake_case : str = SwiftFormerModel(config=a_ ) model.to(a_ ) model.eval() __snake_case : Union[str, Any] = model(a_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dims[-1], 7, 7) ) def SCREAMING_SNAKE_CASE (self , a_ , a_ , a_ ): '''simple docstring''' __snake_case : Dict = self.num_labels __snake_case : Tuple = SwiftFormerForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case : Any = model(a_ , labels=a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) __snake_case : Optional[Any] = SwiftFormerForImageClassification(a_ ) model.to(a_ ) model.eval() __snake_case : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) __snake_case : List[str] = model(a_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' ((__snake_case) , (__snake_case) , (__snake_case)) : Optional[int] = self.prepare_config_and_inputs() __snake_case : Optional[Any] = {'''pixel_values''': pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( __snake_case, __snake_case, unittest.TestCase ): '''simple docstring''' lowerCamelCase__ =(SwiftFormerModel, SwiftFormerForImageClassification) if is_torch_available() else () lowerCamelCase__ =( {'feature-extraction': SwiftFormerModel, 'image-classification': SwiftFormerForImageClassification} if is_torch_available() else {} ) lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False lowerCamelCase__ =False def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Union[str, Any] = SwiftFormerModelTester(self ) __snake_case : Tuple = ConfigTester( self , config_class=a_ , has_text_modality=a_ , hidden_size=37 , num_attention_heads=12 , num_hidden_layers=12 , ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason='''SwiftFormer does not use inputs_embeds''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case , __snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : int = model_class(a_ ) __snake_case : List[str] = model.get_output_embeddings() self.assertTrue(x is None or isinstance(a_ , nn.Linear ) ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case , __snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Dict = model_class(a_ ) __snake_case : List[Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic __snake_case : List[str] = [*signature.parameters.keys()] __snake_case : List[str] = ['''pixel_values'''] self.assertListEqual(arg_names[:1] , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*a_ ) @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' for model_name in SWIFTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: __snake_case : int = SwiftFormerModel.from_pretrained(a_ ) self.assertIsNotNone(a_ ) @unittest.skip(reason='''SwiftFormer does not output attentions''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' def check_hidden_states_output(a_ , a_ , a_ ): __snake_case : List[str] = model_class(a_ ) model.to(a_ ) model.eval() with torch.no_grad(): __snake_case : Union[str, Any] = model(**self._prepare_for_class(a_ , a_ ) ) __snake_case : Optional[Any] = outputs.hidden_states __snake_case : Union[str, Any] = 8 self.assertEqual(len(a_ ) , a_ ) # TODO # SwiftFormer's feature maps are of shape (batch_size, embed_dims, height, width) # with the width and height being successively divided by 2, after every 2 blocks for i in range(len(a_ ) ): self.assertEqual( hidden_states[i].shape , torch.Size( [ self.model_tester.batch_size, self.model_tester.embed_dims[i // 2], (self.model_tester.image_size // 4) // 2 ** (i // 2), (self.model_tester.image_size // 4) // 2 ** (i // 2), ] ) , ) __snake_case , __snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: __snake_case : Any = True check_hidden_states_output(a_ , a_ , a_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] __snake_case : int = True check_hidden_states_output(a_ , a_ , a_ ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' def _config_zero_init(a_ ): __snake_case : Optional[Any] = copy.deepcopy(a_ ) for key in configs_no_init.__dict__.keys(): if "_range" in key or "_std" in key or "initializer_factor" in key or "layer_scale" in key: setattr(a_ , a_ , 1E-10 ) if isinstance(getattr(a_ , a_ , a_ ) , a_ ): __snake_case : Union[str, Any] = _config_zero_init(getattr(a_ , a_ ) ) setattr(a_ , a_ , a_ ) return configs_no_init __snake_case , __snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() __snake_case : Dict = _config_zero_init(a_ ) for model_class in self.all_model_classes: __snake_case : Dict = model_class(config=a_ ) for name, param in model.named_parameters(): if param.requires_grad: self.assertIn( ((param.data.mean() * 1E9) / 1E9).round().item() , [0.0, 1.0] , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) @unittest.skip('''Will be fixed soon by reducing the size of the model used for common tests.''' ) def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' pass def lowercase ( ) ->List[Any]: """simple docstring""" __snake_case : int = Image.open('''./tests/fixtures/tests_samples/COCO/000000039769.png''' ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase ): '''simple docstring''' @cached_property def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' return ViTImageProcessor.from_pretrained('''MBZUAI/swiftformer-xs''' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE (self ): '''simple docstring''' __snake_case : List[Any] = SwiftFormerForImageClassification.from_pretrained('''MBZUAI/swiftformer-xs''' ).to(a_ ) __snake_case : Optional[int] = self.default_image_processor __snake_case : Dict = prepare_img() __snake_case : Tuple = image_processor(images=a_ , return_tensors='''pt''' ).to(a_ ) # forward pass with torch.no_grad(): __snake_case : Any = model(**a_ ) # verify the logits __snake_case : int = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , a_ ) __snake_case : List[Any] = torch.tensor([[-2.1703E00, 2.1107E00, -2.0811E00]] ).to(a_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , a_ , atol=1E-4 ) )
102
"""simple docstring""" import math def lowercase ( _snake_case : int ) ->bool: """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(_snake_case ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def lowercase ( _snake_case : float = 0.1 ) ->int: """simple docstring""" __snake_case : Tuple = 3 __snake_case : Any = 3 while primes / (2 * j - 1) >= ratio: for i in range(j * j + j + 1 , (j + 2) * (j + 2) , j + 1 ): primes += is_prime(_snake_case ) j += 2 return j if __name__ == "__main__": import doctest doctest.testmod()
102
1
'''simple docstring''' import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal A : Optional[int] = datasets.utils.logging.get_logger(__name__) A : List[Any] = ['''names''', '''prefix'''] A : Tuple = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] A : List[Any] = ['''encoding_errors''', '''on_bad_lines'''] A : Union[str, Any] = ['''date_format'''] @dataclass class __lowerCamelCase ( datasets.BuilderConfig ): """simple docstring""" a = "," a = None a = "infer" a = None a = None a = None a = None a = None a = True a = None a = None a = None a = None a = False a = None a = None a = None a = True a = True a = False a = True a = None a = "." a = None a = '"' a = 0 a = None a = None a = None a = None a = True a = True a = 0 a = True a = False a = None a = 1_0000 a = None a = "strict" a = "error" a = None def A ( self : str): if self.delimiter is not None: _A : Any = self.delimiter if self.column_names is not None: _A : List[Any] = self.column_names @property def A ( self : Optional[int]): _A : Dict = { "sep": self.sep, "header": self.header, "names": self.names, "index_col": self.index_col, "usecols": self.usecols, "prefix": self.prefix, "mangle_dupe_cols": self.mangle_dupe_cols, "engine": self.engine, "converters": self.converters, "true_values": self.true_values, "false_values": self.false_values, "skipinitialspace": self.skipinitialspace, "skiprows": self.skiprows, "nrows": self.nrows, "na_values": self.na_values, "keep_default_na": self.keep_default_na, "na_filter": self.na_filter, "verbose": self.verbose, "skip_blank_lines": self.skip_blank_lines, "thousands": self.thousands, "decimal": self.decimal, "lineterminator": self.lineterminator, "quotechar": self.quotechar, "quoting": self.quoting, "escapechar": self.escapechar, "comment": self.comment, "encoding": self.encoding, "dialect": self.dialect, "error_bad_lines": self.error_bad_lines, "warn_bad_lines": self.warn_bad_lines, "skipfooter": self.skipfooter, "doublequote": self.doublequote, "memory_map": self.memory_map, "float_precision": self.float_precision, "chunksize": self.chunksize, "encoding_errors": self.encoding_errors, "on_bad_lines": self.on_bad_lines, "date_format": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , lowerCAmelCase__): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __lowerCamelCase ( datasets.ArrowBasedBuilder ): """simple docstring""" a = CsvConfig def A ( self : Optional[Any]): return datasets.DatasetInfo(features=self.config.features) def A ( self : Tuple , SCREAMING_SNAKE_CASE : str): if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}') _A : List[str] = dl_manager.download_and_extract(self.config.data_files) if isinstance(lowerCAmelCase__ , (str, list, tuple)): _A : int = data_files if isinstance(lowerCAmelCase__ , lowerCAmelCase__): _A : Any = [files] _A : List[Any] = [dl_manager.iter_files(lowerCAmelCase__) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={'files': files})] _A : Optional[Any] = [] for split_name, files in data_files.items(): if isinstance(lowerCAmelCase__ , lowerCAmelCase__): _A : str = [files] _A : Any = [dl_manager.iter_files(lowerCAmelCase__) for file in files] splits.append(datasets.SplitGenerator(name=lowerCAmelCase__ , gen_kwargs={'files': files})) return splits def A ( self : List[Any] , SCREAMING_SNAKE_CASE : pa.Table): if self.config.features is not None: _A : Tuple = self.config.features.arrow_schema if all(not require_storage_cast(lowerCAmelCase__) for feature in self.config.features.values()): # cheaper cast _A : Any = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=lowerCAmelCase__) else: # more expensive cast; allows str <-> int/float or str to Audio for example _A : int = table_cast(lowerCAmelCase__ , lowerCAmelCase__) return pa_table def A ( self : Dict , SCREAMING_SNAKE_CASE : Dict): _A : int = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str _A : Optional[Any] = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(lowerCAmelCase__) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values()) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(lowerCAmelCase__)): _A : Optional[Any] = pd.read_csv(lowerCAmelCase__ , iterator=lowerCAmelCase__ , dtype=lowerCAmelCase__ , **self.config.pd_read_csv_kwargs) try: for batch_idx, df in enumerate(lowerCAmelCase__): _A : Optional[int] = pa.Table.from_pandas(lowerCAmelCase__) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(lowerCAmelCase__) except ValueError as e: logger.error(F'Failed to read file \'{file}\' with error {type(lowerCAmelCase__)}: {e}') raise
367
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available A : Optional[int] = { '''configuration_mask2former''': [ '''MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Mask2FormerConfig''', ], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Tuple = ['''Mask2FormerImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A : Dict = [ '''MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Mask2FormerForUniversalSegmentation''', '''Mask2FormerModel''', '''Mask2FormerPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_maskaformer import MASK2FORMER_PRETRAINED_CONFIG_ARCHIVE_MAP, MaskaFormerConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_maskaformer import MaskaFormerImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_maskaformer import ( MASK2FORMER_PRETRAINED_MODEL_ARCHIVE_LIST, MaskaFormerForUniversalSegmentation, MaskaFormerModel, MaskaFormerPreTrainedModel, ) else: import sys A : Optional[Any] = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
227
0
from collections.abc import Generator from math import sin def A ( _lowerCamelCase ): '''simple docstring''' if len(_lowerCamelCase ) != 32: raise ValueError("Input must be of length 32" ) _lowerCAmelCase : Optional[Any] = b"" for i in [3, 2, 1, 0]: little_endian += string_aa[8 * i : 8 * i + 8] return little_endian def A ( _lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) _lowerCAmelCase : Dict = format(_lowerCamelCase , "08x" )[-8:] _lowerCAmelCase : int = b"" for i in [3, 2, 1, 0]: little_endian_hex += hex_rep[2 * i : 2 * i + 2].encode("utf-8" ) return little_endian_hex def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Tuple = b"" for char in message: bit_string += format(_lowerCamelCase , "08b" ).encode("utf-8" ) _lowerCAmelCase : Tuple = format(len(_lowerCamelCase ) , "064b" ).encode("utf-8" ) # Pad bit_string to a multiple of 512 chars bit_string += b"1" while len(_lowerCamelCase ) % 512 != 448: bit_string += b"0" bit_string += to_little_endian(start_len[32:] ) + to_little_endian(start_len[:32] ) return bit_string def A ( _lowerCamelCase ): '''simple docstring''' if len(_lowerCamelCase ) % 512 != 0: raise ValueError("Input must have length that's a multiple of 512" ) for pos in range(0 , len(_lowerCamelCase ) , 512 ): _lowerCAmelCase : List[str] = bit_string[pos : pos + 512] _lowerCAmelCase : Optional[Any] = [] for i in range(0 , 512 , 32 ): block_words.append(int(to_little_endian(block[i : i + 32] ) , 2 ) ) yield block_words def A ( _lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) _lowerCAmelCase : str = format(_lowerCamelCase , "032b" ) _lowerCAmelCase : Optional[Any] = "" for c in i_str: new_str += "1" if c == "0" else "0" return int(_lowerCamelCase , 2 ) def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' return (a + b) % 2**32 def A ( _lowerCamelCase , _lowerCamelCase ): '''simple docstring''' if i < 0: raise ValueError("Input must be non-negative" ) if shift < 0: raise ValueError("Shift must be non-negative" ) return ((i << shift) ^ (i >> (32 - shift))) % 2**32 def A ( _lowerCamelCase ): '''simple docstring''' _lowerCAmelCase : Dict = preprocess(_lowerCamelCase ) _lowerCAmelCase : Tuple = [int(2**32 * abs(sin(i + 1 ) ) ) for i in range(64 )] # Starting states _lowerCAmelCase : int = 0X67_45_23_01 _lowerCAmelCase : Optional[Any] = 0XEF_CD_AB_89 _lowerCAmelCase : Union[str, Any] = 0X98_BA_DC_FE _lowerCAmelCase : Dict = 0X10_32_54_76 _lowerCAmelCase : int = [ 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 7, 12, 17, 22, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 5, 9, 14, 20, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 4, 11, 16, 23, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, 6, 10, 15, 21, ] # Process bit string in chunks, each with 16 32-char words for block_words in get_block_words(_lowerCamelCase ): _lowerCAmelCase : Dict = aa _lowerCAmelCase : Union[str, Any] = ba _lowerCAmelCase : Dict = ca _lowerCAmelCase : Tuple = da # Hash current chunk for i in range(64 ): if i <= 15: # f = (b & c) | (not_32(b) & d) # Alternate definition for f _lowerCAmelCase : List[Any] = d ^ (b & (c ^ d)) _lowerCAmelCase : int = i elif i <= 31: # f = (d & b) | (not_32(d) & c) # Alternate definition for f _lowerCAmelCase : List[str] = c ^ (d & (b ^ c)) _lowerCAmelCase : Dict = (5 * i + 1) % 16 elif i <= 47: _lowerCAmelCase : Union[str, Any] = b ^ c ^ d _lowerCAmelCase : Optional[int] = (3 * i + 5) % 16 else: _lowerCAmelCase : List[str] = c ^ (b | not_aa(_lowerCamelCase )) _lowerCAmelCase : int = (7 * i) % 16 _lowerCAmelCase : int = (f + a + added_consts[i] + block_words[g]) % 2**32 _lowerCAmelCase : Optional[int] = d _lowerCAmelCase : Tuple = c _lowerCAmelCase : int = b _lowerCAmelCase : Tuple = sum_aa(_lowerCamelCase , left_rotate_aa(_lowerCamelCase , shift_amounts[i] ) ) # Add hashed chunk to running total _lowerCAmelCase : List[str] = sum_aa(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : Any = sum_aa(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : Any = sum_aa(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : str = sum_aa(_lowerCamelCase , _lowerCamelCase ) _lowerCAmelCase : Union[str, Any] = reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) + reformat_hex(_lowerCamelCase ) return digest if __name__ == "__main__": import doctest doctest.testmod()
36
'''simple docstring''' import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( SwiftFormerConfig, SwiftFormerForImageClassification, ViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() __lowercase : Tuple = logging.get_logger(__name__) __lowercase : List[Any] = torch.device('cpu') def lowerCamelCase (): __a : int = 'http://images.cocodataset.org/val2017/000000039769.jpg' __a : Tuple = Image.open(requests.get(_SCREAMING_SNAKE_CASE , stream=_SCREAMING_SNAKE_CASE ).raw ) return im def lowerCamelCase (_SCREAMING_SNAKE_CASE : Union[str, Any] ): if swiftformer_name == "swiftformer_xs": return torch.tensor([-2.1_7_0_3e0_0, 2.1_1_0_7e0_0, -2.0_8_1_1e0_0, 8.8_6_8_5e-0_1, 2.4_3_6_0e-0_1] ) elif swiftformer_name == "swiftformer_s": return torch.tensor([3.9_6_3_6e-0_1, 2.3_4_7_8e-0_1, -1.6_9_6_3e0_0, -1.7_3_8_1e0_0, -8.6_3_3_7e-0_1] ) elif swiftformer_name == "swiftformer_l1": return torch.tensor([-4.2_7_6_8e-0_1, -4.7_4_2_9e-0_1, -1.0_8_9_7e0_0, -1.0_2_4_8e0_0, 3.5_5_2_3e-0_2] ) elif swiftformer_name == "swiftformer_l3": return torch.tensor([-2.5_3_3_0e-0_1, 2.4_2_1_1e-0_1, -6.0_1_8_5e-0_1, -8.2_7_8_9e-0_1, -6.0_4_4_6e-0_2] ) def lowerCamelCase (_SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Tuple , _SCREAMING_SNAKE_CASE : Union[str, Any] ): __a : int = dct.pop(_SCREAMING_SNAKE_CASE ) __a : Tuple = val def lowerCamelCase (_SCREAMING_SNAKE_CASE : str ): __a : Dict = [] for k in state_dict.keys(): __a : List[Any] = k if ".pwconv" in k: __a : List[Any] = k_new.replace('.pwconv' , '.point_wise_conv' ) if ".dwconv" in k: __a : Dict = k_new.replace('.dwconv' , '.depth_wise_conv' ) if ".Proj." in k: __a : Optional[int] = k_new.replace('.Proj.' , '.proj.' ) if "patch_embed" in k_new: __a : List[Any] = k_new.replace('patch_embed' , 'swiftformer.patch_embed.patch_embedding' ) if "network" in k_new: __a : Union[str, Any] = k_new.split('.' ) if ls[2].isdigit(): __a : Union[str, Any] = 'swiftformer.encoder.network.' + ls[1] + '.blocks.' + ls[2] + '.' + '.'.join(ls[3:] ) else: __a : Union[str, Any] = k_new.replace('network' , 'swiftformer.encoder.network' ) rename_keys.append((k, k_new) ) return rename_keys @torch.no_grad() def lowerCamelCase (_SCREAMING_SNAKE_CASE : str , _SCREAMING_SNAKE_CASE : Union[str, Any] , _SCREAMING_SNAKE_CASE : Union[str, Any] ): __a : Union[str, Any] = SwiftFormerConfig() # dataset (ImageNet-21k only or also fine-tuned on ImageNet 2012), patch_size and image_size __a : List[str] = 1_000 __a : Tuple = 'huggingface/label-files' __a : str = 'imagenet-1k-id2label.json' __a : Dict = json.load(open(hf_hub_download(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , repo_type='dataset' ) , 'r' ) ) __a : Optional[int] = {int(_SCREAMING_SNAKE_CASE ): v for k, v in idalabel.items()} __a : Any = idalabel __a : str = {v: k for k, v in idalabel.items()} # size of the architecture if swiftformer_name == "swiftformer_xs": __a : Dict = [3, 3, 6, 4] __a : int = [48, 56, 112, 220] elif swiftformer_name == "swiftformer_s": __a : Dict = [3, 3, 9, 6] __a : List[str] = [48, 64, 168, 224] elif swiftformer_name == "swiftformer_l1": __a : Dict = [4, 3, 10, 5] __a : Optional[int] = [48, 96, 192, 384] elif swiftformer_name == "swiftformer_l3": __a : Tuple = [4, 4, 12, 6] __a : Dict = [64, 128, 320, 512] # load state_dict of original model, remove and rename some keys if original_ckpt: if original_ckpt.startswith('https' ): __a : List[Any] = torch.hub.load_state_dict_from_url(_SCREAMING_SNAKE_CASE , map_location='cpu' , check_hash=_SCREAMING_SNAKE_CASE ) else: __a : Union[str, Any] = torch.load(_SCREAMING_SNAKE_CASE , map_location='cpu' ) __a : Optional[Any] = checkpoint __a : Dict = create_rename_keys(_SCREAMING_SNAKE_CASE ) for rename_key_src, rename_key_dest in rename_keys: rename_key(_SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE , _SCREAMING_SNAKE_CASE ) # load HuggingFace model __a : Tuple = SwiftFormerForImageClassification(_SCREAMING_SNAKE_CASE ).eval() hf_model.load_state_dict(_SCREAMING_SNAKE_CASE ) # prepare test inputs __a : Tuple = prepare_img() __a : str = ViTImageProcessor.from_pretrained('preprocessor_config' ) __a : Tuple = processor(images=_SCREAMING_SNAKE_CASE , return_tensors='pt' ) # compare outputs from both models __a : List[Any] = get_expected_output(_SCREAMING_SNAKE_CASE ) __a : Dict = hf_model(inputs['pixel_values'] ).logits assert hf_logits.shape == torch.Size([1, 1_000] ) assert torch.allclose(hf_logits[0, 0:5] , _SCREAMING_SNAKE_CASE , atol=1e-3 ) Path(_SCREAMING_SNAKE_CASE ).mkdir(exist_ok=_SCREAMING_SNAKE_CASE ) print(F"""Saving model {swiftformer_name} to {pytorch_dump_folder_path}""" ) hf_model.save_pretrained(_SCREAMING_SNAKE_CASE ) if __name__ == "__main__": __lowercase : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( '--swiftformer_name', default='swiftformer_xs', choices=['swiftformer_xs', 'swiftformer_s', 'swiftformer_l1', 'swiftformer_l3'], type=str, help='Name of the SwiftFormer model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='./converted_outputs/', type=str, help='Path to the output PyTorch model directory.', ) parser.add_argument('--original_ckpt', default=None, type=str, help='Path to the original model checkpoint.') __lowercase : Tuple = parser.parse_args() convert_swiftformer_checkpoint(args.swiftformer_name, args.pytorch_dump_folder_path, args.original_ckpt)
27
0
"""simple docstring""" import unittest from transformers import is_flax_available from transformers.testing_utils import require_flax, require_sentencepiece, require_tokenizers, require_torch, slow if is_flax_available(): import optax from flax.training.common_utils import onehot from transformers import AutoTokenizer, FlaxMTaForConditionalGeneration from transformers.models.ta.modeling_flax_ta import shift_tokens_right @require_torch @require_sentencepiece @require_tokenizers @require_flax class A_ (unittest.TestCase ): '''simple docstring''' @slow def UpperCamelCase__ ( self ): """simple docstring""" UpperCAmelCase_ : List[str] = FlaxMTaForConditionalGeneration.from_pretrained("google/mt5-small" ) UpperCAmelCase_ : Any = AutoTokenizer.from_pretrained("google/mt5-small" ) UpperCAmelCase_ : int = tokenizer("Hello there" , return_tensors="np" ).input_ids UpperCAmelCase_ : List[str] = tokenizer("Hi I am" , return_tensors="np" ).input_ids UpperCAmelCase_ : Any = shift_tokens_right(lowercase_ , model.config.pad_token_id , model.config.decoder_start_token_id ) UpperCAmelCase_ : Dict = model(lowercase_ , decoder_input_ids=lowercase_ ).logits UpperCAmelCase_ : Tuple = optax.softmax_cross_entropy(lowercase_ , onehot(lowercase_ , logits.shape[-1] ) ).mean() UpperCAmelCase_ : str = -(labels.shape[-1] * loss.item()) UpperCAmelCase_ : str = -84.91_27 self.assertTrue(abs(mtf_score - EXPECTED_SCORE ) < 1E-4 )
357
"""simple docstring""" import datasets _a = '\\n@InProceedings{conneau2018xnli,\n author = "Conneau, Alexis\n and Rinott, Ruty\n and Lample, Guillaume\n and Williams, Adina\n and Bowman, Samuel R.\n and Schwenk, Holger\n and Stoyanov, Veselin",\n title = "XNLI: Evaluating Cross-lingual Sentence Representations",\n booktitle = "Proceedings of the 2018 Conference on Empirical Methods\n in Natural Language Processing",\n year = "2018",\n publisher = "Association for Computational Linguistics",\n location = "Brussels, Belgium",\n}\n' _a = '\\nXNLI is a subset of a few thousand examples from MNLI which has been translated\ninto a 14 different languages (some low-ish resource). As with MNLI, the goal is\nto predict textual entailment (does sentence A imply/contradict/neither sentence\nB) and is a classification task (given two sentences, predict one of three\nlabels).\n' _a = '\nComputes XNLI score which is just simple accuracy.\nArgs:\n predictions: Predicted labels.\n references: Ground truth labels.\nReturns:\n \'accuracy\': accuracy\nExamples:\n\n >>> predictions = [0, 1]\n >>> references = [0, 1]\n >>> xnli_metric = datasets.load_metric("xnli")\n >>> results = xnli_metric.compute(predictions=predictions, references=references)\n >>> print(results)\n {\'accuracy\': 1.0}\n' def __a ( __lowerCamelCase, __lowerCamelCase ): return (preds == labels).mean() @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION ,_KWARGS_DESCRIPTION ) class A_ (datasets.Metric ): '''simple docstring''' def UpperCamelCase__ ( self ): """simple docstring""" return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), "references": datasets.Value("int64" if self.config_name != "sts-b" else "float32" ), } ) , codebase_urls=[] , reference_urls=[] , format="numpy" , ) def UpperCamelCase__ ( self , lowercase_ , lowercase_ ): """simple docstring""" return {"accuracy": simple_accuracy(lowercase_ , lowercase_ )}
23
0
import copy import tempfile import unittest from transformers import MaMaaaConfig, is_torch_available from transformers.testing_utils import require_sentencepiece, require_tokenizers, require_torch, slow, torch_device from transformers.utils import cached_property from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaTokenizer from transformers.models.mam_aaa.modeling_mam_aaa import MaMaaaDecoder, MaMaaaEncoder def UpperCAmelCase_ ( _A , _A , _A , _A=None , _A=None , _A=None , _A=None , _A=None , ): '''simple docstring''' if attention_mask is None: SCREAMING_SNAKE_CASE__ = input_ids.ne(config.pad_token_id ) if decoder_attention_mask is None: SCREAMING_SNAKE_CASE__ = decoder_input_ids.ne(config.pad_token_id ) if head_mask is None: SCREAMING_SNAKE_CASE__ = torch.ones(config.encoder_layers , config.encoder_attention_heads , device=_A ) if decoder_head_mask is None: SCREAMING_SNAKE_CASE__ = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_A ) if cross_attn_head_mask is None: SCREAMING_SNAKE_CASE__ = torch.ones(config.decoder_layers , config.decoder_attention_heads , device=_A ) return { "input_ids": input_ids, "decoder_input_ids": decoder_input_ids, "attention_mask": attention_mask, "decoder_attention_mask": attention_mask, "head_mask": head_mask, "decoder_head_mask": decoder_head_mask, "cross_attn_head_mask": cross_attn_head_mask, } class UpperCAmelCase__ : """simple docstring""" def __init__( self : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Dict=13 , __lowerCamelCase : Tuple=7 , __lowerCamelCase : Dict=True , __lowerCamelCase : Dict=False , __lowerCamelCase : Optional[Any]=99 , __lowerCamelCase : Optional[Any]=16 , __lowerCamelCase : Optional[int]=2 , __lowerCamelCase : Tuple=4 , __lowerCamelCase : Optional[int]=4 , __lowerCamelCase : List[Any]="relu" , __lowerCamelCase : Any=0.1 , __lowerCamelCase : Optional[Any]=0.1 , __lowerCamelCase : Any=0.0 , __lowerCamelCase : Tuple=0.0 , __lowerCamelCase : Optional[Any]=20 , __lowerCamelCase : Tuple=2 , __lowerCamelCase : Dict=1 , __lowerCamelCase : List[Any]=0 , ) -> str: SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = encoder_layerdrop SCREAMING_SNAKE_CASE__ = decoder_layerdrop SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = eos_token_id SCREAMING_SNAKE_CASE__ = pad_token_id SCREAMING_SNAKE_CASE__ = bos_token_id def lowercase_ ( self : Tuple ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = self.eos_token_id # Eos Token SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) # we need to clamp the input ids here to avoid having pad token in between # this is because for M2M100 the position_ids are prepared such that # all pad tokens have pos id = 2 and rest are between 2..seq_length # and the seq_length here is seq_length - num_pad_tokens # but when using past, there is no way of knowing if the past input ids had # pad tokens in them, which results in incorrect seq_lenth and which in turn results in # position_ids being off by num_pad_tokens in past input SCREAMING_SNAKE_CASE__ = input_ids.clamp(self.pad_token_id + 1 ) SCREAMING_SNAKE_CASE__ = decoder_input_ids.clamp(self.pad_token_id + 1 ) SCREAMING_SNAKE_CASE__ = self.get_config() SCREAMING_SNAKE_CASE__ = prepare_mam_aaa_inputs_dict(__lowerCamelCase , __lowerCamelCase , __lowerCamelCase ) return config, inputs_dict def lowercase_ ( self : Any ) -> str: return MaMaaaConfig( vocab_size=self.vocab_size , d_model=self.hidden_size , encoder_layers=self.num_hidden_layers , decoder_layers=self.num_hidden_layers , encoder_attention_heads=self.num_attention_heads , decoder_attention_heads=self.num_attention_heads , encoder_ffn_dim=self.intermediate_size , decoder_ffn_dim=self.intermediate_size , dropout=self.hidden_dropout_prob , attention_dropout=self.attention_probs_dropout_prob , encoder_layerdrop=self.encoder_layerdrop , decoder_layerdrop=self.decoder_layerdrop , max_position_embeddings=self.max_position_embeddings , eos_token_id=self.eos_token_id , bos_token_id=self.bos_token_id , pad_token_id=self.pad_token_id , ) def lowercase_ ( self : int ) -> int: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() return config, inputs_dict def lowercase_ ( self : Tuple , __lowerCamelCase : Optional[int] , __lowerCamelCase : List[Any] ) -> Any: SCREAMING_SNAKE_CASE__ = MaMaaaModel(config=__lowerCamelCase ).get_decoder().to(__lowerCamelCase ).eval() SCREAMING_SNAKE_CASE__ = inputs_dict['''input_ids'''] SCREAMING_SNAKE_CASE__ = inputs_dict['''attention_mask'''] SCREAMING_SNAKE_CASE__ = inputs_dict['''head_mask'''] # first forward pass SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase , attention_mask=__lowerCamelCase , head_mask=__lowerCamelCase , use_cache=__lowerCamelCase ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids SCREAMING_SNAKE_CASE__ = ids_tensor((self.batch_size, 3) , config.vocab_size ) SCREAMING_SNAKE_CASE__ = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and SCREAMING_SNAKE_CASE__ = torch.cat([input_ids, next_tokens] , dim=-1 ) SCREAMING_SNAKE_CASE__ = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase , attention_mask=__lowerCamelCase )['''last_hidden_state'''] SCREAMING_SNAKE_CASE__ = model(__lowerCamelCase , attention_mask=__lowerCamelCase , past_key_values=__lowerCamelCase )[ '''last_hidden_state''' ] # select random slice SCREAMING_SNAKE_CASE__ = ids_tensor((1,) , output_from_past.shape[-1] ).item() SCREAMING_SNAKE_CASE__ = output_from_no_past[:, -3:, random_slice_idx].detach() SCREAMING_SNAKE_CASE__ = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(__lowerCamelCase , __lowerCamelCase , atol=1e-2 ) ) def lowercase_ ( self : List[str] , __lowerCamelCase : Any , __lowerCamelCase : Tuple ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = MaMaaaModel(config=__lowerCamelCase ).to(__lowerCamelCase ).eval() SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = outputs.encoder_last_hidden_state SCREAMING_SNAKE_CASE__ = outputs.last_hidden_state with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ = model.get_encoder() encoder.save_pretrained(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = MaMaaaEncoder.from_pretrained(__lowerCamelCase ).to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = encoder(inputs_dict['''input_ids'''] , attention_mask=inputs_dict['''attention_mask'''] )[ 0 ] self.parent.assertTrue((encoder_last_hidden_state_a - encoder_last_hidden_state).abs().max().item() < 1e-3 ) with tempfile.TemporaryDirectory() as tmpdirname: SCREAMING_SNAKE_CASE__ = model.get_decoder() decoder.save_pretrained(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = MaMaaaDecoder.from_pretrained(__lowerCamelCase ).to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = decoder( input_ids=inputs_dict['''decoder_input_ids'''] , attention_mask=inputs_dict['''decoder_attention_mask'''] , encoder_hidden_states=__lowerCamelCase , encoder_attention_mask=inputs_dict['''attention_mask'''] , )[0] self.parent.assertTrue((last_hidden_state_a - last_hidden_state).abs().max().item() < 1e-3 ) @require_torch class UpperCAmelCase__ ( A__ , A__ , A__ , unittest.TestCase ): """simple docstring""" a = ( ( MaMaaaModel, MaMaaaForConditionalGeneration, ) if is_torch_available() else () ) a = (MaMaaaForConditionalGeneration,) if is_torch_available() else () a = ( { "conversational": MaMaaaForConditionalGeneration, "feature-extraction": MaMaaaModel, "summarization": MaMaaaForConditionalGeneration, "text2text-generation": MaMaaaForConditionalGeneration, "translation": MaMaaaForConditionalGeneration, } if is_torch_available() else {} ) a = True a = True a = False a = False def lowercase_ ( self : Union[str, Any] , __lowerCamelCase : Tuple , __lowerCamelCase : int , __lowerCamelCase : Optional[Any] , __lowerCamelCase : Optional[int] , __lowerCamelCase : int ) -> Dict: if pipeline_test_casse_name == "TranslationPipelineTests": # Get `ValueError: Translation requires a `src_lang` and a `tgt_lang` for this model`. # `M2M100Config` was never used in pipeline tests: cannot create a simple tokenizer. return True return False def lowercase_ ( self : Union[str, Any] ) -> str: SCREAMING_SNAKE_CASE__ = MaMaaaModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=__lowerCamelCase ) def lowercase_ ( self : str ) -> List[Any]: self.config_tester.run_common_tests() def lowercase_ ( self : int ) -> int: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() for model_class in self.all_model_classes: SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__lowerCamelCase ) SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = model_class.from_pretrained(__lowerCamelCase , output_loading_info=__lowerCamelCase ) self.assertEqual(info['''missing_keys'''] , [] ) def lowercase_ ( self : List[Any] ) -> Optional[int]: SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(*__lowerCamelCase ) def lowercase_ ( self : Union[str, Any] ) -> Optional[Any]: SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() self.model_tester.check_encoder_decoder_model_standalone(*__lowerCamelCase ) def lowercase_ ( self : List[Any] ) -> List[Any]: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in (MaMaaaModel, MaMaaaForConditionalGeneration): SCREAMING_SNAKE_CASE__ = model_class(__lowerCamelCase ) model.to(__lowerCamelCase ) model.eval() SCREAMING_SNAKE_CASE__ = copy.deepcopy(self._prepare_for_class(__lowerCamelCase , __lowerCamelCase ) ) if not self.is_encoder_decoder: SCREAMING_SNAKE_CASE__ = inputs['''input_ids'''] del inputs["input_ids"] else: SCREAMING_SNAKE_CASE__ = inputs['''input_ids'''] SCREAMING_SNAKE_CASE__ = inputs.get('''decoder_input_ids''' , __lowerCamelCase ) del inputs["input_ids"] inputs.pop('''decoder_input_ids''' , __lowerCamelCase ) SCREAMING_SNAKE_CASE__ = model.get_input_embeddings() if not self.is_encoder_decoder: SCREAMING_SNAKE_CASE__ = wte(__lowerCamelCase ) else: SCREAMING_SNAKE_CASE__ = wte(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = wte(__lowerCamelCase ) with torch.no_grad(): model(**__lowerCamelCase )[0] def lowercase_ ( self : Any ) -> Union[str, Any]: SCREAMING_SNAKE_CASE__,SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() SCREAMING_SNAKE_CASE__ = input_dict['''input_ids'''] SCREAMING_SNAKE_CASE__ = input_ids.ne(1 ).to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = MaMaaaForConditionalGeneration(__lowerCamelCase ).eval().to(__lowerCamelCase ) if torch_device == "cuda": model.half() model.generate(__lowerCamelCase , attention_mask=__lowerCamelCase ) model.generate(num_beams=4 , do_sample=__lowerCamelCase , early_stopping=__lowerCamelCase , num_return_sequences=3 ) def UpperCAmelCase_ ( _A ): '''simple docstring''' return torch.tensor(_A , dtype=torch.long , device=_A ) _SCREAMING_SNAKE_CASE : Optional[int] = 1e-4 @require_torch @require_sentencepiece @require_tokenizers @slow class UpperCAmelCase__ ( unittest.TestCase ): """simple docstring""" @cached_property def lowercase_ ( self : List[Any] ) -> Dict: return MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''' ) def lowercase_ ( self : Optional[int] ) -> Tuple: SCREAMING_SNAKE_CASE__ = MaMaaaModel.from_pretrained('''facebook/m2m100_418M''' ).to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = _long_tensor([[12_8028, 98, 12, 3_0527, 2732, 159, 7755, 6_1904, 3_9144, 38, 2]] ) SCREAMING_SNAKE_CASE__ = _long_tensor([[2, 12_8028, 98, 12, 3_0527, 2732, 159, 7755, 6_1904, 3_9144, 38]] ) SCREAMING_SNAKE_CASE__ = prepare_mam_aaa_inputs_dict(model.config , __lowerCamelCase , __lowerCamelCase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase )[0] SCREAMING_SNAKE_CASE__ = torch.Size((1, 11, 1024) ) self.assertEqual(output.shape , __lowerCamelCase ) # change to expected output here SCREAMING_SNAKE_CASE__ = torch.tensor( [[-0.7780, -0.1676, 0.1038], [-6.7556, -1.3992, 0.0567], [-7.5383, -0.5920, -0.2779]] , device=__lowerCamelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) def lowercase_ ( self : str ) -> Dict: SCREAMING_SNAKE_CASE__ = MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''' ).to(__lowerCamelCase ) # change to intended input SCREAMING_SNAKE_CASE__ = _long_tensor([[12_8028, 98, 12, 3_0527, 2732, 159, 7755, 6_1904, 3_9144, 38, 2]] ) SCREAMING_SNAKE_CASE__ = _long_tensor([[2, 12_8028, 98, 12, 3_0527, 2732, 159, 7755, 6_1904, 3_9144, 38]] ) SCREAMING_SNAKE_CASE__ = prepare_mam_aaa_inputs_dict(model.config , __lowerCamelCase , __lowerCamelCase ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(**__lowerCamelCase )[0] SCREAMING_SNAKE_CASE__ = torch.Size((1, 11, model.config.vocab_size) ) self.assertEqual(output.shape , __lowerCamelCase ) # change to expected output here SCREAMING_SNAKE_CASE__ = torch.tensor( [[-1.0448, -1.0411, 3.7992], [-3.2191, -3.2386, -1.3451], [-3.6210, -3.5993, 0.4925]] , device=__lowerCamelCase ) self.assertTrue(torch.allclose(output[:, :3, :3] , __lowerCamelCase , atol=__lowerCamelCase ) ) def lowercase_ ( self : Any ) -> str: SCREAMING_SNAKE_CASE__ = MaMaaaForConditionalGeneration.from_pretrained('''facebook/m2m100_418M''' ).to(__lowerCamelCase ) SCREAMING_SNAKE_CASE__ = MaMaaaTokenizer.from_pretrained('''facebook/m2m100_418M''' , src_lang='''fr''' , tgt_lang='''en''' ) SCREAMING_SNAKE_CASE__ = [ '''L\'affaire NSA souligne l\'absence totale de débat sur le renseignement''', '''Selon moi, il y a deux niveaux de réponse de la part du gouvernement français.''', '''Lorsque François Hollande téléphone à Barack Obama ou quand le ministre des affaires étrangères Laurent''' ''' Fabius convoque l\'ambassadeur des Etats-Unis, ils réagissent à une vraie découverte, qui est celle de''' ''' l\'ampleur de la surveillance américaine sur l\'ensemble des communications en France.''', ] # The below article tests that we don't add any hypotheses outside of the top n_beams SCREAMING_SNAKE_CASE__ = tokenizer(__lowerCamelCase , padding=__lowerCamelCase , return_tensors='''pt''' ) SCREAMING_SNAKE_CASE__ = model.generate( input_ids=dct['''input_ids'''].to(__lowerCamelCase ) , attention_mask=dct['''attention_mask'''].to(__lowerCamelCase ) , num_beams=5 , forced_bos_token_id=tokenizer.get_lang_id('''en''' ) , ) SCREAMING_SNAKE_CASE__ = [ '''The NSA case highlights the total absence of intelligence debate''', '''I think there are two levels of response from the French government.''', '''When François Hollande calls Barack Obama or when Foreign Minister Laurent Fabius calls the U.S.''' ''' Ambassador, they respond to a real discovery, which is that of the scale of U.S. surveillance on all''' ''' communications in France.''', ] SCREAMING_SNAKE_CASE__ = tokenizer.batch_decode( hypotheses_batch.tolist() , clean_up_tokenization_spaces=__lowerCamelCase , skip_special_tokens=__lowerCamelCase ) assert generated == expected_en
314
from functools import reduce _SCREAMING_SNAKE_CASE : Any = ( '''73167176531330624919225119674426574742355349194934''' '''96983520312774506326239578318016984801869478851843''' '''85861560789112949495459501737958331952853208805511''' '''12540698747158523863050715693290963295227443043557''' '''66896648950445244523161731856403098711121722383113''' '''62229893423380308135336276614282806444486645238749''' '''30358907296290491560440772390713810515859307960866''' '''70172427121883998797908792274921901699720888093776''' '''65727333001053367881220235421809751254540594752243''' '''52584907711670556013604839586446706324415722155397''' '''53697817977846174064955149290862569321978468622482''' '''83972241375657056057490261407972968652414535100474''' '''82166370484403199890008895243450658541227588666881''' '''16427171479924442928230863465674813919123162824586''' '''17866458359124566529476545682848912883142607690042''' '''24219022671055626321111109370544217506941658960408''' '''07198403850962455444362981230987879927244284909188''' '''84580156166097919133875499200524063689912560717606''' '''05886116467109405077541002256983155200055935729725''' '''71636269561882670428252483600823257530420752963450''' ) def UpperCAmelCase_ ( _A = N ): '''simple docstring''' return max( # mypy cannot properly interpret reduce int(reduce(lambda _A , _A : str(int(_A ) * int(_A ) ) , n[i : i + 13] ) ) for i in range(len(_A ) - 12 ) ) if __name__ == "__main__": print(F"{solution() = }")
314
1
from __future__ import annotations import requests _UpperCAmelCase : List[Any] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase = 1 , _UpperCAmelCase = "new" , _UpperCAmelCase = None ) -> dict: lowerCamelCase__ : Dict = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(_UpperCAmelCase ) - valid_terms ) ): lowerCamelCase__ : Union[str, Any] = F"""Invalid search term: {invalid_search_terms}""" raise ValueError(_UpperCAmelCase ) lowerCamelCase__ : Tuple = requests.get( F"""https://reddit.com/r/{subreddit}/{age}.json?limit={limit}""" , headers={'User-agent': 'A random string'} , ) if response.status_code == 429: raise requests.HTTPError lowerCamelCase__ : int = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(_UpperCAmelCase )} lowerCamelCase__ : Union[str, Any] = {} for id_ in range(_UpperCAmelCase ): lowerCamelCase__ : str = { item: data['data']['children'][id_]['data'][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
355
from bisect import bisect from itertools import accumulate def SCREAMING_SNAKE_CASE ( _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase , _UpperCAmelCase ) -> Tuple: lowerCamelCase__ : Optional[int] = sorted(zip(_UpperCAmelCase , _UpperCAmelCase ) , key=lambda _UpperCAmelCase : x[0] / x[1] , reverse=_UpperCAmelCase ) lowerCamelCase__ , lowerCamelCase__ : Optional[Any] = [i[0] for i in r], [i[1] for i in r] lowerCamelCase__ : Tuple = list(accumulate(_UpperCAmelCase ) ) lowerCamelCase__ : int = bisect(_UpperCAmelCase , _UpperCAmelCase ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
45
0
"""simple docstring""" from abc import ABC, abstractmethod from argparse import ArgumentParser class _UpperCAmelCase ( lowercase_ ): @staticmethod @abstractmethod def lowerCamelCase ( __UpperCamelCase :ArgumentParser ): raise NotImplementedError() @abstractmethod def lowerCamelCase ( self :List[str] ): raise NotImplementedError()
292
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) _snake_case : Union[str, Any] = { 'configuration_encodec': [ 'ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP', 'EncodecConfig', ], 'feature_extraction_encodec': ['EncodecFeatureExtractor'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _snake_case : int = [ 'ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST', 'EncodecModel', 'EncodecPreTrainedModel', ] if TYPE_CHECKING: from .configuration_encodec import ( ENCODEC_PRETRAINED_CONFIG_ARCHIVE_MAP, EncodecConfig, ) from .feature_extraction_encodec import EncodecFeatureExtractor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_encodec import ( ENCODEC_PRETRAINED_MODEL_ARCHIVE_LIST, EncodecModel, EncodecPreTrainedModel, ) else: import sys _snake_case : str = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
292
1
import subprocess import sys from transformers import BertConfig, BertModel, BertTokenizer, pipeline from transformers.testing_utils import TestCasePlus, require_torch class lowerCamelCase__ ( _a ): @require_torch def _lowerCamelCase ( self : Union[str, Any] ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched a__: Dict ="\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " a__: Union[str, Any] ="\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " a__: Dict ="\nimport socket\ndef offline_socket(*args, **kwargs): raise RuntimeError(\"Offline mode is enabled, we shouldn't access internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache a__: Tuple ="hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(_a ) BertModel.from_pretrained(_a ) BertTokenizer.from_pretrained(_a ) pipeline(task="fill-mask" , model=_a ) # baseline - just load from_pretrained with normal network a__: Optional[int] =[sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed a__: str =self.get_env() # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__: Union[str, Any] ="1" a__: Any =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def _lowerCamelCase ( self : str ): # python one-liner segments # this must be loaded before socket.socket is monkey-patched a__: Optional[int] ="\nfrom transformers import BertConfig, BertModel, BertTokenizer, pipeline\n " a__: Tuple ="\nmname = \"hf-internal-testing/tiny-random-bert\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nBertTokenizer.from_pretrained(mname)\npipe = pipeline(task=\"fill-mask\", model=mname)\nprint(\"success\")\n " a__: List[Any] ="\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Faking flaky internet\")\nsocket.socket = offline_socket\n " # Force fetching the files so that we can use the cache a__: str ="hf-internal-testing/tiny-random-bert" BertConfig.from_pretrained(_a ) BertModel.from_pretrained(_a ) BertTokenizer.from_pretrained(_a ) pipeline(task="fill-mask" , model=_a ) # baseline - just load from_pretrained with normal network a__: Any =[sys.executable, "-c", "\n".join([load, run, mock] )] # should succeed a__: Optional[Any] =self.get_env() a__: Union[str, Any] =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def _lowerCamelCase ( self : List[str] ): # this test is a bit tricky since TRANSFORMERS_OFFLINE can only be changed before # `transformers` is loaded, and it's too late for inside pytest - so we are changing it # while running an external program # python one-liner segments # this must be loaded before socket.socket is monkey-patched a__: Tuple ="\nfrom transformers import BertConfig, BertModel, BertTokenizer\n " a__: str ="\nmname = \"hf-internal-testing/tiny-random-bert-sharded\"\nBertConfig.from_pretrained(mname)\nBertModel.from_pretrained(mname)\nprint(\"success\")\n " a__: int ="\nimport socket\ndef offline_socket(*args, **kwargs): raise ValueError(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " # baseline - just load from_pretrained with normal network a__: Union[str, Any] =[sys.executable, "-c", "\n".join([load, run] )] # should succeed a__: Optional[Any] =self.get_env() a__: Optional[Any] =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # next emulate no network a__: int =[sys.executable, "-c", "\n".join([load, mock, run] )] # Doesn't fail anymore since the model is in the cache due to other tests, so commenting this. # env["TRANSFORMERS_OFFLINE"] = "0" # result = subprocess.run(cmd, env=env, check=False, capture_output=True) # self.assertEqual(result.returncode, 1, result.stderr) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__: Tuple ="1" a__: Dict =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) @require_torch def _lowerCamelCase ( self : Optional[Any] ): a__: Dict ="\nfrom transformers import pipeline\n " a__: Union[str, Any] ="\nmname = \"hf-internal-testing/tiny-random-bert\"\npipe = pipeline(model=mname)\n " a__: List[str] ="\nimport socket\ndef offline_socket(*args, **kwargs): raise socket.error(\"Offline mode is enabled\")\nsocket.socket = offline_socket\n " a__: Dict =self.get_env() a__: Optional[Any] ="1" a__: Dict =[sys.executable, "-c", "\n".join([load, mock, run] )] a__: Optional[int] =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 1 , result.stderr ) self.assertIn( "You cannot infer task automatically within `pipeline` when using offline mode" , result.stderr.decode().replace("\n" , "" ) , ) @require_torch def _lowerCamelCase ( self : Optional[int] ): a__: Optional[Any] ="\nfrom transformers import AutoModel\n " a__: str ="\nmname = \"hf-internal-testing/test_dynamic_model\"\nAutoModel.from_pretrained(mname, trust_remote_code=True)\nprint(\"success\")\n " # baseline - just load from_pretrained with normal network a__: Tuple =[sys.executable, "-c", "\n".join([load, run] )] # should succeed a__: Any =self.get_env() a__: int =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() ) # should succeed as TRANSFORMERS_OFFLINE=1 tells it to use local files a__: List[Any] ="1" a__: int =subprocess.run(_a , env=_a , check=_a , capture_output=_a ) self.assertEqual(result.returncode , 0 , result.stderr ) self.assertIn("success" , result.stdout.decode() )
42
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging __UpperCAmelCase = logging.get_logger(__name__) __UpperCAmelCase = { '''google/mobilenet_v1_1.0_224''': '''https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json''', '''google/mobilenet_v1_0.75_192''': '''https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json''', # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class lowerCamelCase__ ( _a ): _lowerCAmelCase = '''mobilenet_v1''' def __init__( self : int , _a : Tuple=3 , _a : str=2_2_4 , _a : Dict=1.0 , _a : List[Any]=8 , _a : Tuple="relu6" , _a : Dict=True , _a : Optional[int]=0.9_9_9 , _a : List[Any]=0.0_2 , _a : Optional[Any]=0.0_0_1 , **_a : Optional[int] , ): super().__init__(**_a ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) a__: str =num_channels a__: Union[str, Any] =image_size a__: Dict =depth_multiplier a__: Union[str, Any] =min_depth a__: Any =hidden_act a__: int =tf_padding a__: Dict =classifier_dropout_prob a__: Any =initializer_range a__: List[str] =layer_norm_eps class lowerCamelCase__ ( _a ): _lowerCAmelCase = version.parse('''1.11''' ) @property def _lowerCamelCase ( self : int ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def _lowerCamelCase ( self : Tuple ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def _lowerCamelCase ( self : Dict ): return 1e-4
42
1
"""simple docstring""" from dataclasses import dataclass from typing import Dict, Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, apply_forward_hook from .attention_processor import AttentionProcessor, AttnProcessor from .modeling_utils import ModelMixin from .vae import Decoder, DecoderOutput, DiagonalGaussianDistribution, Encoder @dataclass class UpperCamelCase ( lowercase ): UpperCAmelCase : "DiagonalGaussianDistribution" class UpperCamelCase ( lowercase , lowercase ): UpperCAmelCase : Dict = True @register_to_config def __init__(self : Tuple , _A : int = 3 , _A : int = 3 , _A : Tuple[str] = ("DownEncoderBlock2D",) , _A : Tuple[str] = ("UpDecoderBlock2D",) , _A : Tuple[int] = (64,) , _A : int = 1 , _A : str = "silu" , _A : int = 4 , _A : int = 32 , _A : int = 32 , _A : float = 0.18_215 , ) -> List[str]: super().__init__() # pass init params to Encoder __snake_case : List[str] = Encoder( in_channels=_A , out_channels=_A , down_block_types=_A , block_out_channels=_A , layers_per_block=_A , act_fn=_A , norm_num_groups=_A , double_z=_A , ) # pass init params to Decoder __snake_case : Optional[int] = Decoder( in_channels=_A , out_channels=_A , up_block_types=_A , block_out_channels=_A , layers_per_block=_A , norm_num_groups=_A , act_fn=_A , ) __snake_case : Any = nn.Convad(2 * latent_channels , 2 * latent_channels , 1) __snake_case : Tuple = nn.Convad(_A , _A , 1) __snake_case : int = False __snake_case : int = False # only relevant if vae tiling is enabled __snake_case : str = self.config.sample_size __snake_case : Any = ( self.config.sample_size[0] if isinstance(self.config.sample_size , (list, tuple)) else self.config.sample_size ) __snake_case : str = int(sample_size / (2 ** (len(self.config.block_out_channels) - 1))) __snake_case : str = 0.25 def _lowercase (self : str , _A : int , _A : Union[str, Any]=False) -> Optional[int]: if isinstance(_A , (Encoder, Decoder)): __snake_case : str = value def _lowercase (self : Tuple , _A : bool = True) -> Any: __snake_case : List[Any] = use_tiling def _lowercase (self : str) -> List[Any]: self.enable_tiling(_A) def _lowercase (self : Optional[Any]) -> Optional[Any]: __snake_case : Any = True def _lowercase (self : str) -> Optional[int]: __snake_case : Dict = False @property # Copied from diffusers.models.unet_2d_condition.UNet2DConditionModel.attn_processors def _lowercase (self : Any) -> Dict[str, AttentionProcessor]: __snake_case : str = {} def fn_recursive_add_processors(_A : str , _A : torch.nn.Module , _A : Dict[str, AttentionProcessor]): if hasattr(_A , 'set_processor'): __snake_case : Tuple = module.processor for sub_name, child in module.named_children(): fn_recursive_add_processors(f"{name}.{sub_name}" , _A , _A) return processors for name, module in self.named_children(): fn_recursive_add_processors(_A , _A , _A) return processors def _lowercase (self : int , _A : Union[AttentionProcessor, Dict[str, AttentionProcessor]]) -> str: __snake_case : Optional[int] = len(self.attn_processors.keys()) if isinstance(_A , _A) and len(_A) != count: raise ValueError( f"A dict of processors was passed, but the number of processors {len(_A)} does not match the" f" number of attention layers: {count}. Please make sure to pass {count} processor classes.") def fn_recursive_attn_processor(_A : str , _A : torch.nn.Module , _A : List[str]): if hasattr(_A , 'set_processor'): if not isinstance(_A , _A): module.set_processor(_A) else: module.set_processor(processor.pop(f"{name}.processor")) for sub_name, child in module.named_children(): fn_recursive_attn_processor(f"{name}.{sub_name}" , _A , _A) for name, module in self.named_children(): fn_recursive_attn_processor(_A , _A , _A) def _lowercase (self : Union[str, Any]) -> int: self.set_attn_processor(AttnProcessor()) @apply_forward_hook def _lowercase (self : Optional[int] , _A : torch.FloatTensor , _A : bool = True) -> AutoencoderKLOutput: if self.use_tiling and (x.shape[-1] > self.tile_sample_min_size or x.shape[-2] > self.tile_sample_min_size): return self.tiled_encode(_A , return_dict=_A) if self.use_slicing and x.shape[0] > 1: __snake_case : Union[str, Any] = [self.encoder(_A) for x_slice in x.split(1)] __snake_case : Any = torch.cat(_A) else: __snake_case : Tuple = self.encoder(_A) __snake_case : Dict = self.quant_conv(_A) __snake_case : Optional[Any] = DiagonalGaussianDistribution(_A) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_A) def _lowercase (self : Dict , _A : torch.FloatTensor , _A : bool = True) -> Union[DecoderOutput, torch.FloatTensor]: if self.use_tiling and (z.shape[-1] > self.tile_latent_min_size or z.shape[-2] > self.tile_latent_min_size): return self.tiled_decode(_A , return_dict=_A) __snake_case : Dict = self.post_quant_conv(_A) __snake_case : Union[str, Any] = self.decoder(_A) if not return_dict: return (dec,) return DecoderOutput(sample=_A) @apply_forward_hook def _lowercase (self : Tuple , _A : torch.FloatTensor , _A : bool = True) -> Union[DecoderOutput, torch.FloatTensor]: if self.use_slicing and z.shape[0] > 1: __snake_case : Optional[int] = [self._decode(_A).sample for z_slice in z.split(1)] __snake_case : Optional[int] = torch.cat(_A) else: __snake_case : Any = self._decode(_A).sample if not return_dict: return (decoded,) return DecoderOutput(sample=_A) def _lowercase (self : Dict , _A : List[Any] , _A : List[Any] , _A : Optional[int]) -> List[str]: __snake_case : List[str] = min(a.shape[2] , b.shape[2] , _A) for y in range(_A): __snake_case : Union[str, Any] = a[:, :, -blend_extent + y, :] * (1 - y / blend_extent) + b[:, :, y, :] * (y / blend_extent) return b def _lowercase (self : List[Any] , _A : Any , _A : Tuple , _A : Optional[int]) -> Union[str, Any]: __snake_case : List[str] = min(a.shape[3] , b.shape[3] , _A) for x in range(_A): __snake_case : str = a[:, :, :, -blend_extent + x] * (1 - x / blend_extent) + b[:, :, :, x] * (x / blend_extent) return b def _lowercase (self : List[Any] , _A : torch.FloatTensor , _A : bool = True) -> AutoencoderKLOutput: __snake_case : int = int(self.tile_sample_min_size * (1 - self.tile_overlap_factor)) __snake_case : Any = int(self.tile_latent_min_size * self.tile_overlap_factor) __snake_case : Union[str, Any] = self.tile_latent_min_size - blend_extent # Split the image into 512x512 tiles and encode them separately. __snake_case : Optional[Any] = [] for i in range(0 , x.shape[2] , _A): __snake_case : Optional[int] = [] for j in range(0 , x.shape[3] , _A): __snake_case : Optional[Any] = x[:, :, i : i + self.tile_sample_min_size, j : j + self.tile_sample_min_size] __snake_case : List[Any] = self.encoder(_A) __snake_case : int = self.quant_conv(_A) row.append(_A) rows.append(_A) __snake_case : Dict = [] for i, row in enumerate(_A): __snake_case : Optional[Any] = [] for j, tile in enumerate(_A): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __snake_case : Optional[int] = self.blend_v(rows[i - 1][j] , _A , _A) if j > 0: __snake_case : List[str] = self.blend_h(row[j - 1] , _A , _A) result_row.append(tile[:, :, :row_limit, :row_limit]) result_rows.append(torch.cat(_A , dim=3)) __snake_case : str = torch.cat(_A , dim=2) __snake_case : List[Any] = DiagonalGaussianDistribution(_A) if not return_dict: return (posterior,) return AutoencoderKLOutput(latent_dist=_A) def _lowercase (self : Union[str, Any] , _A : torch.FloatTensor , _A : bool = True) -> Union[DecoderOutput, torch.FloatTensor]: __snake_case : Optional[int] = int(self.tile_latent_min_size * (1 - self.tile_overlap_factor)) __snake_case : Tuple = int(self.tile_sample_min_size * self.tile_overlap_factor) __snake_case : Optional[Any] = self.tile_sample_min_size - blend_extent # Split z into overlapping 64x64 tiles and decode them separately. # The tiles have an overlap to avoid seams between tiles. __snake_case : int = [] for i in range(0 , z.shape[2] , _A): __snake_case : Dict = [] for j in range(0 , z.shape[3] , _A): __snake_case : int = z[:, :, i : i + self.tile_latent_min_size, j : j + self.tile_latent_min_size] __snake_case : Optional[int] = self.post_quant_conv(_A) __snake_case : List[str] = self.decoder(_A) row.append(_A) rows.append(_A) __snake_case : Optional[int] = [] for i, row in enumerate(_A): __snake_case : List[Any] = [] for j, tile in enumerate(_A): # blend the above tile and the left tile # to the current tile and add the current tile to the result row if i > 0: __snake_case : List[str] = self.blend_v(rows[i - 1][j] , _A , _A) if j > 0: __snake_case : int = self.blend_h(row[j - 1] , _A , _A) result_row.append(tile[:, :, :row_limit, :row_limit]) result_rows.append(torch.cat(_A , dim=3)) __snake_case : Union[str, Any] = torch.cat(_A , dim=2) if not return_dict: return (dec,) return DecoderOutput(sample=_A) def _lowercase (self : List[Any] , _A : torch.FloatTensor , _A : bool = False , _A : bool = True , _A : Optional[torch.Generator] = None , ) -> Union[DecoderOutput, torch.FloatTensor]: __snake_case : Optional[Any] = sample __snake_case : List[str] = self.encode(_A).latent_dist if sample_posterior: __snake_case : List[Any] = posterior.sample(generator=_A) else: __snake_case : int = posterior.mode() __snake_case : Union[str, Any] = self.decode(_A).sample if not return_dict: return (dec,) return DecoderOutput(sample=_A)
172
"""simple docstring""" _a : Tuple= 8.3_1_4_4_5_9_8 def __UpperCAmelCase ( UpperCAmelCase_ : float , UpperCAmelCase_ : float ) -> float: '''simple docstring''' if temperature < 0: raise Exception('Temperature cannot be less than 0 K' ) if molar_mass <= 0: raise Exception('Molar mass cannot be less than or equal to 0 kg/mol' ) else: return (3 * UNIVERSAL_GAS_CONSTANT * temperature / molar_mass) ** 0.5 if __name__ == "__main__": import doctest # run doctest doctest.testmod() # example _a : Any= 300 _a : Optional[Any]= 28 _a : Optional[int]= rms_speed_of_molecule(temperature, molar_mass) print(f'''Vrms of Nitrogen gas at 300 K is {vrms} m/s''')
172
1
'''simple docstring''' def _UpperCamelCase ( UpperCamelCase__ ): if not all(char in """01""" for char in bin_string ): raise ValueError("""Non-binary value was passed to the function""" ) if not bin_string: raise ValueError("""Empty string was passed to the function""" ) UpperCAmelCase__ : str = """""" while len(UpperCamelCase__ ) % 3 != 0: UpperCAmelCase__ : Optional[Any] = """0""" + bin_string UpperCAmelCase__ : Optional[int] = [ bin_string[index : index + 3] for index in range(len(UpperCamelCase__ ) ) if index % 3 == 0 ] for bin_group in bin_string_in_3_list: UpperCAmelCase__ : Optional[Any] = 0 for index, val in enumerate(UpperCamelCase__ ): oct_val += int(2 ** (2 - index) * int(UpperCamelCase__ ) ) oct_string += str(UpperCamelCase__ ) return oct_string if __name__ == "__main__": from doctest import testmod testmod()
358
'''simple docstring''' import numpy class _snake_case : def __init__( self , _lowerCamelCase , _lowerCamelCase): UpperCAmelCase__ : Dict = input_array # Random initial weights are assigned where first argument is the # number of nodes in previous layer and second argument is the # number of nodes in the next layer. # Random initial weights are assigned. # self.input_array.shape[1] is used to represent number of nodes in input layer. # First hidden layer consists of 4 nodes. UpperCAmelCase__ : Optional[Any] = numpy.random.rand( self.input_array.shape[1] , 4) # Random initial values for the first hidden layer. # First hidden layer has 4 nodes. # Second hidden layer has 3 nodes. UpperCAmelCase__ : Optional[int] = numpy.random.rand( 4 , 3) # Random initial values for the second hidden layer. # Second hidden layer has 3 nodes. # Output layer has 1 node. UpperCAmelCase__ : Any = numpy.random.rand(3 , 1) # Real output values provided. UpperCAmelCase__ : Tuple = output_array # Predicted output values by the neural network. # Predicted_output array initially consists of zeroes. UpperCAmelCase__ : Union[str, Any] = numpy.zeros(output_array.shape) def snake_case__ ( self): UpperCAmelCase__ : List[str] = sigmoid( numpy.dot(self.input_array , self.input_layer_and_first_hidden_layer_weights)) # layer_between_first_hidden_layer_and_second_hidden_layer is the layer # connecting the first hidden set of nodes with the second hidden set of nodes. UpperCAmelCase__ : Any = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , )) # layer_between_second_hidden_layer_and_output is the layer connecting # second hidden layer with the output node. UpperCAmelCase__ : Tuple = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , )) return self.layer_between_second_hidden_layer_and_output def snake_case__ ( self): UpperCAmelCase__ : Optional[int] = numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer.T , 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output) , ) UpperCAmelCase__ : str = numpy.dot( self.layer_between_input_and_first_hidden_layer.T , numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer) , ) UpperCAmelCase__ : Any = numpy.dot( self.input_array.T , numpy.dot( numpy.dot( 2 * (self.output_array - self.predicted_output) * sigmoid_derivative(self.predicted_output) , self.second_hidden_layer_and_output_layer_weights.T , ) * sigmoid_derivative( self.layer_between_first_hidden_layer_and_second_hidden_layer) , self.first_hidden_layer_and_second_hidden_layer_weights.T , ) * sigmoid_derivative(self.layer_between_input_and_first_hidden_layer) , ) self.input_layer_and_first_hidden_layer_weights += ( updated_input_layer_and_first_hidden_layer_weights ) self.first_hidden_layer_and_second_hidden_layer_weights += ( updated_first_hidden_layer_and_second_hidden_layer_weights ) self.second_hidden_layer_and_output_layer_weights += ( updated_second_hidden_layer_and_output_layer_weights ) def snake_case__ ( self , _lowerCamelCase , _lowerCamelCase , _lowerCamelCase): for iteration in range(1 , iterations + 1): UpperCAmelCase__ : Optional[Any] = self.feedforward() self.back_propagation() if give_loss: UpperCAmelCase__ : str = numpy.mean(numpy.square(output - self.feedforward())) print(f'''Iteration {iteration} Loss: {loss}''') def snake_case__ ( self , _lowerCamelCase): UpperCAmelCase__ : List[Any] = input_arr UpperCAmelCase__ : Tuple = sigmoid( numpy.dot(self.array , self.input_layer_and_first_hidden_layer_weights)) UpperCAmelCase__ : List[Any] = sigmoid( numpy.dot( self.layer_between_input_and_first_hidden_layer , self.first_hidden_layer_and_second_hidden_layer_weights , )) UpperCAmelCase__ : Optional[Any] = sigmoid( numpy.dot( self.layer_between_first_hidden_layer_and_second_hidden_layer , self.second_hidden_layer_and_output_layer_weights , )) return int(self.layer_between_second_hidden_layer_and_output > 0.6) def _UpperCamelCase ( UpperCamelCase__ ): return 1 / (1 + numpy.exp(-value )) def _UpperCamelCase ( UpperCamelCase__ ): return (value) * (1 - (value)) def _UpperCamelCase ( ): UpperCAmelCase__ : Union[str, Any] = numpy.array( ( [0, 0, 0], [0, 0, 1], [0, 1, 0], [0, 1, 1], [1, 0, 0], [1, 0, 1], [1, 1, 0], [1, 1, 1], ) , dtype=numpy.floataa , ) # True output values for the given input values. UpperCAmelCase__ : str = numpy.array(([0], [1], [1], [0], [1], [0], [0], [1]) , dtype=numpy.floataa ) # Calling neural network class. UpperCAmelCase__ : List[Any] = TwoHiddenLayerNeuralNetwork( input_array=UpperCamelCase__ , output_array=UpperCamelCase__ ) # Calling training function. # Set give_loss to True if you want to see loss in every iteration. neural_network.train(output=UpperCamelCase__ , iterations=1_0 , give_loss=UpperCamelCase__ ) return neural_network.predict(numpy.array(([1, 1, 1]) , dtype=numpy.floataa ) ) if __name__ == "__main__": example()
283
0
from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( IMAGENET_STANDARD_MEAN, IMAGENET_STANDARD_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging if is_vision_available(): import PIL A : str = logging.get_logger(__name__) class _lowercase ( __SCREAMING_SNAKE_CASE): """simple docstring""" A__ = ["pixel_values"] def __init__( self : Union[str, Any] , __lowerCamelCase : Optional[int] = True , __lowerCamelCase : Optional[Any] = None , __lowerCamelCase : str = None , __lowerCamelCase : List[str] = PILImageResampling.BILINEAR , __lowerCamelCase : Dict = True , __lowerCamelCase : str = 1 / 255 , __lowerCamelCase : Dict = True , __lowerCamelCase : Tuple = None , __lowerCamelCase : Any = None , **__lowerCamelCase : List[Any] , ): '''simple docstring''' super().__init__(**__UpperCAmelCase ) lowerCamelCase__ : List[Any] = size if size is not None else {"shortest_edge": 384} lowerCamelCase__ : Dict = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) lowerCamelCase__ : List[str] = do_resize lowerCamelCase__ : Union[str, Any] = size # Default value set here for backwards compatibility where the value in config is None lowerCamelCase__ : str = crop_pct if crop_pct is not None else 224 / 256 lowerCamelCase__ : Optional[Any] = resample lowerCamelCase__ : List[str] = do_rescale lowerCamelCase__ : Union[str, Any] = rescale_factor lowerCamelCase__ : str = do_normalize lowerCamelCase__ : int = image_mean if image_mean is not None else IMAGENET_STANDARD_MEAN lowerCamelCase__ : Optional[Any] = image_std if image_std is not None else IMAGENET_STANDARD_STD def lowerCAmelCase ( self : Dict , __lowerCamelCase : List[str] , __lowerCamelCase : int , __lowerCamelCase : List[Any] , __lowerCamelCase : Optional[Any] = PILImageResampling.BICUBIC , __lowerCamelCase : int = None , **__lowerCamelCase : Union[str, Any] , ): '''simple docstring''' lowerCamelCase__ : List[Any] = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) if "shortest_edge" not in size: raise ValueError(f"Size dictionary must contain \'shortest_edge\' key. Got {size.keys()}" ) lowerCamelCase__ : Any = size["shortest_edge"] if shortest_edge < 384: # maintain same ratio, resizing shortest edge to shortest_edge/crop_pct lowerCamelCase__ : Tuple = int(shortest_edge / crop_pct ) lowerCamelCase__ : Optional[int] = get_resize_output_image_size(__UpperCAmelCase , size=__UpperCAmelCase , default_to_square=__UpperCAmelCase ) lowerCamelCase__ : str = resize(image=__UpperCAmelCase , size=__UpperCAmelCase , resample=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) # then crop to (shortest_edge, shortest_edge) return center_crop(image=__UpperCAmelCase , size=(shortest_edge, shortest_edge) , data_format=__UpperCAmelCase , **__UpperCAmelCase ) else: # warping (no cropping) when evaluated at 384 or larger return resize( __UpperCAmelCase , size=(shortest_edge, shortest_edge) , resample=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def lowerCAmelCase ( self : Tuple , __lowerCamelCase : int , __lowerCamelCase : Optional[int] , __lowerCamelCase : Optional[int] = None , **__lowerCamelCase : Tuple , ): '''simple docstring''' return rescale(__UpperCAmelCase , scale=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def lowerCAmelCase ( self : Dict , __lowerCamelCase : Optional[int] , __lowerCamelCase : int , __lowerCamelCase : int , __lowerCamelCase : List[Any] = None , **__lowerCamelCase : Optional[int] , ): '''simple docstring''' return normalize(__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase , data_format=__UpperCAmelCase , **__UpperCAmelCase ) def lowerCAmelCase ( self : int , __lowerCamelCase : Tuple , __lowerCamelCase : List[Any] = None , __lowerCamelCase : int = None , __lowerCamelCase : List[Any] = None , __lowerCamelCase : str = None , __lowerCamelCase : List[Any] = None , __lowerCamelCase : Union[str, Any] = None , __lowerCamelCase : Dict = None , __lowerCamelCase : Any = None , __lowerCamelCase : Union[str, Any] = None , __lowerCamelCase : Optional[int] = None , __lowerCamelCase : int = ChannelDimension.FIRST , **__lowerCamelCase : str , ): '''simple docstring''' lowerCamelCase__ : List[str] = do_resize if do_resize is not None else self.do_resize lowerCamelCase__ : Any = crop_pct if crop_pct is not None else self.crop_pct lowerCamelCase__ : int = resample if resample is not None else self.resample lowerCamelCase__ : Tuple = do_rescale if do_rescale is not None else self.do_rescale lowerCamelCase__ : Optional[int] = rescale_factor if rescale_factor is not None else self.rescale_factor lowerCamelCase__ : Optional[int] = do_normalize if do_normalize is not None else self.do_normalize lowerCamelCase__ : List[Any] = image_mean if image_mean is not None else self.image_mean lowerCamelCase__ : Any = image_std if image_std is not None else self.image_std lowerCamelCase__ : str = size if size is not None else self.size lowerCamelCase__ : List[str] = get_size_dict(__UpperCAmelCase , default_to_square=__UpperCAmelCase ) lowerCamelCase__ : Tuple = make_list_of_images(__UpperCAmelCase ) if not valid_images(__UpperCAmelCase ): raise ValueError( "Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, " "torch.Tensor, tf.Tensor or jax.ndarray." ) if do_resize and size is None or resample is None: raise ValueError("Size and resample must be specified if do_resize is True." ) if do_resize and size["shortest_edge"] < 384 and crop_pct is None: raise ValueError("crop_pct must be specified if size < 384." ) if do_rescale and rescale_factor is None: raise ValueError("Rescale factor must be specified if do_rescale is True." ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("Image mean and std must be specified if do_normalize is True." ) # All transformations expect numpy arrays. lowerCamelCase__ : Optional[int] = [to_numpy_array(__UpperCAmelCase ) for image in images] if do_resize: lowerCamelCase__ : int = [self.resize(image=__UpperCAmelCase , size=__UpperCAmelCase , crop_pct=__UpperCAmelCase , resample=__UpperCAmelCase ) for image in images] if do_rescale: lowerCamelCase__ : List[Any] = [self.rescale(image=__UpperCAmelCase , scale=__UpperCAmelCase ) for image in images] if do_normalize: lowerCamelCase__ : int = [self.normalize(image=__UpperCAmelCase , mean=__UpperCAmelCase , std=__UpperCAmelCase ) for image in images] lowerCamelCase__ : List[str] = [to_channel_dimension_format(__UpperCAmelCase , __UpperCAmelCase ) for image in images] lowerCamelCase__ : Tuple = {"pixel_values": images} return BatchFeature(data=__UpperCAmelCase , tensor_type=__UpperCAmelCase )
184
"""simple docstring""" UpperCamelCase : dict[str, float] = { "km/h": 1.0, "m/s": 3.6, "mph": 1.60_93_44, "knot": 1.8_52, } UpperCamelCase : dict[str, float] = { "km/h": 1.0, "m/s": 0.2_77_77_77_78, "mph": 0.6_21_37_11_92, "knot": 0.5_39_95_68_03, } def A ( snake_case :float , snake_case :str , snake_case :str ) -> float: if unit_to not in speed_chart or unit_from not in speed_chart_inverse: __UpperCamelCase = ( f'Incorrect \'from_type\' or \'to_type\' value: {unit_from!r}, {unit_to!r}\n' f'Valid values are: {", ".join(snake_case )}' ) raise ValueError(snake_case ) return round(speed * speed_chart[unit_from] * speed_chart_inverse[unit_to] , 3 ) if __name__ == "__main__": import doctest doctest.testmod()
316
0
"""simple docstring""" import gc import unittest import numpy as np import torch from diffusers import ( AudioDiffusionPipeline, AutoencoderKL, DDIMScheduler, DDPMScheduler, DiffusionPipeline, Mel, UNetaDConditionModel, UNetaDModel, ) from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu enable_full_determinism() class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> List[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() @property def _UpperCamelCase ( self ) -> str: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :Union[str, Any] = UNetaDModel( sample_size=(32, 64) ,in_channels=1 ,out_channels=1 ,layers_per_block=2 ,block_out_channels=(1_28, 1_28) ,down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') ,up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') ,) return model @property def _UpperCamelCase ( self ) -> Union[str, Any]: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :Optional[Any] = UNetaDConditionModel( sample_size=(64, 32) ,in_channels=1 ,out_channels=1 ,layers_per_block=2 ,block_out_channels=(1_28, 1_28) ,down_block_types=('''CrossAttnDownBlock2D''', '''DownBlock2D''') ,up_block_types=('''UpBlock2D''', '''CrossAttnUpBlock2D''') ,cross_attention_dim=10 ,) return model @property def _UpperCamelCase ( self ) -> int: """simple docstring""" torch.manual_seed(0 ) __SCREAMING_SNAKE_CASE :List[str] = AutoencoderKL( sample_size=(1_28, 64) ,in_channels=1 ,out_channels=1 ,latent_channels=1 ,layers_per_block=2 ,block_out_channels=(1_28, 1_28) ,down_block_types=('''DownEncoderBlock2D''', '''DownEncoderBlock2D''') ,up_block_types=('''UpDecoderBlock2D''', '''UpDecoderBlock2D''') ,) __SCREAMING_SNAKE_CASE :Union[str, Any] = UNetaDModel( sample_size=(64, 32) ,in_channels=1 ,out_channels=1 ,layers_per_block=2 ,block_out_channels=(1_28, 1_28) ,down_block_types=('''AttnDownBlock2D''', '''DownBlock2D''') ,up_block_types=('''UpBlock2D''', '''AttnUpBlock2D''') ,) return vqvae, unet @slow def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :Optional[int] = '''cpu''' # ensure determinism for the device-dependent torch.Generator __SCREAMING_SNAKE_CASE :List[str] = Mel( x_res=self.dummy_unet.config.sample_size[1] ,y_res=self.dummy_unet.config.sample_size[0] ,) __SCREAMING_SNAKE_CASE :Tuple = DDPMScheduler() __SCREAMING_SNAKE_CASE :Tuple = AudioDiffusionPipeline(vqvae=SCREAMING_SNAKE_CASE__ ,unet=self.dummy_unet ,mel=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[Any] = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Any = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(42 ) __SCREAMING_SNAKE_CASE :Any = pipe(generator=SCREAMING_SNAKE_CASE__ ,steps=4 ) __SCREAMING_SNAKE_CASE :Tuple = output.audios[0] __SCREAMING_SNAKE_CASE :int = output.images[0] __SCREAMING_SNAKE_CASE :Dict = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(42 ) __SCREAMING_SNAKE_CASE :Union[str, Any] = pipe(generator=SCREAMING_SNAKE_CASE__ ,steps=4 ,return_dict=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[int] = output[0][0] assert audio.shape == (1, (self.dummy_unet.config.sample_size[1] - 1) * mel.hop_length) assert ( image.height == self.dummy_unet.config.sample_size[0] and image.width == self.dummy_unet.config.sample_size[1] ) __SCREAMING_SNAKE_CASE :Any = np.frombuffer(image.tobytes() ,dtype='''uint8''' )[:10] __SCREAMING_SNAKE_CASE :Any = np.frombuffer(image_from_tuple.tobytes() ,dtype='''uint8''' )[:10] __SCREAMING_SNAKE_CASE :Optional[Any] = np.array([69, 2_55, 2_55, 2_55, 0, 0, 77, 1_81, 12, 1_27] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() == 0 __SCREAMING_SNAKE_CASE :Dict = Mel( x_res=self.dummy_vqvae_and_unet[0].config.sample_size[1] ,y_res=self.dummy_vqvae_and_unet[0].config.sample_size[0] ,) __SCREAMING_SNAKE_CASE :Optional[int] = DDIMScheduler() __SCREAMING_SNAKE_CASE :Union[str, Any] = self.dummy_vqvae_and_unet __SCREAMING_SNAKE_CASE :List[Any] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] ,unet=dummy_vqvae_and_unet[1] ,mel=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Any = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) np.random.seed(0 ) __SCREAMING_SNAKE_CASE :Any = np.random.uniform(-1 ,1 ,((dummy_vqvae_and_unet[0].config.sample_size[1] - 1) * mel.hop_length,) ) __SCREAMING_SNAKE_CASE :Dict = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(42 ) __SCREAMING_SNAKE_CASE :List[str] = pipe(raw_audio=SCREAMING_SNAKE_CASE__ ,generator=SCREAMING_SNAKE_CASE__ ,start_step=5 ,steps=10 ) __SCREAMING_SNAKE_CASE :Dict = output.images[0] assert ( image.height == self.dummy_vqvae_and_unet[0].config.sample_size[0] and image.width == self.dummy_vqvae_and_unet[0].config.sample_size[1] ) __SCREAMING_SNAKE_CASE :int = np.frombuffer(image.tobytes() ,dtype='''uint8''' )[:10] __SCREAMING_SNAKE_CASE :List[Any] = np.array([1_20, 1_17, 1_10, 1_09, 1_38, 1_67, 1_38, 1_48, 1_32, 1_21] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 __SCREAMING_SNAKE_CASE :Union[str, Any] = self.dummy_unet_condition __SCREAMING_SNAKE_CASE :List[str] = AudioDiffusionPipeline( vqvae=self.dummy_vqvae_and_unet[0] ,unet=SCREAMING_SNAKE_CASE__ ,mel=SCREAMING_SNAKE_CASE__ ,scheduler=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) np.random.seed(0 ) __SCREAMING_SNAKE_CASE :Optional[Any] = torch.rand((1, 1, 10) ) __SCREAMING_SNAKE_CASE :Union[str, Any] = pipe(generator=SCREAMING_SNAKE_CASE__ ,encoding=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :List[str] = output.images[0] __SCREAMING_SNAKE_CASE :List[str] = np.frombuffer(image.tobytes() ,dtype='''uint8''' )[:10] __SCREAMING_SNAKE_CASE :Any = np.array([1_07, 1_03, 1_20, 1_27, 1_42, 1_22, 1_13, 1_22, 97, 1_11] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0 @slow @require_torch_gpu class _SCREAMING_SNAKE_CASE( unittest.TestCase ): def _UpperCamelCase ( self ) -> str: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCamelCase ( self ) -> str: """simple docstring""" __SCREAMING_SNAKE_CASE :str = torch_device __SCREAMING_SNAKE_CASE :str = DiffusionPipeline.from_pretrained('''teticio/audio-diffusion-ddim-256''' ) __SCREAMING_SNAKE_CASE :List[str] = pipe.to(SCREAMING_SNAKE_CASE__ ) pipe.set_progress_bar_config(disable=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = torch.Generator(device=SCREAMING_SNAKE_CASE__ ).manual_seed(42 ) __SCREAMING_SNAKE_CASE :str = pipe(generator=SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Dict = output.audios[0] __SCREAMING_SNAKE_CASE :List[str] = output.images[0] assert audio.shape == (1, (pipe.unet.config.sample_size[1] - 1) * pipe.mel.hop_length) assert image.height == pipe.unet.config.sample_size[0] and image.width == pipe.unet.config.sample_size[1] __SCREAMING_SNAKE_CASE :int = np.frombuffer(image.tobytes() ,dtype='''uint8''' )[:10] __SCREAMING_SNAKE_CASE :Any = np.array([1_51, 1_67, 1_54, 1_44, 1_22, 1_34, 1_21, 1_05, 70, 26] ) assert np.abs(image_slice.flatten() - expected_slice ).max() == 0
364
"""simple docstring""" import numpy as np import torch import tqdm from ...models.unet_ad import UNetaDModel from ...pipelines import DiffusionPipeline from ...utils import randn_tensor from ...utils.dummy_pt_objects import DDPMScheduler class _SCREAMING_SNAKE_CASE( A ): def __init__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,) -> Optional[int]: """simple docstring""" super().__init__() __SCREAMING_SNAKE_CASE :Any = value_function __SCREAMING_SNAKE_CASE :List[str] = unet __SCREAMING_SNAKE_CASE :int = scheduler __SCREAMING_SNAKE_CASE :Optional[int] = env __SCREAMING_SNAKE_CASE :Optional[int] = env.get_dataset() __SCREAMING_SNAKE_CASE :str = {} for key in self.data.keys(): try: __SCREAMING_SNAKE_CASE :Optional[Any] = self.data[key].mean() except: # noqa: E722 pass __SCREAMING_SNAKE_CASE :Dict = {} for key in self.data.keys(): try: __SCREAMING_SNAKE_CASE :str = self.data[key].std() except: # noqa: E722 pass __SCREAMING_SNAKE_CASE :Optional[int] = env.observation_space.shape[0] __SCREAMING_SNAKE_CASE :int = env.action_space.shape[0] def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> str: """simple docstring""" return (x_in - self.means[key]) / self.stds[key] def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> Dict: """simple docstring""" return x_in * self.stds[key] + self.means[key] def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" if type(SCREAMING_SNAKE_CASE__ ) is dict: return {k: self.to_torch(SCREAMING_SNAKE_CASE__ ) for k, v in x_in.items()} elif torch.is_tensor(SCREAMING_SNAKE_CASE__ ): return x_in.to(self.unet.device ) return torch.tensor(SCREAMING_SNAKE_CASE__ ,device=self.unet.device ) def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> Union[str, Any]: """simple docstring""" for key, val in cond.items(): __SCREAMING_SNAKE_CASE :Dict = val.clone() return x_in def _UpperCamelCase ( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) -> int: """simple docstring""" __SCREAMING_SNAKE_CASE :Union[str, Any] = x.shape[0] __SCREAMING_SNAKE_CASE :List[Any] = None for i in tqdm.tqdm(self.scheduler.timesteps ): # create batch of timesteps to pass into model __SCREAMING_SNAKE_CASE :Tuple = torch.full((batch_size,) ,SCREAMING_SNAKE_CASE__ ,device=self.unet.device ,dtype=torch.long ) for _ in range(SCREAMING_SNAKE_CASE__ ): with torch.enable_grad(): x.requires_grad_() # permute to match dimension for pre-trained models __SCREAMING_SNAKE_CASE :str = self.value_function(x.permute(0 ,2 ,1 ) ,SCREAMING_SNAKE_CASE__ ).sample __SCREAMING_SNAKE_CASE :Union[str, Any] = torch.autograd.grad([y.sum()] ,[x] )[0] __SCREAMING_SNAKE_CASE :int = self.scheduler._get_variance(SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :Optional[Any] = torch.exp(0.5 * posterior_variance ) __SCREAMING_SNAKE_CASE :List[str] = model_std * grad __SCREAMING_SNAKE_CASE :Dict = 0 __SCREAMING_SNAKE_CASE :List[Any] = x.detach() __SCREAMING_SNAKE_CASE :Union[str, Any] = x + scale * grad __SCREAMING_SNAKE_CASE :Any = self.reset_xa(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,self.action_dim ) __SCREAMING_SNAKE_CASE :Optional[int] = self.unet(x.permute(0 ,2 ,1 ) ,SCREAMING_SNAKE_CASE__ ).sample.permute(0 ,2 ,1 ) # TODO: verify deprecation of this kwarg __SCREAMING_SNAKE_CASE :Optional[int] = self.scheduler.step(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,predict_epsilon=SCREAMING_SNAKE_CASE__ )['''prev_sample'''] # apply conditions to the trajectory (set the initial state) __SCREAMING_SNAKE_CASE :List[str] = self.reset_xa(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,self.action_dim ) __SCREAMING_SNAKE_CASE :Dict = self.to_torch(SCREAMING_SNAKE_CASE__ ) return x, y def __call__( self ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__=64 ,SCREAMING_SNAKE_CASE__=32 ,SCREAMING_SNAKE_CASE__=2 ,SCREAMING_SNAKE_CASE__=0.1 ) -> Union[str, Any]: """simple docstring""" __SCREAMING_SNAKE_CASE :Tuple = self.normalize(SCREAMING_SNAKE_CASE__ ,'''observations''' ) __SCREAMING_SNAKE_CASE :List[Any] = obs[None].repeat(SCREAMING_SNAKE_CASE__ ,axis=0 ) __SCREAMING_SNAKE_CASE :str = {0: self.to_torch(SCREAMING_SNAKE_CASE__ )} __SCREAMING_SNAKE_CASE :Optional[Any] = (batch_size, planning_horizon, self.state_dim + self.action_dim) # generate initial noise and apply our conditions (to make the trajectories start at current state) __SCREAMING_SNAKE_CASE :Optional[int] = randn_tensor(SCREAMING_SNAKE_CASE__ ,device=self.unet.device ) __SCREAMING_SNAKE_CASE :Tuple = self.reset_xa(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,self.action_dim ) __SCREAMING_SNAKE_CASE :Any = self.to_torch(SCREAMING_SNAKE_CASE__ ) # run the diffusion process __SCREAMING_SNAKE_CASE ,__SCREAMING_SNAKE_CASE :Tuple = self.run_diffusion(SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ,SCREAMING_SNAKE_CASE__ ) # sort output trajectories by value __SCREAMING_SNAKE_CASE :Any = y.argsort(0 ,descending=SCREAMING_SNAKE_CASE__ ).squeeze() __SCREAMING_SNAKE_CASE :Any = x[sorted_idx] __SCREAMING_SNAKE_CASE :str = sorted_values[:, :, : self.action_dim] __SCREAMING_SNAKE_CASE :Union[str, Any] = actions.detach().cpu().numpy() __SCREAMING_SNAKE_CASE :Optional[int] = self.de_normalize(SCREAMING_SNAKE_CASE__ ,key='''actions''' ) # select the action with the highest value if y is not None: __SCREAMING_SNAKE_CASE :Optional[int] = 0 else: # if we didn't run value guiding, select a random action __SCREAMING_SNAKE_CASE :Any = np.random.randint(0 ,SCREAMING_SNAKE_CASE__ ) __SCREAMING_SNAKE_CASE :int = denorm_actions[selected_index, 0] return denorm_actions
239
0
"""simple docstring""" def UpperCAmelCase__ ( SCREAMING_SNAKE_CASE : list ): '''simple docstring''' lowerCAmelCase = 0 while len(SCREAMING_SNAKE_CASE ) > 1: lowerCAmelCase = 0 # Consider two files with minimum cost to be merged for _ in range(2 ): lowerCAmelCase = files.index(min(SCREAMING_SNAKE_CASE ) ) temp += files[min_index] files.pop(SCREAMING_SNAKE_CASE ) files.append(SCREAMING_SNAKE_CASE ) optimal_merge_cost += temp return optimal_merge_cost if __name__ == "__main__": import doctest doctest.testmod()
46
def lowerCAmelCase__ ( SCREAMING_SNAKE_CASE_: int ) -> list: '''simple docstring''' A__ = int(SCREAMING_SNAKE_CASE_ ) if n_element < 1: A__ = ValueError("a should be a positive number" ) raise my_error A__ = [1] A__ , A__ , A__ = (0, 0, 0) A__ = 1 while index < n_element: while hamming_list[i] * 2 <= hamming_list[-1]: i += 1 while hamming_list[j] * 3 <= hamming_list[-1]: j += 1 while hamming_list[k] * 5 <= hamming_list[-1]: k += 1 hamming_list.append( min(hamming_list[i] * 2 , hamming_list[j] * 3 , hamming_list[k] * 5 ) ) index += 1 return hamming_list if __name__ == "__main__": lowerCAmelCase__ = input("""Enter the last number (nth term) of the Hamming Number Series: """) print("""Formula of Hamming Number Series => 2^i * 3^j * 5^k""") lowerCAmelCase__ = hamming(int(n)) print("""-----------------------------------------------------""") print(f"""The list with nth numbers is: {hamming_numbers}""") print("""-----------------------------------------------------""")
68
0
import warnings from typing import Dict, List, Optional, Tuple from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging lowercase : Dict = logging.get_logger(__name__) class A ( __snake_case ): __magic_name__ = ['''input_ids''', '''attention_mask'''] def __init__( self , SCREAMING_SNAKE_CASE="</s>" , SCREAMING_SNAKE_CASE="<unk>" , SCREAMING_SNAKE_CASE="<pad>" , SCREAMING_SNAKE_CASE=125 , SCREAMING_SNAKE_CASE=None , **SCREAMING_SNAKE_CASE , ) -> Any: """simple docstring""" if extra_ids > 0 and additional_special_tokens is None: A : Optional[Any] = [F'<extra_id_{i}>' for i in range(lowerCamelCase_ )] elif extra_ids > 0 and additional_special_tokens is not None: # Check that we have the right number of extra_id special tokens A : Optional[int] = len(set(filter(lambda SCREAMING_SNAKE_CASE : bool('''extra_id''' in str(lowerCamelCase_ ) ) , lowerCamelCase_ ) ) ) if extra_tokens != extra_ids: raise ValueError( F'Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are' ''' provided to ByT5Tokenizer. In this case the additional_special_tokens must include the''' ''' extra_ids tokens''' ) A : Optional[Any] = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else pad_token A : str = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else eos_token A : int = AddedToken(lowerCamelCase_ , lstrip=lowerCamelCase_ , rstrip=lowerCamelCase_ ) if isinstance(lowerCamelCase_ , lowerCamelCase_ ) else unk_token super().__init__( eos_token=lowerCamelCase_ , unk_token=lowerCamelCase_ , pad_token=lowerCamelCase_ , extra_ids=lowerCamelCase_ , additional_special_tokens=lowerCamelCase_ , **lowerCamelCase_ , ) A : List[Any] = extra_ids A : Any = 2**8 # utf is 8 bits # define special tokens dict A : Optional[Any] = { self.pad_token: 0, self.eos_token: 1, self.unk_token: 2, } A : Dict = len(self.special_tokens_encoder ) A : Optional[int] = len(lowerCamelCase_ ) for i, token in enumerate(lowerCamelCase_ ): A : Union[str, Any] = self.vocab_size + i - n A : Any = {v: k for k, v in self.special_tokens_encoder.items()} @property def __lowerCAmelCase ( self ) -> Union[str, Any]: """simple docstring""" return self._utf_vocab_size + self._num_special_tokens + self._extra_ids def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None , SCREAMING_SNAKE_CASE = False ) -> Optional[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=lowerCamelCase_ , token_ids_a=lowerCamelCase_ , already_has_special_tokens=lowerCamelCase_ ) # normal case: some special tokens if token_ids_a is None: return ([0] * len(lowerCamelCase_ )) + [1] return ([0] * len(lowerCamelCase_ )) + [1] + ([0] * len(lowerCamelCase_ )) + [1] def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> List[Any]: """simple docstring""" if len(lowerCamelCase_ ) > 0 and token_ids[-1] == self.eos_token_id: warnings.warn( F'This sequence already has {self.eos_token}. In future versions this behavior may lead to duplicated' ''' eos tokens being added.''' ) return token_ids else: return token_ids + [self.eos_token_id] def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> int: """simple docstring""" A : Any = [self.eos_token_id] if token_ids_a is None: return len(token_ids_a + eos ) * [0] return len(token_ids_a + eos + token_ids_a + eos ) * [0] def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> Tuple: """simple docstring""" A : List[Any] = self._add_eos_if_not_present(lowerCamelCase_ ) if token_ids_a is None: return token_ids_a else: A : int = self._add_eos_if_not_present(lowerCamelCase_ ) return token_ids_a + token_ids_a def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" A : int = [chr(lowerCamelCase_ ) for i in text.encode('''utf-8''' )] return tokens def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" if token in self.special_tokens_encoder: A : List[str] = self.special_tokens_encoder[token] elif token in self.added_tokens_encoder: A : List[Any] = self.added_tokens_encoder[token] elif len(lowerCamelCase_ ) != 1: A : Tuple = self.unk_token_id else: A : Optional[int] = ord(lowerCamelCase_ ) + self._num_special_tokens return token_id def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[int]: """simple docstring""" if index in self.special_tokens_decoder: A : Tuple = self.special_tokens_decoder[index] else: A : Tuple = chr(index - self._num_special_tokens ) return token def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE ) -> Optional[Any]: """simple docstring""" A : List[Any] = b'''''' for token in tokens: if token in self.special_tokens_decoder: A : int = self.special_tokens_decoder[token].encode('''utf-8''' ) elif token in self.added_tokens_decoder: A : Optional[Any] = self.special_tokens_decoder[token].encode('''utf-8''' ) elif token in self.special_tokens_encoder: A : List[str] = token.encode('''utf-8''' ) elif token in self.added_tokens_encoder: A : List[str] = token.encode('''utf-8''' ) else: A : List[Any] = bytes([ord(lowerCamelCase_ )] ) bstring += tok_string A : Dict = bstring.decode('''utf-8''' , errors='''ignore''' ) return string def __lowerCAmelCase ( self , SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE = None ) -> int: """simple docstring""" return ()
355
'''simple docstring''' # Function to print upper half of diamond (pyramid) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' for i in range(0 , snake_case__ ): for _ in range(0 , n - i - 1 ): # printing spaces print(''' ''' , end='''''' ) for _ in range(0 , i + 1 ): # printing stars print('''* ''' , end='''''' ) print() def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' for i in range(snake_case__ , 0 , -1 ): for _ in range(snake_case__ , 0 , -1 ): # printing stars print('''* ''' , end='''''' ) print() for _ in range(n - i + 1 , 0 , -1 ): # printing spaces print(''' ''' , end='''''' ) def lowerCAmelCase_ ( snake_case__ ): '''simple docstring''' if n <= 0: print(''' ... .... nothing printing :(''' ) return floyd(snake_case__ ) # upper half reverse_floyd(snake_case__ ) # lower half if __name__ == "__main__": print(R'| /\ | |- | |- |--| |\ /| |-') print(R'|/ \| |- |_ |_ |__| | \/ | |_') lowercase : List[str] = 1 while K: lowercase : List[Any] = int(input('enter the number and , and see the magic : ')) print() pretty_print(user_number) lowercase : Any = int(input('press 0 to exit... and 1 to continue...')) print('Good Bye...')
311
0
'''simple docstring''' import flax.linen as nn import jax.numpy as jnp from .attention_flax import FlaxTransformeraDModel from .resnet_flax import FlaxDownsampleaD, FlaxResnetBlockaD, FlaxUpsampleaD class lowercase_ ( nn.Module ): __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 0.0 __UpperCAmelCase = 1 __UpperCAmelCase = 1 __UpperCAmelCase = True __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = jnp.floataa def __a ( self ): UpperCamelCase__ = [] UpperCamelCase__ = [] for i in range(self.num_layers ): UpperCamelCase__ = self.in_channels if i == 0 else self.out_channels UpperCamelCase__ = FlaxResnetBlockaD( in_channels=a , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(a ) UpperCamelCase__ = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(a ) UpperCamelCase__ = resnets UpperCamelCase__ = attentions if self.add_downsample: UpperCamelCase__ = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , a , a , a , a=True ): UpperCamelCase__ = () for resnet, attn in zip(self.resnets , self.attentions ): UpperCamelCase__ = resnet(a , a , deterministic=a ) UpperCamelCase__ = attn(a , a , deterministic=a ) output_states += (hidden_states,) if self.add_downsample: UpperCamelCase__ = self.downsamplers_a(a ) output_states += (hidden_states,) return hidden_states, output_states class lowercase_ ( nn.Module ): __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 0.0 __UpperCAmelCase = 1 __UpperCAmelCase = True __UpperCAmelCase = jnp.floataa def __a ( self ): UpperCamelCase__ = [] for i in range(self.num_layers ): UpperCamelCase__ = self.in_channels if i == 0 else self.out_channels UpperCamelCase__ = FlaxResnetBlockaD( in_channels=a , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(a ) UpperCamelCase__ = resnets if self.add_downsample: UpperCamelCase__ = FlaxDownsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , a , a , a=True ): UpperCamelCase__ = () for resnet in self.resnets: UpperCamelCase__ = resnet(a , a , deterministic=a ) output_states += (hidden_states,) if self.add_downsample: UpperCamelCase__ = self.downsamplers_a(a ) output_states += (hidden_states,) return hidden_states, output_states class lowercase_ ( nn.Module ): __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 0.0 __UpperCAmelCase = 1 __UpperCAmelCase = 1 __UpperCAmelCase = True __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = jnp.floataa def __a ( self ): UpperCamelCase__ = [] UpperCamelCase__ = [] for i in range(self.num_layers ): UpperCamelCase__ = self.in_channels if (i == self.num_layers - 1) else self.out_channels UpperCamelCase__ = self.prev_output_channel if i == 0 else self.out_channels UpperCamelCase__ = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(a ) UpperCamelCase__ = FlaxTransformeraDModel( in_channels=self.out_channels , n_heads=self.num_attention_heads , d_head=self.out_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , only_cross_attention=self.only_cross_attention , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(a ) UpperCamelCase__ = resnets UpperCamelCase__ = attentions if self.add_upsample: UpperCamelCase__ = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , a , a , a , a , a=True ): for resnet, attn in zip(self.resnets , self.attentions ): # pop res hidden states UpperCamelCase__ = res_hidden_states_tuple[-1] UpperCamelCase__ = res_hidden_states_tuple[:-1] UpperCamelCase__ = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) UpperCamelCase__ = resnet(a , a , deterministic=a ) UpperCamelCase__ = attn(a , a , deterministic=a ) if self.add_upsample: UpperCamelCase__ = self.upsamplers_a(a ) return hidden_states class lowercase_ ( nn.Module ): __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 42 __UpperCAmelCase = 0.0 __UpperCAmelCase = 1 __UpperCAmelCase = True __UpperCAmelCase = jnp.floataa def __a ( self ): UpperCamelCase__ = [] for i in range(self.num_layers ): UpperCamelCase__ = self.in_channels if (i == self.num_layers - 1) else self.out_channels UpperCamelCase__ = self.prev_output_channel if i == 0 else self.out_channels UpperCamelCase__ = FlaxResnetBlockaD( in_channels=resnet_in_channels + res_skip_channels , out_channels=self.out_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(a ) UpperCamelCase__ = resnets if self.add_upsample: UpperCamelCase__ = FlaxUpsampleaD(self.out_channels , dtype=self.dtype ) def __call__( self , a , a , a , a=True ): for resnet in self.resnets: # pop res hidden states UpperCamelCase__ = res_hidden_states_tuple[-1] UpperCamelCase__ = res_hidden_states_tuple[:-1] UpperCamelCase__ = jnp.concatenate((hidden_states, res_hidden_states) , axis=-1 ) UpperCamelCase__ = resnet(a , a , deterministic=a ) if self.add_upsample: UpperCamelCase__ = self.upsamplers_a(a ) return hidden_states class lowercase_ ( nn.Module ): __UpperCAmelCase = 42 __UpperCAmelCase = 0.0 __UpperCAmelCase = 1 __UpperCAmelCase = 1 __UpperCAmelCase = False __UpperCAmelCase = False __UpperCAmelCase = jnp.floataa def __a ( self ): # there is always at least one resnet UpperCamelCase__ = [ FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) ] UpperCamelCase__ = [] for _ in range(self.num_layers ): UpperCamelCase__ = FlaxTransformeraDModel( in_channels=self.in_channels , n_heads=self.num_attention_heads , d_head=self.in_channels // self.num_attention_heads , depth=1 , use_linear_projection=self.use_linear_projection , use_memory_efficient_attention=self.use_memory_efficient_attention , dtype=self.dtype , ) attentions.append(a ) UpperCamelCase__ = FlaxResnetBlockaD( in_channels=self.in_channels , out_channels=self.in_channels , dropout_prob=self.dropout , dtype=self.dtype , ) resnets.append(a ) UpperCamelCase__ = resnets UpperCamelCase__ = attentions def __call__( self , a , a , a , a=True ): UpperCamelCase__ = self.resnets[0](a , a ) for attn, resnet in zip(self.attentions , self.resnets[1:] ): UpperCamelCase__ = attn(a , a , deterministic=a ) UpperCamelCase__ = resnet(a , a , deterministic=a ) return hidden_states
80
"""simple docstring""" import copy from ...configuration_utils import PretrainedConfig from ...utils import add_start_docstrings A_ = r''' [`RagConfig`] stores the configuration of a *RagModel*. Configuration objects inherit from [`PretrainedConfig`] and can be used to control the model outputs. Read the documentation from [`PretrainedConfig`] for more information. Args: title_sep (`str`, *optional*, defaults to `" / "`): Separator inserted between the title and the text of the retrieved document when calling [`RagRetriever`]. doc_sep (`str`, *optional*, defaults to `" // "`): Separator inserted between the text of the retrieved document and the original input when calling [`RagRetriever`]. n_docs (`int`, *optional*, defaults to 5): Number of documents to retrieve. max_combined_length (`int`, *optional*, defaults to 300): Max length of contextualized input returned by [`~RagRetriever.__call__`]. retrieval_vector_size (`int`, *optional*, defaults to 768): Dimensionality of the document embeddings indexed by [`RagRetriever`]. retrieval_batch_size (`int`, *optional*, defaults to 8): Retrieval batch size, defined as the number of queries issues concurrently to the faiss index encapsulated [`RagRetriever`]. dataset (`str`, *optional*, defaults to `"wiki_dpr"`): A dataset identifier of the indexed dataset in HuggingFace Datasets (list all available datasets and ids using `datasets.list_datasets()`). dataset_split (`str`, *optional*, defaults to `"train"`) Which split of the `dataset` to load. index_name (`str`, *optional*, defaults to `"compressed"`) The index name of the index associated with the `dataset`. One can choose between `"legacy"`, `"exact"` and `"compressed"`. index_path (`str`, *optional*) The path to the serialized faiss index on disk. passages_path (`str`, *optional*): A path to text passages compatible with the faiss index. Required if using [`~models.rag.retrieval_rag.LegacyIndex`] use_dummy_dataset (`bool`, *optional*, defaults to `False`) Whether to load a "dummy" variant of the dataset specified by `dataset`. label_smoothing (`float`, *optional*, defaults to 0.0): Only relevant if `return_loss` is set to `True`. Controls the `epsilon` parameter value for label smoothing in the loss calculation. If set to 0, no label smoothing is performed. do_marginalize (`bool`, *optional*, defaults to `False`): If `True`, the logits are marginalized over all documents by making use of `torch.nn.functional.log_softmax`. reduce_loss (`bool`, *optional*, defaults to `False`): Whether or not to reduce the NLL loss using the `torch.Tensor.sum` operation. do_deduplication (`bool`, *optional*, defaults to `True`): Whether or not to deduplicate the generations from different context documents for a given input. Has to be set to `False` if used while training with distributed backend. exclude_bos_score (`bool`, *optional*, defaults to `False`): Whether or not to disregard the BOS token when computing the loss. output_retrieved(`bool`, *optional*, defaults to `False`): If set to `True`, `retrieved_doc_embeds`, `retrieved_doc_ids`, `context_input_ids` and `context_attention_mask` are returned. See returned tensors for more detail. use_cache (`bool`, *optional*, defaults to `True`): Whether or not the model should return the last key/values attentions (not used by all models). forced_eos_token_id (`int`, *optional*): The id of the token to force as the last generated token when `max_length` is reached. Usually set to `eos_token_id`. ''' @add_start_docstrings(__a ) class lowercase( __a ): '''simple docstring''' lowercase__ = "rag" lowercase__ = True def __init__( self: Union[str, Any], a_: int=None, a_: Tuple=True, a_: Optional[int]=None, a_: List[str]=None, a_: int=None, a_: Optional[Any]=None, a_: List[str]=None, a_: Optional[Any]=" / ", a_: Tuple=" // ", a_: List[Any]=5, a_: Dict=300, a_: Tuple=768, a_: Optional[Any]=8, a_: int="wiki_dpr", a_: Any="train", a_: Optional[int]="compressed", a_: Optional[int]=None, a_: List[Any]=None, a_: Optional[Any]=False, a_: str=False, a_: Dict=0.0, a_: Union[str, Any]=True, a_: Union[str, Any]=False, a_: str=False, a_: List[str]=False, a_: Union[str, Any]=True, a_: Any=None, **a_: List[Any], ): '''simple docstring''' super().__init__( bos_token_id=a_, pad_token_id=a_, eos_token_id=a_, decoder_start_token_id=a_, forced_eos_token_id=a_, is_encoder_decoder=a_, prefix=a_, vocab_size=a_, **a_, ) assert ( "question_encoder" in kwargs and "generator" in kwargs ), "Config has to be initialized with question_encoder and generator config" _snake_case : Union[str, Any] = kwargs.pop("""question_encoder""" ) _snake_case : List[str] = question_encoder_config.pop("""model_type""" ) _snake_case : Union[str, Any] = kwargs.pop("""generator""" ) _snake_case : Any = decoder_config.pop("""model_type""" ) from ..auto.configuration_auto import AutoConfig _snake_case : Union[str, Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Optional[Any] = AutoConfig.for_model(a_, **a_ ) _snake_case : Any = reduce_loss _snake_case : Optional[int] = label_smoothing _snake_case : Dict = exclude_bos_score _snake_case : int = do_marginalize _snake_case : Optional[Any] = title_sep _snake_case : Any = doc_sep _snake_case : List[str] = n_docs _snake_case : Tuple = max_combined_length _snake_case : Optional[Any] = dataset _snake_case : Union[str, Any] = dataset_split _snake_case : Tuple = index_name _snake_case : Any = retrieval_vector_size _snake_case : Union[str, Any] = retrieval_batch_size _snake_case : str = passages_path _snake_case : Tuple = index_path _snake_case : List[Any] = use_dummy_dataset _snake_case : Optional[Any] = output_retrieved _snake_case : Tuple = do_deduplication _snake_case : Union[str, Any] = use_cache if self.forced_eos_token_id is None: _snake_case : Dict = getattr(self.generator, """forced_eos_token_id""", a_ ) @classmethod def UpperCamelCase_ ( cls: Any, a_: PretrainedConfig, a_: PretrainedConfig, **a_: Optional[Any] ): '''simple docstring''' return cls(question_encoder=question_encoder_config.to_dict(), generator=generator_config.to_dict(), **a_ ) def UpperCamelCase_ ( self: Tuple ): '''simple docstring''' _snake_case : Optional[int] = copy.deepcopy(self.__dict__ ) _snake_case : List[str] = self.question_encoder.to_dict() _snake_case : Tuple = self.generator.to_dict() _snake_case : Dict = self.__class__.model_type return output
64
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _A = { '''configuration_squeezebert''': [ '''SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''SqueezeBertConfig''', '''SqueezeBertOnnxConfig''', ], '''tokenization_squeezebert''': ['''SqueezeBertTokenizer'''], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = ['''SqueezeBertTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _A = [ '''SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''SqueezeBertForMaskedLM''', '''SqueezeBertForMultipleChoice''', '''SqueezeBertForQuestionAnswering''', '''SqueezeBertForSequenceClassification''', '''SqueezeBertForTokenClassification''', '''SqueezeBertModel''', '''SqueezeBertModule''', '''SqueezeBertPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_squeezebert import ( SQUEEZEBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, SqueezeBertConfig, SqueezeBertOnnxConfig, ) from .tokenization_squeezebert import SqueezeBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_squeezebert_fast import SqueezeBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_squeezebert import ( SQUEEZEBERT_PRETRAINED_MODEL_ARCHIVE_LIST, SqueezeBertForMaskedLM, SqueezeBertForMultipleChoice, SqueezeBertForQuestionAnswering, SqueezeBertForSequenceClassification, SqueezeBertForTokenClassification, SqueezeBertModel, SqueezeBertModule, SqueezeBertPreTrainedModel, ) else: import sys _A = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
167
import unittest import numpy as np import torch from diffusers import KarrasVePipeline, KarrasVeScheduler, UNetaDModel from diffusers.utils.testing_utils import enable_full_determinism, require_torch, slow, torch_device enable_full_determinism() class A ( unittest.TestCase ): @property def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" torch.manual_seed(0 ) lowerCAmelCase_ = UNetaDModel( block_out_channels=(32, 64), layers_per_block=2, sample_size=32, in_channels=3, out_channels=3, down_block_types=('''DownBlock2D''', '''AttnDownBlock2D'''), up_block_types=('''AttnUpBlock2D''', '''UpBlock2D'''), ) return model def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = self.dummy_uncond_unet lowerCAmelCase_ = KarrasVeScheduler() lowerCAmelCase_ = KarrasVePipeline(unet=UpperCamelCase__, scheduler=UpperCamelCase__ ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCAmelCase_ = torch.manual_seed(0 ) lowerCAmelCase_ = pipe(num_inference_steps=2, generator=UpperCamelCase__, output_type='''numpy''' ).images lowerCAmelCase_ = torch.manual_seed(0 ) lowerCAmelCase_ = pipe(num_inference_steps=2, generator=UpperCamelCase__, output_type='''numpy''', return_dict=UpperCamelCase__ )[0] lowerCAmelCase_ = image[0, -3:, -3:, -1] lowerCAmelCase_ = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 32, 32, 3) lowerCAmelCase_ = np.array([0.0, 1.0, 0.0, 0.0, 0.0, 1.0, 0.0, 0.0, 0.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 @slow @require_torch class A ( unittest.TestCase ): def SCREAMING_SNAKE_CASE__ ( self ): """simple docstring""" lowerCAmelCase_ = '''google/ncsnpp-celebahq-256''' lowerCAmelCase_ = UNetaDModel.from_pretrained(UpperCamelCase__ ) lowerCAmelCase_ = KarrasVeScheduler() lowerCAmelCase_ = KarrasVePipeline(unet=UpperCamelCase__, scheduler=UpperCamelCase__ ) pipe.to(UpperCamelCase__ ) pipe.set_progress_bar_config(disable=UpperCamelCase__ ) lowerCAmelCase_ = torch.manual_seed(0 ) lowerCAmelCase_ = pipe(num_inference_steps=20, generator=UpperCamelCase__, output_type='''numpy''' ).images lowerCAmelCase_ = image[0, -3:, -3:, -1] assert image.shape == (1, 256, 256, 3) lowerCAmelCase_ = np.array([0.578, 0.5_811, 0.5_924, 0.5_809, 0.587, 0.5_886, 0.5_861, 0.5_802, 0.586] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
167
1
'''simple docstring''' import gc import tempfile import unittest import numpy as np import torch from diffusers import VersatileDiffusionTextToImagePipeline from diffusers.utils.testing_utils import nightly, require_torch_gpu, torch_device lowercase : Optional[Any] = False class __UpperCAmelCase ( unittest.TestCase ): pass @nightly @require_torch_gpu class __UpperCAmelCase ( unittest.TestCase ): def lowerCamelCase ( self ): """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained('shi-labs/versatile-diffusion' ) # remove text_unet pipe.remove_unused_weights() pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images with tempfile.TemporaryDirectory() as tmpdirname: pipe.save_pretrained(lowerCAmelCase_ ) _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained(lowerCAmelCase_ ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = generator.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=2 , output_type='numpy' ).images assert np.abs(image - new_image ).sum() < 1E-5, "Models don't have the same forward pass" def lowerCamelCase ( self ): """simple docstring""" _snake_case = VersatileDiffusionTextToImagePipeline.from_pretrained( 'shi-labs/versatile-diffusion' , torch_dtype=torch.floataa ) pipe.to(lowerCAmelCase_ ) pipe.set_progress_bar_config(disable=lowerCAmelCase_ ) _snake_case = 'A painting of a squirrel eating a burger ' _snake_case = torch.manual_seed(0 ) _snake_case = pipe( prompt=lowerCAmelCase_ , generator=lowerCAmelCase_ , guidance_scale=7.5 , num_inference_steps=50 , output_type='numpy' ).images _snake_case = image[0, 2_53:2_56, 2_53:2_56, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case = np.array([0.3367, 0.3169, 0.2656, 0.3870, 0.4790, 0.3796, 0.4009, 0.4878, 0.4778] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2
42
from __future__ import annotations def _a ( SCREAMING_SNAKE_CASE , SCREAMING_SNAKE_CASE ): """simple docstring""" lowercase__ = 0 lowercase__ = len(SCREAMING_SNAKE_CASE ) - 1 while i < j: if nums[i] + nums[j] == target: return [i, j] elif nums[i] + nums[j] < target: lowercase__ = i + 1 else: lowercase__ = j - 1 return [] if __name__ == "__main__": import doctest doctest.testmod() print(f"""{two_pointer([2, 7, 11, 15], 9) = }""")
110
0
"""simple docstring""" def a__ ( ): UpperCAmelCase_ = [31, 28, 31, 30, 31, 30, 31, 31, 30, 31, 30, 31] UpperCAmelCase_ = 6 UpperCAmelCase_ = 1 UpperCAmelCase_ = 1901 UpperCAmelCase_ = 0 while year < 2001: day += 7 if (year % 4 == 0 and year % 100 != 0) or (year % 400 == 0): if day > days_per_month[month - 1] and month != 2: month += 1 UpperCAmelCase_ = day - days_per_month[month - 2] elif day > 29 and month == 2: month += 1 UpperCAmelCase_ = day - 29 else: if day > days_per_month[month - 1]: month += 1 UpperCAmelCase_ = day - days_per_month[month - 2] if month > 12: year += 1 UpperCAmelCase_ = 1 if year < 2001 and day == 1: sundays += 1 return sundays if __name__ == "__main__": print(solution())
241
"""simple docstring""" import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def a__ ( lowerCAmelCase__ , lowerCAmelCase__ ): UpperCAmelCase_ = checkpoint UpperCAmelCase_ = {} UpperCAmelCase_ = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase_ = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase_ = vae_state_dict["quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["quant_conv.bias"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase_ = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(lowerCAmelCase__ ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase_ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase_ = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(lowerCAmelCase__ ) } for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase_ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) conv_attn_to_linear(lowerCAmelCase__ ) for i in range(lowerCAmelCase__ ): UpperCAmelCase_ = num_up_blocks - 1 - i UpperCAmelCase_ = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase_ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase_ = 2 for i in range(1 , num_mid_res_blocks + 1 ): UpperCAmelCase_ = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase_ = renew_vae_resnet_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) UpperCAmelCase_ = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase_ = renew_vae_attention_paths(lowerCAmelCase__ ) UpperCAmelCase_ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ , lowerCAmelCase__ , additional_replacements=[meta_path] , config=lowerCAmelCase__ ) conv_attn_to_linear(lowerCAmelCase__ ) return new_checkpoint def a__ ( lowerCAmelCase__ , lowerCAmelCase__ , ): # Only support V1 UpperCAmelCase_ = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase_ = io.BytesIO(r.content ) UpperCAmelCase_ = OmegaConf.load(lowerCAmelCase__ ) UpperCAmelCase_ = 512 UpperCAmelCase_ = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase_ = {} with safe_open(lowerCAmelCase__ , framework="pt" , device="cpu" ) as f: for key in f.keys(): UpperCAmelCase_ = f.get_tensor(lowerCAmelCase__ ) else: UpperCAmelCase_ = torch.load(lowerCAmelCase__ , map_location=lowerCAmelCase__ )["state_dict"] # Convert the VAE model. UpperCAmelCase_ = create_vae_diffusers_config(lowerCAmelCase__ , image_size=lowerCAmelCase__ ) UpperCAmelCase_ = custom_convert_ldm_vae_checkpoint(lowerCAmelCase__ , lowerCAmelCase__ ) UpperCAmelCase_ = AutoencoderKL(**lowerCAmelCase__ ) vae.load_state_dict(lowerCAmelCase__ ) vae.save_pretrained(lowerCAmelCase__ ) if __name__ == "__main__": lowerCamelCase = argparse.ArgumentParser() parser.add_argument("""--vae_pt_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") parser.add_argument("""--dump_path""", default=None, type=str, required=True, help="""Path to the VAE.pt to convert.""") lowerCamelCase = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
241
1
"""simple docstring""" def SCREAMING_SNAKE_CASE ( _lowerCamelCase : int ) -> bool: if not isinstance(_lowerCamelCase ,_lowerCamelCase ): raise ValueError("""check_bouncy() accepts only integer arguments""" ) _lowerCAmelCase : Any = str(_lowerCamelCase ) _lowerCAmelCase : List[Any] = """""".join(sorted(_lowerCamelCase ) ) return sorted_str_n != str_n and sorted_str_n[::-1] != str_n def SCREAMING_SNAKE_CASE ( _lowerCamelCase : float = 99 ) -> int: if not 0 < percent < 100: raise ValueError("""solution() only accepts values from 0 to 100""" ) _lowerCAmelCase : int = 0 _lowerCAmelCase : str = 1 while True: if check_bouncy(_lowerCamelCase ): bouncy_num += 1 if (bouncy_num / num) * 100 >= percent: return num num += 1 if __name__ == "__main__": from doctest import testmod testmod() print(F"""{solution(99)}""")
44
"""simple docstring""" import numpy as np import qiskit def SCREAMING_SNAKE_CASE ( _lowerCamelCase : int = 8 ,_lowerCamelCase : int | None = None ) -> str: _lowerCAmelCase : int = np.random.default_rng(seed=_lowerCamelCase ) # Roughly 25% of the qubits will contribute to the key. # So we take more than we need. _lowerCAmelCase : Tuple = 6 * key_len # Measurement basis for Alice's qubits. _lowerCAmelCase : Dict = rng.integers(2 ,size=_lowerCamelCase ) # The set of states Alice will prepare. _lowerCAmelCase : Tuple = rng.integers(2 ,size=_lowerCamelCase ) # Measurement basis for Bob's qubits. _lowerCAmelCase : Union[str, Any] = rng.integers(2 ,size=_lowerCamelCase ) # Quantum Circuit to simulate BB84 _lowerCAmelCase : Dict = qiskit.QuantumCircuit(_lowerCamelCase ,name="""BB84""" ) # Alice prepares her qubits according to rules above. for index, _ in enumerate(_lowerCamelCase ): if alice_state[index] == 1: bbaa_circ.x(_lowerCamelCase ) if alice_basis[index] == 1: bbaa_circ.h(_lowerCamelCase ) bbaa_circ.barrier() # Bob measures the received qubits according to rules above. for index, _ in enumerate(_lowerCamelCase ): if bob_basis[index] == 1: bbaa_circ.h(_lowerCamelCase ) bbaa_circ.barrier() bbaa_circ.measure_all() # Simulate the quantum circuit. _lowerCAmelCase : int = qiskit.Aer.get_backend("""aer_simulator""" ) # We only need to run one shot because the key is unique. # Multiple shots will produce the same key. _lowerCAmelCase : List[str] = qiskit.execute(_lowerCamelCase ,_lowerCamelCase ,shots=1 ,seed_simulator=_lowerCamelCase ) # Returns the result of measurement. _lowerCAmelCase : List[Any] = job.result().get_counts(_lowerCamelCase ).most_frequent() # Extracting the generated key from the simulation results. # Only keep measurement results where Alice and Bob chose the same basis. _lowerCAmelCase : str = """""".join( [ result_bit for alice_basis_bit, bob_basis_bit, result_bit in zip( _lowerCamelCase ,_lowerCamelCase ,_lowerCamelCase ) if alice_basis_bit == bob_basis_bit ] ) # Get final key. Pad with 0 if too short, otherwise truncate. _lowerCAmelCase : List[Any] = gen_key[:key_len] if len(_lowerCamelCase ) >= key_len else gen_key.ljust(_lowerCamelCase ,"""0""" ) return key if __name__ == "__main__": print(F"""The generated key is : {bbaa(8, seed=0)}""") from doctest import testmod testmod()
44
1
'''simple docstring''' from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available a_ : Optional[int] = { 'configuration_pegasus_x': ['PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP', 'PegasusXConfig'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: a_ : Optional[Any] = [ 'PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST', 'PegasusXForConditionalGeneration', 'PegasusXModel', 'PegasusXPreTrainedModel', ] if TYPE_CHECKING: from .configuration_pegasus_x import PEGASUS_X_PRETRAINED_CONFIG_ARCHIVE_MAP, PegasusXConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pegasus_x import ( PEGASUS_X_PRETRAINED_MODEL_ARCHIVE_LIST, PegasusXForConditionalGeneration, PegasusXModel, PegasusXPreTrainedModel, ) else: import sys a_ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
355
'''simple docstring''' import argparse import os # New Code # import evaluate import torch from datasets import load_dataset from torch.optim import AdamW from torch.utils.data import DataLoader from transformers import AutoModelForSequenceClassification, AutoTokenizer, get_linear_schedule_with_warmup, set_seed from accelerate import Accelerator, DistributedType from accelerate.utils import find_executable_batch_size ######################################################################## # This is a fully working simple example to use Accelerate, # specifically showcasing how to ensure out-of-memory errors never # interrupt training, and builds off the `nlp_example.py` script. # # This example trains a Bert base model on GLUE MRPC # in any of the following settings (with the same script): # - single CPU or single GPU # - multi GPUS (using PyTorch distributed mode) # - (multi) TPUs # - fp16 (mixed-precision) or fp32 (normal precision) # # New additions from the base script can be found quickly by # looking for the # New Code # tags # # To run it in each of these various modes, follow the instructions # in the readme for examples: # https://github.com/huggingface/accelerate/tree/main/examples # ######################################################################## a_ : Tuple = 16 a_ : Optional[int] = 32 def a_ ( __snake_case : Accelerator , __snake_case : int = 16 ) -> str: """simple docstring""" lowerCamelCase_ =AutoTokenizer.from_pretrained('''bert-base-cased''' ) lowerCamelCase_ =load_dataset('''glue''' , '''mrpc''' ) def tokenize_function(__snake_case : int ): # max_length=None => use the model max length (it's actually the default) lowerCamelCase_ =tokenizer(examples['''sentence1'''] , examples['''sentence2'''] , truncation=__snake_case , max_length=__snake_case ) return outputs # Apply the method we just defined to all the examples in all the splits of the dataset # starting with the main process first: with accelerator.main_process_first(): lowerCamelCase_ =datasets.map( __snake_case , batched=__snake_case , remove_columns=['''idx''', '''sentence1''', '''sentence2'''] , ) # We also rename the 'label' column to 'labels' which is the expected name for labels by the models of the # transformers library lowerCamelCase_ =tokenized_datasets.rename_column('''label''' , '''labels''' ) def collate_fn(__snake_case : Any ): # On TPU it's best to pad everything to the same length or training will be very slow. lowerCamelCase_ =128 if accelerator.distributed_type == DistributedType.TPU else None # When using mixed precision we want round multiples of 8/16 if accelerator.mixed_precision == "fp8": lowerCamelCase_ =16 elif accelerator.mixed_precision != "no": lowerCamelCase_ =8 else: lowerCamelCase_ =None return tokenizer.pad( __snake_case , padding='''longest''' , max_length=__snake_case , pad_to_multiple_of=__snake_case , return_tensors='''pt''' , ) # Instantiate dataloaders. lowerCamelCase_ =DataLoader( tokenized_datasets['''train'''] , shuffle=__snake_case , collate_fn=__snake_case , batch_size=__snake_case ) lowerCamelCase_ =DataLoader( tokenized_datasets['''validation'''] , shuffle=__snake_case , collate_fn=__snake_case , batch_size=__snake_case ) return train_dataloader, eval_dataloader # For testing only if os.environ.get("""TESTING_MOCKED_DATALOADERS""", None) == "1": from accelerate.test_utils.training import mocked_dataloaders a_ : Tuple = mocked_dataloaders # noqa: F811 def a_ ( __snake_case : List[str] , __snake_case : Tuple ) -> Optional[Any]: """simple docstring""" # For testing only if os.environ.get('''TESTING_MOCKED_DATALOADERS''' , __snake_case ) == "1": lowerCamelCase_ =2 # Initialize accelerator lowerCamelCase_ =Accelerator(cpu=args.cpu , mixed_precision=args.mixed_precision ) # Sample hyper-parameters for learning rate, batch size, seed and a few other HPs lowerCamelCase_ =config['''lr'''] lowerCamelCase_ =int(config['''num_epochs'''] ) lowerCamelCase_ =int(config['''seed'''] ) lowerCamelCase_ =int(config['''batch_size'''] ) lowerCamelCase_ =evaluate.load('''glue''' , '''mrpc''' ) # New Code # # We now can define an inner training loop function. It should take a batch size as the only parameter, # and build the dataloaders in there. # It also gets our decorator @find_executable_batch_size(starting_batch_size=__snake_case ) def inner_training_loop(__snake_case : Union[str, Any] ): # And now just move everything below under this function # We need to bring in the Accelerator object from earlier nonlocal accelerator # And reset all of its attributes that could hold onto any memory: accelerator.free_memory() # Then we can declare the model, optimizer, and everything else: set_seed(__snake_case ) # Instantiate the model (we build the model here so that the seed also control new weights initialization) lowerCamelCase_ =AutoModelForSequenceClassification.from_pretrained('''bert-base-cased''' , return_dict=__snake_case ) # We could avoid this line since the accelerator is set with `device_placement=True` (default value). # Note that if you are placing tensors on devices manually, this line absolutely needs to be before the optimizer # creation otherwise training will not work on TPU (`accelerate` will kindly throw an error to make us aware of that). lowerCamelCase_ =model.to(accelerator.device ) # Instantiate optimizer lowerCamelCase_ =AdamW(params=model.parameters() , lr=__snake_case ) lowerCamelCase_, lowerCamelCase_ =get_dataloaders(__snake_case , __snake_case ) # Instantiate scheduler lowerCamelCase_ =get_linear_schedule_with_warmup( optimizer=__snake_case , num_warmup_steps=100 , num_training_steps=(len(__snake_case ) * num_epochs) , ) # Prepare everything # There is no specific order to remember, we just need to unpack the objects in the same order we gave them to the # prepare method. lowerCamelCase_, lowerCamelCase_, lowerCamelCase_, lowerCamelCase_, lowerCamelCase_ =accelerator.prepare( __snake_case , __snake_case , __snake_case , __snake_case , __snake_case ) # Now we train the model for epoch in range(__snake_case ): model.train() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) lowerCamelCase_ =model(**__snake_case ) lowerCamelCase_ =outputs.loss accelerator.backward(__snake_case ) optimizer.step() lr_scheduler.step() optimizer.zero_grad() model.eval() for step, batch in enumerate(__snake_case ): # We could avoid this line since we set the accelerator with `device_placement=True`. batch.to(accelerator.device ) with torch.no_grad(): lowerCamelCase_ =model(**__snake_case ) lowerCamelCase_ =outputs.logits.argmax(dim=-1 ) lowerCamelCase_, lowerCamelCase_ =accelerator.gather_for_metrics((predictions, batch['''labels''']) ) metric.add_batch( predictions=__snake_case , references=__snake_case , ) lowerCamelCase_ =metric.compute() # Use accelerator.print to print only on the main process. accelerator.print(F'''epoch {epoch}:''' , __snake_case ) # New Code # # And call it at the end with no arguments # Note: You could also refactor this outside of your training loop function inner_training_loop() def a_ ( ) -> Dict: """simple docstring""" lowerCamelCase_ =argparse.ArgumentParser(description='''Simple example of training script.''' ) parser.add_argument( '''--mixed_precision''' , type=__snake_case , default=__snake_case , choices=['''no''', '''fp16''', '''bf16''', '''fp8'''] , help='''Whether to use mixed precision. Choose''' '''between fp16 and bf16 (bfloat16). Bf16 requires PyTorch >= 1.10.''' '''and an Nvidia Ampere GPU.''' , ) parser.add_argument('''--cpu''' , action='''store_true''' , help='''If passed, will train on the CPU.''' ) lowerCamelCase_ =parser.parse_args() lowerCamelCase_ ={'''lr''': 2e-5, '''num_epochs''': 3, '''seed''': 42, '''batch_size''': 16} training_function(__snake_case , __snake_case ) if __name__ == "__main__": main()
6
0
'''simple docstring''' import os import tempfile from functools import partial from unittest import TestCase from unittest.mock import patch import numpy as np import pytest from datasets.arrow_dataset import Dataset from datasets.search import ElasticSearchIndex, FaissIndex, MissingIndex from .utils import require_elasticsearch, require_faiss __lowerCamelCase = pytest.mark.integration @require_faiss class A__ ( _snake_case ): def snake_case_ ( self ) -> Any: '''simple docstring''' A_ = Dataset.from_dict({"""filename""": ["""my_name-train""" + """_""" + str(UpperCamelCase__ ) for x in np.arange(30 ).tolist()]} ) return dset def snake_case_ ( self ) -> Tuple: '''simple docstring''' import faiss A_ = self._create_dummy_dataset() A_ = dset.map( lambda UpperCamelCase__ , UpperCamelCase__ : {"vecs": i * np.ones(5 , dtype=np.floataa )} , with_indices=UpperCamelCase__ , keep_in_memory=UpperCamelCase__ ) A_ = dset.add_faiss_index("""vecs""" , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT ) A_ , A_ = dset.get_nearest_examples("""vecs""" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0] , """my_name-train_29""" ) dset.drop_index("""vecs""" ) def snake_case_ ( self ) -> Tuple: '''simple docstring''' import faiss A_ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="""vecs""" , batch_size=100 , metric_type=faiss.METRIC_INNER_PRODUCT , ) A_ , A_ = dset.get_nearest_examples("""vecs""" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0] , """my_name-train_29""" ) def snake_case_ ( self ) -> Any: '''simple docstring''' import faiss A_ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="""vecs""" , metric_type=faiss.METRIC_INNER_PRODUCT , ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCamelCase__ ) as tmp_file: dset.save_faiss_index("""vecs""" , tmp_file.name ) dset.load_faiss_index("""vecs2""" , tmp_file.name ) os.unlink(tmp_file.name ) A_ , A_ = dset.get_nearest_examples("""vecs2""" , np.ones(5 , dtype=np.floataa ) ) self.assertEqual(examples["""filename"""][0] , """my_name-train_29""" ) def snake_case_ ( self ) -> Dict: '''simple docstring''' A_ = self._create_dummy_dataset() dset.add_faiss_index_from_external_arrays( external_arrays=np.ones((30, 5) ) * np.arange(30 ).reshape(-1 , 1 ) , index_name="""vecs""" ) dset.drop_index("""vecs""" ) self.assertRaises(UpperCamelCase__ , partial(dset.get_nearest_examples , """vecs2""" , np.ones(5 , dtype=np.floataa ) ) ) def snake_case_ ( self ) -> Any: '''simple docstring''' from elasticsearch import Elasticsearch A_ = self._create_dummy_dataset() with patch("""elasticsearch.Elasticsearch.search""" ) as mocked_search, patch( """elasticsearch.client.IndicesClient.create""" ) as mocked_index_create, patch("""elasticsearch.helpers.streaming_bulk""" ) as mocked_bulk: A_ = {"""acknowledged""": True} mocked_bulk.return_value([(True, None)] * 30 ) A_ = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 29}]}} A_ = Elasticsearch() dset.add_elasticsearch_index("""filename""" , es_client=UpperCamelCase__ ) A_ , A_ = dset.get_nearest_examples("""filename""" , """my_name-train_29""" ) self.assertEqual(examples["""filename"""][0] , """my_name-train_29""" ) @require_faiss class A__ ( _snake_case ): def snake_case_ ( self ) -> Dict: '''simple docstring''' import faiss A_ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) # add vectors index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsNotNone(index.faiss_index ) self.assertEqual(index.faiss_index.ntotal , 5 ) index.add_vectors(np.zeros((5, 5) , dtype=np.floataa ) ) self.assertEqual(index.faiss_index.ntotal , 10 ) # single query A_ = np.zeros(5 , dtype=np.floataa ) A_ = 1 A_ , A_ = index.search(UpperCamelCase__ ) self.assertRaises(UpperCamelCase__ , index.search , query.reshape(-1 , 1 ) ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) # batched queries A_ = np.eye(5 , dtype=np.floataa )[::-1] A_ , A_ = index.search_batch(UpperCamelCase__ ) self.assertRaises(UpperCamelCase__ , index.search_batch , queries[0] ) A_ = [scores[0] for scores in total_scores] A_ = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCamelCase__ ) , 0 ) self.assertListEqual([4, 3, 2, 1, 0] , UpperCamelCase__ ) def snake_case_ ( self ) -> Optional[Any]: '''simple docstring''' import faiss A_ = FaissIndex(string_factory="""Flat""" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) A_ = FaissIndex(string_factory="""LSH""" ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexLSH ) with self.assertRaises(UpperCamelCase__ ): A_ = FaissIndex(string_factory="""Flat""" , custom_index=faiss.IndexFlat(5 ) ) def snake_case_ ( self ) -> Any: '''simple docstring''' import faiss A_ = faiss.IndexFlat(5 ) A_ = FaissIndex(custom_index=UpperCamelCase__ ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) self.assertIsInstance(index.faiss_index , faiss.IndexFlat ) def snake_case_ ( self ) -> str: '''simple docstring''' import faiss A_ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5 , dtype=np.floataa ) ) # Setting delete=False and unlinking manually is not pretty... but it is required on Windows to # ensure somewhat stable behaviour. If we don't, we get PermissionErrors. This is an age-old issue. # see https://bugs.python.org/issue14243 and # https://stackoverflow.com/questions/23212435/permission-denied-to-write-to-my-temporary-file/23212515 with tempfile.NamedTemporaryFile(delete=UpperCamelCase__ ) as tmp_file: index.save(tmp_file.name ) A_ = FaissIndex.load(tmp_file.name ) os.unlink(tmp_file.name ) A_ = np.zeros(5 , dtype=np.floataa ) A_ = 1 A_ , A_ = index.search(UpperCamelCase__ ) self.assertGreater(scores[0] , 0 ) self.assertEqual(indices[0] , 1 ) @require_faiss def UpperCAmelCase__ ( UpperCAmelCase__ ) -> List[Any]: import faiss A_ = FaissIndex(metric_type=faiss.METRIC_INNER_PRODUCT ) index.add_vectors(np.eye(5, dtype=np.floataa ) ) A_ = """index.faiss""" A_ = F'''mock://{index_name}''' index.save(UpperCAmelCase__, storage_options=mockfs.storage_options ) A_ = FaissIndex.load(UpperCAmelCase__, storage_options=mockfs.storage_options ) A_ = np.zeros(5, dtype=np.floataa ) A_ = 1 A_ , A_ = index.search(UpperCAmelCase__ ) assert scores[0] > 0 assert indices[0] == 1 @require_elasticsearch class A__ ( _snake_case ): def snake_case_ ( self ) -> Union[str, Any]: '''simple docstring''' from elasticsearch import Elasticsearch with patch("""elasticsearch.Elasticsearch.search""" ) as mocked_search, patch( """elasticsearch.client.IndicesClient.create""" ) as mocked_index_create, patch("""elasticsearch.helpers.streaming_bulk""" ) as mocked_bulk: A_ = Elasticsearch() A_ = {"""acknowledged""": True} A_ = ElasticSearchIndex(es_client=UpperCamelCase__ ) mocked_bulk.return_value([(True, None)] * 3 ) index.add_documents(["""foo""", """bar""", """foobar"""] ) # single query A_ = """foo""" A_ = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 0}]}} A_ , A_ = index.search(UpperCamelCase__ ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # single query with timeout A_ = """foo""" A_ = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 0}]}} A_ , A_ = index.search(UpperCamelCase__ , request_timeout=30 ) self.assertEqual(scores[0] , 1 ) self.assertEqual(indices[0] , 0 ) # batched queries A_ = ["""foo""", """bar""", """foobar"""] A_ = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 1}]}} A_ , A_ = index.search_batch(UpperCamelCase__ ) A_ = [scores[0] for scores in total_scores] A_ = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCamelCase__ ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCamelCase__ ) # batched queries with timeout A_ = ["""foo""", """bar""", """foobar"""] A_ = {"""hits""": {"""hits""": [{"""_score""": 1, """_id""": 1}]}} A_ , A_ = index.search_batch(UpperCamelCase__ , request_timeout=30 ) A_ = [scores[0] for scores in total_scores] A_ = [indices[0] for indices in total_indices] self.assertGreater(np.min(UpperCamelCase__ ) , 0 ) self.assertListEqual([1, 1, 1] , UpperCamelCase__ )
162
'''simple docstring''' import os from typing import Dict, List, Union import tensorflow as tf from keras_nlp.tokenizers import BytePairTokenizer from tensorflow_text import pad_model_inputs from .tokenization_gpta import GPTaTokenizer class A__ ( tf.keras.layers.Layer ): def __init__( self , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = None , UpperCamelCase__ = None ) -> List[str]: '''simple docstring''' super().__init__() A_ = pad_token_id A_ = max_length A_ = vocab A_ = merges A_ = BytePairTokenizer(UpperCamelCase__ , UpperCamelCase__ , sequence_length=UpperCamelCase__ ) @classmethod def snake_case_ ( cls , UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) -> Union[str, Any]: '''simple docstring''' A_ = [""" """.join(UpperCamelCase__ ) for m in tokenizer.bpe_ranks.keys()] A_ = tokenizer.get_vocab() return cls(UpperCamelCase__ , UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def snake_case_ ( cls , UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) -> Dict: '''simple docstring''' A_ = GPTaTokenizer.from_pretrained(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) return cls.from_tokenizer(UpperCamelCase__ , *UpperCamelCase__ , **UpperCamelCase__ ) @classmethod def snake_case_ ( cls , UpperCamelCase__ ) -> List[Any]: '''simple docstring''' return cls(**UpperCamelCase__ ) def snake_case_ ( self ) -> int: '''simple docstring''' return { "vocab": self.vocab, "merges": self.merges, "max_length": self.max_length, "pad_token_id": self.pad_token_id, } def snake_case_ ( self , UpperCamelCase__ , UpperCamelCase__ = None ) -> Dict: '''simple docstring''' A_ = self.tf_tokenizer(UpperCamelCase__ ) A_ = tf.ones_like(UpperCamelCase__ ) if self.pad_token_id is not None: # pad the tokens up to max length A_ = max_length if max_length is not None else self.max_length if max_length is not None: A_ , A_ = pad_model_inputs( UpperCamelCase__ , max_seq_length=UpperCamelCase__ , pad_value=self.pad_token_id ) return {"attention_mask": attention_mask, "input_ids": input_ids}
162
1
import argparse import logging import sys from unittest.mock import patch import run_glue_deebert from transformers.testing_utils import TestCasePlus, get_gpu_count, require_torch_non_multi_gpu, slow logging.basicConfig(level=logging.DEBUG) a_ = logging.getLogger() def __lowercase ( ): UpperCamelCase_ : Dict = argparse.ArgumentParser() parser.add_argument('-f' ) UpperCamelCase_ : Optional[Any] = parser.parse_args() return args.f class _lowercase ( snake_case_ ): def SCREAMING_SNAKE_CASE__ ( self : str ) -> None: """simple docstring""" UpperCamelCase_ : Any = logging.StreamHandler(sys.stdout ) logger.addHandler(snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] , snake_case : List[str] ) -> Optional[Any]: """simple docstring""" UpperCamelCase_ : List[Any] = get_gpu_count() if n_gpu > 1: pass # XXX: doesn't quite work with n_gpu > 1 https://github.com/huggingface/transformers/issues/10560 # script = f"{self.examples_dir_str}/research_projects/deebert/run_glue_deebert.py" # distributed_args = f"-m torch.distributed.launch --nproc_per_node={n_gpu} {script}".split() # cmd = [sys.executable] + distributed_args + args # execute_subprocess_async(cmd, env=self.get_env()) # XXX: test the results - need to save them first into .json file else: args.insert(0 , 'run_glue_deebert.py' ) with patch.object(snake_case , 'argv' , snake_case ): UpperCamelCase_ : Optional[Any] = run_glue_deebert.main() for value in result.values(): self.assertGreaterEqual(snake_case , 0.666 ) @slow @require_torch_non_multi_gpu def SCREAMING_SNAKE_CASE__ ( self : int ) -> List[str]: """simple docstring""" UpperCamelCase_ : Dict = '\n --model_type roberta\n --model_name_or_path roberta-base\n --task_name MRPC\n --do_train\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --max_seq_length 128\n --per_gpu_eval_batch_size=1\n --per_gpu_train_batch_size=8\n --learning_rate 2e-4\n --num_train_epochs 3\n --overwrite_output_dir\n --seed 42\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --save_steps 0\n --overwrite_cache\n --eval_after_first_stage\n '.split() self.run_and_check(snake_case ) UpperCamelCase_ : List[Any] = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --eval_each_highway\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(snake_case ) UpperCamelCase_ : Dict = '\n --model_type roberta\n --model_name_or_path ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --task_name MRPC\n --do_eval\n --do_lower_case\n --data_dir ./tests/fixtures/tests_samples/MRPC/\n --output_dir ./examples/deebert/saved_models/roberta-base/MRPC/two_stage\n --plot_data_dir ./examples/deebert/results/\n --max_seq_length 128\n --early_exit_entropy 0.1\n --eval_highway\n --overwrite_cache\n --per_gpu_eval_batch_size=1\n '.split() self.run_and_check(snake_case )
366
import inspect import math import tempfile import unittest import numpy as np from transformers import ViTMAEConfig from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ViTMAEForPreTraining, ViTMAEModel from transformers.models.vit.modeling_vit import VIT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _lowercase : def __init__( self : List[Any] , snake_case : List[str] , snake_case : str=1_3 , snake_case : Any=3_0 , snake_case : Tuple=2 , snake_case : List[Any]=3 , snake_case : str=True , snake_case : List[Any]=True , snake_case : List[Any]=3_2 , snake_case : Union[str, Any]=5 , snake_case : Dict=4 , snake_case : str=3_7 , snake_case : Optional[int]="gelu" , snake_case : Any=0.1 , snake_case : Optional[Any]=0.1 , snake_case : Union[str, Any]=1_0 , snake_case : List[str]=0.02 , snake_case : Optional[int]=3 , snake_case : str=0.6 , snake_case : Any=None , ) -> Dict: """simple docstring""" UpperCamelCase_ : int = parent UpperCamelCase_ : Optional[Any] = batch_size UpperCamelCase_ : List[str] = image_size UpperCamelCase_ : Optional[Any] = patch_size UpperCamelCase_ : Optional[int] = num_channels UpperCamelCase_ : Union[str, Any] = is_training UpperCamelCase_ : Dict = use_labels UpperCamelCase_ : Tuple = hidden_size UpperCamelCase_ : str = num_hidden_layers UpperCamelCase_ : Tuple = num_attention_heads UpperCamelCase_ : Any = intermediate_size UpperCamelCase_ : Dict = hidden_act UpperCamelCase_ : Tuple = hidden_dropout_prob UpperCamelCase_ : Dict = attention_probs_dropout_prob UpperCamelCase_ : Any = type_sequence_label_size UpperCamelCase_ : str = initializer_range UpperCamelCase_ : Tuple = mask_ratio UpperCamelCase_ : int = scope # in ViTMAE, the expected sequence length = (num_patches + 1) * (1 - config.mask_ratio), rounded above # (we add 1 for the [CLS] token) UpperCamelCase_ : List[Any] = (image_size // patch_size) ** 2 UpperCamelCase_ : Dict = int(math.ceil((1 - mask_ratio) * (num_patches + 1) ) ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> str: """simple docstring""" UpperCamelCase_ : Optional[int] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCamelCase_ : List[str] = None if self.use_labels: UpperCamelCase_ : Optional[int] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCamelCase_ : List[str] = self.get_config() return config, pixel_values, labels def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> Tuple: """simple docstring""" return ViTMAEConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case , initializer_range=self.initializer_range , mask_ratio=self.mask_ratio , ) def SCREAMING_SNAKE_CASE__ ( self : List[str] , snake_case : Dict , snake_case : List[str] , snake_case : List[str] ) -> Optional[int]: """simple docstring""" UpperCamelCase_ : Optional[int] = ViTMAEModel(config=snake_case ) model.to(snake_case ) model.eval() UpperCamelCase_ : int = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] , snake_case : List[str] , snake_case : Optional[int] , snake_case : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_ : Optional[Any] = ViTMAEForPreTraining(snake_case ) model.to(snake_case ) model.eval() UpperCamelCase_ : Tuple = model(snake_case ) UpperCamelCase_ : Tuple = (self.image_size // self.patch_size) ** 2 UpperCamelCase_ : Tuple = self.patch_size**2 * self.num_channels self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) # test greyscale images UpperCamelCase_ : Optional[int] = 1 UpperCamelCase_ : Dict = ViTMAEForPreTraining(snake_case ) model.to(snake_case ) model.eval() UpperCamelCase_ : List[Any] = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) UpperCamelCase_ : Tuple = model(snake_case ) UpperCamelCase_ : Optional[Any] = self.patch_size**2 self.parent.assertEqual(result.logits.shape , (self.batch_size, num_patches, expected_num_channels) ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> List[str]: """simple docstring""" UpperCamelCase_ : List[Any] = self.prepare_config_and_inputs() UpperCamelCase_, UpperCamelCase_, UpperCamelCase_ : int = config_and_inputs UpperCamelCase_ : Union[str, Any] = {'pixel_values': pixel_values} return config, inputs_dict @require_torch class _lowercase ( snake_case_ , snake_case_ , unittest.TestCase ): lowercase = (ViTMAEModel, ViTMAEForPreTraining) if is_torch_available() else () lowercase = {'feature-extraction': ViTMAEModel} if is_torch_available() else {} lowercase = False lowercase = False lowercase = False lowercase = False def SCREAMING_SNAKE_CASE__ ( self : int ) -> Any: """simple docstring""" UpperCamelCase_ : Any = ViTMAEModelTester(self ) UpperCamelCase_ : Tuple = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case , hidden_size=3_7 ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> str: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason='ViTMAE does not use inputs_embeds' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" pass def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Dict: """simple docstring""" UpperCamelCase_, UpperCamelCase_ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ : Any = model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCamelCase_ : int = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case , nn.Linear ) ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> List[str]: """simple docstring""" UpperCamelCase_, UpperCamelCase_ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ : List[str] = model_class(snake_case ) UpperCamelCase_ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCamelCase_ : Union[str, Any] = [*signature.parameters.keys()] UpperCamelCase_ : Optional[int] = ['pixel_values'] self.assertListEqual(arg_names[:1] , snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> int: """simple docstring""" UpperCamelCase_ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> int: """simple docstring""" UpperCamelCase_ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_pretraining(*snake_case ) def SCREAMING_SNAKE_CASE__ ( self : str , snake_case : List[str] , snake_case : Optional[int] , snake_case : Dict ) -> Dict: """simple docstring""" np.random.seed(2 ) UpperCamelCase_ : Optional[int] = int((pt_model.config.image_size // pt_model.config.patch_size) ** 2 ) UpperCamelCase_ : List[str] = np.random.uniform(size=(self.model_tester.batch_size, num_patches) ) UpperCamelCase_ : int = torch.from_numpy(snake_case ) # Add `noise` argument. # PT inputs will be prepared in `super().check_pt_tf_models()` with this added `noise` argument UpperCamelCase_ : Tuple = pt_noise super().check_pt_tf_models(snake_case , snake_case , snake_case ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Union[str, Any]: """simple docstring""" UpperCamelCase_, UpperCamelCase_ : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCamelCase_ : Tuple = model_class(snake_case ) model.to(snake_case ) model.eval() # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): UpperCamelCase_ : str = model(**self._prepare_for_class(snake_case , snake_case ) ) UpperCamelCase_ : Any = outputs[0].cpu().numpy() UpperCamelCase_ : int = 0 with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case ) UpperCamelCase_ : Union[str, Any] = model_class.from_pretrained(snake_case ) model.to(snake_case ) # make random mask reproducible torch.manual_seed(2 ) with torch.no_grad(): UpperCamelCase_ : Any = model(**self._prepare_for_class(snake_case , snake_case ) ) # Make sure we don't have nans UpperCamelCase_ : Optional[Any] = after_outputs[0].cpu().numpy() UpperCamelCase_ : Union[str, Any] = 0 UpperCamelCase_ : int = np.amax(np.abs(out_a - out_a ) ) self.assertLessEqual(snake_case , 1e-5 ) @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self : Any ) -> int: """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Any: """simple docstring""" pass @unittest.skip( reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load\n to get deterministic results.' ) def SCREAMING_SNAKE_CASE__ ( self : List[str] ) -> Tuple: """simple docstring""" pass @unittest.skip(reason='ViTMAE returns a random mask + ids_restore in each forward pass. See test_save_load' ) def SCREAMING_SNAKE_CASE__ ( self : Tuple ) -> int: """simple docstring""" pass @unittest.skip('Will be fixed soon by reducing the size of the model used for common tests.' ) def SCREAMING_SNAKE_CASE__ ( self : Optional[Any] ) -> Optional[Any]: """simple docstring""" pass @slow def SCREAMING_SNAKE_CASE__ ( self : int ) -> List[Any]: """simple docstring""" for model_name in VIT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCamelCase_ : Dict = ViTMAEModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def __lowercase ( ): UpperCamelCase_ : List[Any] = Image.open('./tests/fixtures/tests_samples/COCO/000000039769.png' ) return image @require_torch @require_vision class _lowercase ( unittest.TestCase ): @cached_property def SCREAMING_SNAKE_CASE__ ( self : Union[str, Any] ) -> Dict: """simple docstring""" return ViTImageProcessor.from_pretrained('facebook/vit-mae-base' ) if is_vision_available() else None @slow def SCREAMING_SNAKE_CASE__ ( self : Optional[int] ) -> Optional[int]: """simple docstring""" np.random.seed(2 ) UpperCamelCase_ : List[str] = ViTMAEForPreTraining.from_pretrained('facebook/vit-mae-base' ).to(snake_case ) UpperCamelCase_ : Tuple = self.default_image_processor UpperCamelCase_ : Union[str, Any] = prepare_img() UpperCamelCase_ : Optional[int] = image_processor(images=snake_case , return_tensors='pt' ).to(snake_case ) # prepare a noise vector that will be also used for testing the TF model # (this way we can ensure that the PT and TF models operate on the same inputs) UpperCamelCase_ : int = ViTMAEConfig() UpperCamelCase_ : Any = int((vit_mae_config.image_size // vit_mae_config.patch_size) ** 2 ) UpperCamelCase_ : List[str] = np.random.uniform(size=(1, num_patches) ) # forward pass with torch.no_grad(): UpperCamelCase_ : str = model(**snake_case , noise=torch.from_numpy(snake_case ).to(device=snake_case ) ) # verify the logits UpperCamelCase_ : Dict = torch.Size((1, 1_9_6, 7_6_8) ) self.assertEqual(outputs.logits.shape , snake_case ) UpperCamelCase_ : Union[str, Any] = torch.tensor( [[-0.0548, -1.7023, -0.9325], [0.3721, -0.5670, -0.2233], [0.8235, -1.3878, -0.3524]] ) self.assertTrue(torch.allclose(outputs.logits[0, :3, :3] , expected_slice.to(snake_case ) , atol=1e-4 ) )
50
0
"""simple docstring""" import doctest import logging import os import unittest from pathlib import Path from typing import List, Union import transformers from transformers.testing_utils import require_tf, require_torch, slow _a = logging.getLogger() @unittest.skip("Temporarily disable the doc tests." ) @require_torch @require_tf @slow class _lowerCAmelCase ( unittest.TestCase ): """simple docstring""" def _lowercase ( self : List[str], UpperCAmelCase__ : Path, UpperCAmelCase__ : Union[str, None] = None, UpperCAmelCase__ : Union[List[str], None] = None, UpperCAmelCase__ : Union[str, List[str], None] = None, UpperCAmelCase__ : bool = True, ): __lowercase = [file for file in os.listdir(UpperCAmelCase__ ) if os.path.isfile(os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) )] if identifier is not None: __lowercase = [file for file in files if identifier in file] if n_identifier is not None: if isinstance(UpperCAmelCase__, UpperCAmelCase__ ): for n_ in n_identifier: __lowercase = [file for file in files if n_ not in file] else: __lowercase = [file for file in files if n_identifier not in file] __lowercase = ignore_files or [] ignore_files.append("__init__.py" ) __lowercase = [file for file in files if file not in ignore_files] for file in files: # Open all files print("Testing", UpperCAmelCase__ ) if only_modules: __lowercase = file.split("." )[0] try: __lowercase = getattr(UpperCAmelCase__, UpperCAmelCase__ ) __lowercase = doctest.DocTestSuite(UpperCAmelCase__ ) __lowercase = unittest.TextTestRunner().run(UpperCAmelCase__ ) self.assertIs(len(result.failures ), 0 ) except AttributeError: logger.info(F"""{module_identifier} is not a module.""" ) else: __lowercase = doctest.testfile(str(".." / directory / file ), optionflags=doctest.ELLIPSIS ) self.assertIs(result.failed, 0 ) def _lowercase ( self : Tuple ): __lowercase = Path("src/transformers" ) __lowercase = "modeling" __lowercase = [ "modeling_ctrl.py", "modeling_tf_ctrl.py", ] self.analyze_directory(UpperCAmelCase__, identifier=UpperCAmelCase__, ignore_files=UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ): __lowercase = Path("src/transformers" ) __lowercase = "tokenization" self.analyze_directory(UpperCAmelCase__, identifier=UpperCAmelCase__ ) def _lowercase ( self : int ): __lowercase = Path("src/transformers" ) __lowercase = "configuration" self.analyze_directory(UpperCAmelCase__, identifier=UpperCAmelCase__ ) def _lowercase ( self : Optional[int] ): __lowercase = Path("src/transformers" ) __lowercase = ["configuration", "modeling", "tokenization"] self.analyze_directory(UpperCAmelCase__, n_identifier=UpperCAmelCase__ ) def _lowercase ( self : Optional[Any] ): __lowercase = Path("docs/source" ) __lowercase = ["favicon.ico"] self.analyze_directory(UpperCAmelCase__, ignore_files=UpperCAmelCase__, only_modules=UpperCAmelCase__ )
17
"""simple docstring""" from collections.abc import Sequence def _A ( UpperCamelCase_ : Sequence[float], UpperCamelCase_ : float) -> float: '''simple docstring''' return sum(c * (x**i) for i, c in enumerate(UpperCamelCase_)) def _A ( UpperCamelCase_ : Sequence[float], UpperCamelCase_ : float) -> float: '''simple docstring''' __lowercase = 0.0 for coeff in reversed(UpperCamelCase_): __lowercase = result * x + coeff return result if __name__ == "__main__": _a = (0.0, 0.0, 5.0, 9.3, 7.0) _a = 10.0 print(evaluate_poly(poly, x)) print(horner(poly, x))
17
1
"""simple docstring""" import unittest from transformers import AutoTokenizer, NystromformerConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, NystromformerModel, ) from transformers.models.nystromformer.modeling_nystromformer import NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST class lowercase__ : def __init__( self : Any , UpperCAmelCase_ : str , UpperCAmelCase_ : Optional[Any]=13 , UpperCAmelCase_ : Optional[int]=7 , UpperCAmelCase_ : List[str]=True , UpperCAmelCase_ : Any=True , UpperCAmelCase_ : Optional[int]=True , UpperCAmelCase_ : Dict=True , UpperCAmelCase_ : Tuple=99 , UpperCAmelCase_ : Any=32 , UpperCAmelCase_ : Dict=5 , UpperCAmelCase_ : Union[str, Any]=4 , UpperCAmelCase_ : Optional[int]=37 , UpperCAmelCase_ : Tuple="gelu" , UpperCAmelCase_ : Any=0.1 , UpperCAmelCase_ : str=0.1 , UpperCAmelCase_ : List[str]=512 , UpperCAmelCase_ : List[str]=16 , UpperCAmelCase_ : int=2 , UpperCAmelCase_ : str=0.02 , UpperCAmelCase_ : Optional[int]=3 , UpperCAmelCase_ : str=4 , UpperCAmelCase_ : Any=None , ): SCREAMING_SNAKE_CASE__ = parent SCREAMING_SNAKE_CASE__ = batch_size SCREAMING_SNAKE_CASE__ = seq_length SCREAMING_SNAKE_CASE__ = is_training SCREAMING_SNAKE_CASE__ = use_input_mask SCREAMING_SNAKE_CASE__ = use_token_type_ids SCREAMING_SNAKE_CASE__ = use_labels SCREAMING_SNAKE_CASE__ = vocab_size SCREAMING_SNAKE_CASE__ = hidden_size SCREAMING_SNAKE_CASE__ = num_hidden_layers SCREAMING_SNAKE_CASE__ = num_attention_heads SCREAMING_SNAKE_CASE__ = intermediate_size SCREAMING_SNAKE_CASE__ = hidden_act SCREAMING_SNAKE_CASE__ = hidden_dropout_prob SCREAMING_SNAKE_CASE__ = attention_probs_dropout_prob SCREAMING_SNAKE_CASE__ = max_position_embeddings SCREAMING_SNAKE_CASE__ = type_vocab_size SCREAMING_SNAKE_CASE__ = type_sequence_label_size SCREAMING_SNAKE_CASE__ = initializer_range SCREAMING_SNAKE_CASE__ = num_labels SCREAMING_SNAKE_CASE__ = num_choices SCREAMING_SNAKE_CASE__ = scope def A_ ( self : List[Any] ): SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) SCREAMING_SNAKE_CASE__ = None if self.use_input_mask: SCREAMING_SNAKE_CASE__ = random_attention_mask([self.batch_size, self.seq_length] ) SCREAMING_SNAKE_CASE__ = None if self.use_token_type_ids: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None SCREAMING_SNAKE_CASE__ = None if self.use_labels: SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) SCREAMING_SNAKE_CASE__ = ids_tensor([self.batch_size] , self.num_choices ) SCREAMING_SNAKE_CASE__ = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def A_ ( self : Optional[Any] ): return NystromformerConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=UpperCAmelCase_ , initializer_range=self.initializer_range , ) def A_ ( self : Any , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Tuple ): SCREAMING_SNAKE_CASE__ = NystromformerModel(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def A_ ( self : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE__ = NystromformerForMaskedLM(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def A_ ( self : List[Any] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : str , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Dict ): SCREAMING_SNAKE_CASE__ = NystromformerForQuestionAnswering(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = model( UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , start_positions=UpperCAmelCase_ , end_positions=UpperCAmelCase_ , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def A_ ( self : List[Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Any , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : Union[str, Any] , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : int , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = NystromformerForSequenceClassification(UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def A_ ( self : Union[str, Any] , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : int , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : Optional[int] , UpperCAmelCase_ : Optional[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = self.num_labels SCREAMING_SNAKE_CASE__ = NystromformerForTokenClassification(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def A_ ( self : List[str] , UpperCAmelCase_ : List[Any] , UpperCAmelCase_ : str , UpperCAmelCase_ : List[str] , UpperCAmelCase_ : str , UpperCAmelCase_ : Any , UpperCAmelCase_ : Tuple , UpperCAmelCase_ : Optional[int] ): SCREAMING_SNAKE_CASE__ = self.num_choices SCREAMING_SNAKE_CASE__ = NystromformerForMultipleChoice(config=UpperCAmelCase_ ) model.to(UpperCAmelCase_ ) model.eval() SCREAMING_SNAKE_CASE__ = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() SCREAMING_SNAKE_CASE__ = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() SCREAMING_SNAKE_CASE__ = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() SCREAMING_SNAKE_CASE__ = model( UpperCAmelCase_ , attention_mask=UpperCAmelCase_ , token_type_ids=UpperCAmelCase_ , labels=UpperCAmelCase_ , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def A_ ( self : Any ): SCREAMING_SNAKE_CASE__ = self.prepare_config_and_inputs() ( ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ( SCREAMING_SNAKE_CASE__ ) , ) = config_and_inputs SCREAMING_SNAKE_CASE__ = {'input_ids': input_ids, 'token_type_ids': token_type_ids, 'attention_mask': input_mask} return config, inputs_dict @require_torch class lowercase__ ( _UpperCAmelCase , _UpperCAmelCase , unittest.TestCase ): A__ : Optional[Any] =( ( NystromformerModel, NystromformerForMaskedLM, NystromformerForMultipleChoice, NystromformerForQuestionAnswering, NystromformerForSequenceClassification, NystromformerForTokenClassification, ) if is_torch_available() else () ) A__ : str =( { """feature-extraction""": NystromformerModel, """fill-mask""": NystromformerForMaskedLM, """question-answering""": NystromformerForQuestionAnswering, """text-classification""": NystromformerForSequenceClassification, """token-classification""": NystromformerForTokenClassification, """zero-shot""": NystromformerForSequenceClassification, } if is_torch_available() else {} ) A__ : Union[str, Any] =False A__ : str =False def A_ ( self : Dict ): SCREAMING_SNAKE_CASE__ = NystromformerModelTester(self ) SCREAMING_SNAKE_CASE__ = ConfigTester(self , config_class=UpperCAmelCase_ , hidden_size=37 ) def A_ ( self : Any ): self.config_tester.run_common_tests() def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def A_ ( self : str ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: SCREAMING_SNAKE_CASE__ = type self.model_tester.create_and_check_model(*UpperCAmelCase_ ) def A_ ( self : Any ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*UpperCAmelCase_ ) def A_ ( self : Optional[int] ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*UpperCAmelCase_ ) def A_ ( self : int ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*UpperCAmelCase_ ) def A_ ( self : Union[str, Any] ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*UpperCAmelCase_ ) def A_ ( self : Tuple ): SCREAMING_SNAKE_CASE__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*UpperCAmelCase_ ) @slow def A_ ( self : str ): for model_name in NYSTROMFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: SCREAMING_SNAKE_CASE__ = NystromformerModel.from_pretrained(UpperCAmelCase_ ) self.assertIsNotNone(UpperCAmelCase_ ) @require_torch class lowercase__ ( unittest.TestCase ): @slow def A_ ( self : Optional[Any] ): SCREAMING_SNAKE_CASE__ = NystromformerModel.from_pretrained('uw-madison/nystromformer-512' ) SCREAMING_SNAKE_CASE__ = torch.tensor([[0, 1, 2, 3, 4, 5]] ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(UpperCAmelCase_ )[0] SCREAMING_SNAKE_CASE__ = torch.Size((1, 6, 768) ) self.assertEqual(output.shape , UpperCAmelCase_ ) SCREAMING_SNAKE_CASE__ = torch.tensor( [[[-0.4_532, -0.0_936, 0.5_137], [-0.2_676, 0.0_628, 0.6_186], [-0.3_629, -0.1_726, 0.4_716]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , UpperCAmelCase_ , atol=1e-4 ) ) @slow def A_ ( self : Any ): SCREAMING_SNAKE_CASE__ = 'the [MASK] of Belgium is Brussels' SCREAMING_SNAKE_CASE__ = AutoTokenizer.from_pretrained('uw-madison/nystromformer-512' ) SCREAMING_SNAKE_CASE__ = NystromformerForMaskedLM.from_pretrained('uw-madison/nystromformer-512' ) SCREAMING_SNAKE_CASE__ = tokenizer(UpperCAmelCase_ , return_tensors='pt' ) with torch.no_grad(): SCREAMING_SNAKE_CASE__ = model(encoding.input_ids ).logits SCREAMING_SNAKE_CASE__ = token_logits[:, 2, :].argmax(-1 )[0] self.assertEqual(tokenizer.decode(UpperCAmelCase_ ) , 'capital' )
356
import json import os import tempfile import transformers import datasets from utils import generate_example_dataset, get_duration __snake_case = 50_00_00 __snake_case ,__snake_case = os.path.split(__file__) __snake_case = os.path.join(RESULTS_BASEPATH, """results""", RESULTS_FILENAME.replace(""".py""", """.json""")) @get_duration def _lowercase ( UpperCamelCase_ , **UpperCamelCase_ ) -> Union[str, Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = dataset.map(**UpperCamelCase_ ) @get_duration def _lowercase ( UpperCamelCase_ , **UpperCamelCase_ ) -> List[Any]: '''simple docstring''' SCREAMING_SNAKE_CASE__ = dataset.filter(**UpperCamelCase_ ) def _lowercase ( ) -> Dict: '''simple docstring''' SCREAMING_SNAKE_CASE__ = {'num examples': SPEED_TEST_N_EXAMPLES} with tempfile.TemporaryDirectory() as tmp_dir: SCREAMING_SNAKE_CASE__ = datasets.Features({'text': datasets.Value('string' ), 'numbers': datasets.Value('float32' )} ) SCREAMING_SNAKE_CASE__ = generate_example_dataset( os.path.join(UpperCamelCase_ , 'dataset.arrow' ) , UpperCamelCase_ , num_examples=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = transformers.AutoTokenizer.from_pretrained('bert-base-cased' , use_fast=UpperCamelCase_ ) def tokenize(UpperCamelCase_ ): return tokenizer(examples['text'] ) SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , batched=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=lambda UpperCamelCase_ : None , batched=UpperCamelCase_ ) with dataset.formatted_as(type='numpy' ): SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=lambda UpperCamelCase_ : None , batched=UpperCamelCase_ ) with dataset.formatted_as(type='pandas' ): SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=lambda UpperCamelCase_ : None , batched=UpperCamelCase_ ) with dataset.formatted_as(type='torch' , columns='numbers' ): SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=lambda UpperCamelCase_ : None , batched=UpperCamelCase_ ) with dataset.formatted_as(type='tensorflow' , columns='numbers' ): SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=lambda UpperCamelCase_ : None , batched=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = map(UpperCamelCase_ , function=UpperCamelCase_ , batched=UpperCamelCase_ ) SCREAMING_SNAKE_CASE__ = filter(UpperCamelCase_ ) # Activate later when tokenizer support batched inputs # with dataset.formatted_as(type='numpy'): # times[func.__name__ + " fast-tokenizer batched numpy"] = func(dataset, function=tokenize, batched=True) with open(UpperCamelCase_ , 'wb' ) as f: f.write(json.dumps(UpperCamelCase_ ).encode('utf-8' ) ) if __name__ == "__main__": # useful to run the profiler benchmark_map_filter()
169
0
from __future__ import annotations import math def __snake_case ( __UpperCamelCase : int ,__UpperCamelCase : int ,__UpperCamelCase : bool ,__UpperCamelCase : list[int] ,__UpperCamelCase : float ): """simple docstring""" if depth < 0: raise ValueError("Depth cannot be less than 0" ) if len(__UpperCamelCase ) == 0: raise ValueError("Scores cannot be empty" ) if depth == height: return scores[node_index] if is_max: return max( minimax(depth + 1 ,node_index * 2 ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) ,minimax(depth + 1 ,node_index * 2 + 1 ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) ,) return min( minimax(depth + 1 ,node_index * 2 ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) ,minimax(depth + 1 ,node_index * 2 + 1 ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) ,) def __snake_case ( ): """simple docstring""" A_ = [90, 23, 6, 33, 21, 65, 123, 3_4423] A_ = math.log(len(__UpperCamelCase ) ,2 ) print("Optimal value : " ,end="" ) print(minimax(0 ,0 ,__UpperCamelCase ,__UpperCamelCase ,__UpperCamelCase ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
312
def __snake_case ( __UpperCamelCase : int = 1000 ): """simple docstring""" return sum(2 * a * ((a - 1) // 2) for a in range(3 ,n + 1 ) ) if __name__ == "__main__": print(solution())
312
1
import os from tempfile import TemporaryDirectory from unittest import TestCase import pytest from absl.testing import parameterized from datasets import config from datasets.arrow_reader import HF_GCP_BASE_URL from datasets.builder import DatasetBuilder from datasets.dataset_dict import IterableDatasetDict from datasets.iterable_dataset import IterableDataset from datasets.load import dataset_module_factory, import_main_class from datasets.utils.file_utils import cached_path lowerCAmelCase_ = [ {"""dataset""": """wikipedia""", """config_name""": """20220301.de"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.en"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.fr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.frr"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.it"""}, {"""dataset""": """wikipedia""", """config_name""": """20220301.simple"""}, {"""dataset""": """snli""", """config_name""": """plain_text"""}, {"""dataset""": """eli5""", """config_name""": """LFQA_reddit"""}, {"""dataset""": """wiki40b""", """config_name""": """en"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.compressed"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.nq.no_index"""}, {"""dataset""": """wiki_dpr""", """config_name""": """psgs_w100.multiset.no_index"""}, {"""dataset""": """natural_questions""", """config_name""": """default"""}, ] def lowerCamelCase_ ( lowerCAmelCase: Dict=True )-> List[Any]: if with_config: return [ { "testcase_name": d["dataset"] + "/" + d["config_name"], "dataset": d["dataset"], "config_name": d["config_name"], } for d in DATASETS_ON_HF_GCP ] else: return [ {"testcase_name": dataset, "dataset": dataset} for dataset in {d["dataset"] for d in DATASETS_ON_HF_GCP} ] @parameterized.named_parameters(list_datasets_on_hf_gcp_parameters(with_config=UpperCAmelCase_ ) ) class _lowerCAmelCase ( UpperCAmelCase_ ): '''simple docstring''' a_ : Any =None a_ : Any =None def UpperCamelCase_ ( self : List[Any] , UpperCamelCase : str , UpperCamelCase : Optional[Any] ): '''simple docstring''' with TemporaryDirectory() as tmp_dir: _snake_case : int = dataset_module_factory(UpperCamelCase , cache_dir=UpperCamelCase ) _snake_case : Optional[Any] = import_main_class(dataset_module.module_path , dataset=UpperCamelCase ) _snake_case : DatasetBuilder = builder_cls( cache_dir=UpperCamelCase , config_name=UpperCamelCase , hash=dataset_module.hash , ) _snake_case : Tuple = '/'.join( [ HF_GCP_BASE_URL, builder_instance._relative_data_dir(with_hash=UpperCamelCase ).replace(os.sep , '/' ), config.DATASET_INFO_FILENAME, ] ) _snake_case : List[str] = cached_path(UpperCamelCase , cache_dir=UpperCamelCase ) self.assertTrue(os.path.exists(UpperCamelCase ) ) @pytest.mark.integration def lowerCamelCase_ ( lowerCAmelCase: List[Any] )-> str: _snake_case : int = tmp_path_factory.mktemp('test_hf_gcp' ) / 'test_wikipedia_simple' _snake_case : Tuple = dataset_module_factory('wikipedia' , cache_dir=lowerCAmelCase ) _snake_case : str = import_main_class(dataset_module.module_path ) _snake_case : DatasetBuilder = builder_cls( cache_dir=lowerCAmelCase , config_name='20220301.frr' , hash=dataset_module.hash , ) # use the HF cloud storage, not the original download_and_prepare that uses apache-beam _snake_case : Dict = None builder_instance.download_and_prepare() _snake_case : List[str] = builder_instance.as_dataset() assert ds @pytest.mark.integration def lowerCamelCase_ ( lowerCAmelCase: Optional[int] )-> Optional[Any]: _snake_case : Optional[int] = dataset_module_factory('wikipedia' , cache_dir=lowerCAmelCase ) _snake_case : Any = import_main_class(dataset_module.module_path , dataset=lowerCAmelCase ) _snake_case : DatasetBuilder = builder_cls( cache_dir=lowerCAmelCase , config_name='20220301.frr' , hash=dataset_module.hash , ) _snake_case : Optional[int] = builder_instance.as_streaming_dataset() assert ds assert isinstance(lowerCAmelCase , lowerCAmelCase ) assert "train" in ds assert isinstance(ds['train'] , lowerCAmelCase ) assert next(iter(ds['train'] ) )
362
from __future__ import annotations lowerCAmelCase_ = """#""" class _lowerCAmelCase : '''simple docstring''' def __init__( self : Any ): '''simple docstring''' _snake_case : dict = {} def UpperCamelCase_ ( self : Optional[int] , UpperCamelCase : str ): '''simple docstring''' _snake_case : List[Any] = self._trie for char in text: if char not in trie: _snake_case : int = {} _snake_case : int = trie[char] _snake_case : Optional[Any] = True def UpperCamelCase_ ( self : Union[str, Any] , UpperCamelCase : str ): '''simple docstring''' _snake_case : Optional[int] = self._trie for char in prefix: if char in trie: _snake_case : Optional[Any] = trie[char] else: return [] return self._elements(UpperCamelCase ) def UpperCamelCase_ ( self : List[Any] , UpperCamelCase : dict ): '''simple docstring''' _snake_case : int = [] for c, v in d.items(): _snake_case : Dict = [' '] if c == END else [(c + s) for s in self._elements(UpperCamelCase )] result.extend(UpperCamelCase ) return tuple(UpperCamelCase ) lowerCAmelCase_ = Trie() lowerCAmelCase_ = ("""depart""", """detergent""", """daring""", """dog""", """deer""", """deal""") for word in words: trie.insert_word(word) def lowerCamelCase_ ( lowerCAmelCase: str )-> tuple: _snake_case : List[Any] = trie.find_word(lowerCAmelCase ) return tuple(string + word for word in suffixes ) def lowerCamelCase_ ( )-> None: print(autocomplete_using_trie('de' ) ) if __name__ == "__main__": import doctest doctest.testmod() main()
260
0