code
stringlengths
86
54.5k
code_codestyle
int64
0
371
style_context
stringlengths
87
49.2k
style_context_codestyle
int64
0
349
label
int64
0
1
"""simple docstring""" import json import re from typing import TYPE_CHECKING, List, Optional, Tuple, Union import numpy as np from ...utils import is_tf_available, is_torch_available, logging if TYPE_CHECKING: if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf from tokenizers import pre_tokenizers from ...tokenization_utils_base import BatchEncoding from ...tokenization_utils_fast import PreTrainedTokenizerFast from .tokenization_codegen import CodeGenTokenizer A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''vocab.json''', '''merges_file''': '''merges.txt''', '''tokenizer_file''': '''tokenizer.json'''} A_ = { '''vocab_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/vocab.json''', }, '''merges_file''': { '''Salesforce/codegen-350M-mono''': '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/merges.txt''', }, '''tokenizer_file''': { '''Salesforce/codegen-350M-mono''': ( '''https://huggingface.co/Salesforce/codegen-350M-mono/resolve/main/tokenizer.json''' ), }, } A_ = { '''Salesforce/codegen-350M-mono''': 2048, } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = ['input_ids', 'attention_mask'] snake_case_ = CodeGenTokenizer def __init__( self : List[str] , snake_case : Optional[int]=None , snake_case : List[Any]=None , snake_case : Union[str, Any]=None , snake_case : Tuple="<|endoftext|>" , snake_case : Optional[int]="<|endoftext|>" , snake_case : int="<|endoftext|>" , snake_case : Dict=False , **snake_case : int , ): '''simple docstring''' super().__init__( snake_case , snake_case , tokenizer_file=snake_case , unk_token=snake_case , bos_token=snake_case , eos_token=snake_case , add_prefix_space=snake_case , **snake_case , ) if kwargs.pop("""add_bos_token""" , snake_case ): A__ : List[Any] = kwargs.pop("""name_or_path""" , """""" ) raise ValueError( """Currenty GPT2's fast tokenizer does NOT support adding a BOS token.""" """Instead you should use GPT2's slow tokenizer class `CodeGenTokenizer` as follows: \n""" F'`CodeGenTokenizer.from_pretrained(\'{model_id}\')`\nor\n' F'`AutoTokenizer.from_pretrained(\'{model_id}\', use_fast=False)`\n' """This issue will be fixed soon, see: https://github.com/huggingface/tokenizers/pull/1005.""" """ so that the fast tokenizer works correctly.""" ) A__ : Optional[Any] = json.loads(self.backend_tokenizer.pre_tokenizer.__getstate__() ) if pre_tok_state.get("""add_prefix_space""" , snake_case ) != add_prefix_space: A__ : int = getattr(snake_case , pre_tok_state.pop("""type""" ) ) A__ : Optional[Any] = add_prefix_space A__ : Any = pre_tok_class(**snake_case ) A__ : Dict = add_prefix_space def _UpperCamelCase ( self : int , *snake_case : Dict , **snake_case : Dict ): '''simple docstring''' A__ : Dict = kwargs.get("""is_split_into_words""" , snake_case ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._batch_encode_plus(*snake_case , **snake_case ) def _UpperCamelCase ( self : Dict , *snake_case : Optional[int] , **snake_case : Any ): '''simple docstring''' A__ : Any = kwargs.get("""is_split_into_words""" , snake_case ) assert self.add_prefix_space or not is_split_into_words, ( F'You need to instantiate {self.__class__.__name__} with add_prefix_space=True ' "to use it with pretokenized inputs." ) return super()._encode_plus(*snake_case , **snake_case ) def _UpperCamelCase ( self : Dict , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' A__ : List[str] = self._tokenizer.model.save(snake_case , name=snake_case ) return tuple(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Union[int, List[int], "np.ndarray", "torch.Tensor", "tf.Tensor"] , snake_case : bool = False , snake_case : bool = None , snake_case : Optional[List[str]] = None , **snake_case : int , ): '''simple docstring''' A__ : List[Any] = super().decode( token_ids=snake_case , skip_special_tokens=snake_case , clean_up_tokenization_spaces=snake_case , **snake_case , ) if truncate_before_pattern is not None and len(snake_case ) > 0: A__ : Optional[int] = self.truncate(snake_case , snake_case ) return decoded_text def _UpperCamelCase ( self : str , snake_case : Any , snake_case : Tuple ): '''simple docstring''' def find_re(snake_case : Any , snake_case : Tuple , snake_case : Optional[int] ): A__ : Tuple = pattern.search(snake_case , snake_case ) return m.start() if m else -1 A__ : Any = [re.compile(snake_case , re.MULTILINE ) for pattern in truncate_before_pattern] A__ : Dict = list(re.finditer("""^print""" , snake_case , re.MULTILINE ) ) if len(snake_case ) > 1: A__ : str = completion[: prints[1].start()] A__ : List[Any] = list(re.finditer("""^def""" , snake_case , re.MULTILINE ) ) if len(snake_case ) > 1: A__ : Union[str, Any] = completion[: defs[1].start()] A__ : List[str] = 0 A__ : Optional[int] = [ pos for pos in [find_re(snake_case , snake_case , snake_case ) for terminal in terminals] if pos != -1 ] if len(snake_case ) > 0: return completion[: min(snake_case )] else: return completion
296
"""simple docstring""" from collections import defaultdict from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_5_0_0_0_0_0 ) ->int: A__ : defaultdict = defaultdict(UpperCAmelCase__ ) A__ : Any = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, UpperCAmelCase__, 2 ): if gcd(UpperCAmelCase__, UpperCAmelCase__ ) > 1: continue A__ : str = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'{solution() = }')
296
1
"""simple docstring""" from argparse import ArgumentParser from ..pipelines import Pipeline, PipelineDataFormat, get_supported_tasks, pipeline from ..utils import logging from . import BaseTransformersCLICommand A_ = logging.get_logger(__name__) # pylint: disable=invalid-name def _lowerCAmelCase ( UpperCAmelCase__ : str ) ->List[str]: if not path: return "pipe" for ext in PipelineDataFormat.SUPPORTED_FORMATS: if path.endswith(UpperCAmelCase__ ): return ext raise Exception( f'Unable to determine file format from file extension {path}. ' f'Please provide the format through --format {PipelineDataFormat.SUPPORTED_FORMATS}' ) def _lowerCAmelCase ( UpperCAmelCase__ : Dict ) ->Optional[Any]: A__ : int = pipeline( task=args.task, model=args.model if args.model else None, config=args.config, tokenizer=args.tokenizer, device=args.device, ) A__ : int = try_infer_format_from_ext(args.input ) if args.format == """infer""" else args.format A__ : Dict = PipelineDataFormat.from_str( format=UpperCAmelCase__, output_path=args.output, input_path=args.input, column=args.column if args.column else nlp.default_input_names, overwrite=args.overwrite, ) return RunCommand(UpperCAmelCase__, UpperCAmelCase__ ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Any , snake_case : Pipeline , snake_case : PipelineDataFormat ): '''simple docstring''' A__ : Union[str, Any] = nlp A__ : Optional[int] = reader @staticmethod def _UpperCamelCase ( snake_case : ArgumentParser ): '''simple docstring''' A__ : Optional[Any] = parser.add_parser("""run""" , help="""Run a pipeline through the CLI""" ) run_parser.add_argument("""--task""" , choices=get_supported_tasks() , help="""Task to run""" ) run_parser.add_argument("""--input""" , type=snake_case , help="""Path to the file to use for inference""" ) run_parser.add_argument("""--output""" , type=snake_case , help="""Path to the file that will be used post to write results.""" ) run_parser.add_argument("""--model""" , type=snake_case , help="""Name or path to the model to instantiate.""" ) run_parser.add_argument("""--config""" , type=snake_case , help="""Name or path to the model's config to instantiate.""" ) run_parser.add_argument( """--tokenizer""" , type=snake_case , help="""Name of the tokenizer to use. (default: same as the model name)""" ) run_parser.add_argument( """--column""" , type=snake_case , help="""Name of the column to use as input. (For multi columns input as QA use column1,columns2)""" , ) run_parser.add_argument( """--format""" , type=snake_case , default="""infer""" , choices=PipelineDataFormat.SUPPORTED_FORMATS , help="""Input format to read from""" , ) run_parser.add_argument( """--device""" , type=snake_case , default=-1 , help="""Indicate the device to run onto, -1 indicates CPU, >= 0 indicates GPU (default: -1)""" , ) run_parser.add_argument("""--overwrite""" , action="""store_true""" , help="""Allow overwriting the output file.""" ) run_parser.set_defaults(func=snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ : List[str] = self._nlp, [] for entry in self._reader: A__ : Tuple = nlp(**snake_case ) if self._reader.is_multi_columns else nlp(snake_case ) if isinstance(snake_case , snake_case ): outputs.append(snake_case ) else: outputs += output # Saving data if self._nlp.binary_output: A__ : List[Any] = self._reader.save_binary(snake_case ) logger.warning(F'Current pipeline requires output to be in binary format, saving at {binary_path}' ) else: self._reader.save(snake_case )
296
"""simple docstring""" import os from distutils.util import strtobool def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any] ) ->List[str]: for e in env_keys: A__ : List[Any] = int(os.environ.get(UpperCAmelCase__, -1 ) ) if val >= 0: return val return default def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : str=False ) ->List[str]: A__ : List[Any] = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return strtobool(UpperCAmelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]="no" ) ->int: A__ : str = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return value
296
1
"""simple docstring""" import absl # noqa: F401 # Here to have a nice missing dependency error message early on import nltk # noqa: F401 # Here to have a nice missing dependency error message early on import numpy # noqa: F401 # Here to have a nice missing dependency error message early on import six # noqa: F401 # Here to have a nice missing dependency error message early on from rouge_score import rouge_scorer, scoring import datasets A_ = '''\ @inproceedings{lin-2004-rouge, title = "{ROUGE}: A Package for Automatic Evaluation of Summaries", author = "Lin, Chin-Yew", booktitle = "Text Summarization Branches Out", month = jul, year = "2004", address = "Barcelona, Spain", publisher = "Association for Computational Linguistics", url = "https://www.aclweb.org/anthology/W04-1013", pages = "74--81", } ''' A_ = '''\ ROUGE, or Recall-Oriented Understudy for Gisting Evaluation, is a set of metrics and a software package used for evaluating automatic summarization and machine translation software in natural language processing. The metrics compare an automatically produced summary or translation against a reference or a set of references (human-produced) summary or translation. Note that ROUGE is case insensitive, meaning that upper case letters are treated the same way as lower case letters. This metrics is a wrapper around Google Research reimplementation of ROUGE: https://github.com/google-research/google-research/tree/master/rouge ''' A_ = ''' Calculates average rouge scores for a list of hypotheses and references Args: predictions: list of predictions to score. Each prediction should be a string with tokens separated by spaces. references: list of reference for each prediction. Each reference should be a string with tokens separated by spaces. rouge_types: A list of rouge types to calculate. Valid names: `"rouge{n}"` (e.g. `"rouge1"`, `"rouge2"`) where: {n} is the n-gram based scoring, `"rougeL"`: Longest common subsequence based scoring. `"rougeLSum"`: rougeLsum splits text using `"\n"`. See details in https://github.com/huggingface/datasets/issues/617 use_stemmer: Bool indicating whether Porter stemmer should be used to strip word suffixes. use_aggregator: Return aggregates if this is set to True Returns: rouge1: rouge_1 (precision, recall, f1), rouge2: rouge_2 (precision, recall, f1), rougeL: rouge_l (precision, recall, f1), rougeLsum: rouge_lsum (precision, recall, f1) Examples: >>> rouge = datasets.load_metric(\'rouge\') >>> predictions = ["hello there", "general kenobi"] >>> references = ["hello there", "general kenobi"] >>> results = rouge.compute(predictions=predictions, references=references) >>> print(list(results.keys())) [\'rouge1\', \'rouge2\', \'rougeL\', \'rougeLsum\'] >>> print(results["rouge1"]) AggregateScore(low=Score(precision=1.0, recall=1.0, fmeasure=1.0), mid=Score(precision=1.0, recall=1.0, fmeasure=1.0), high=Score(precision=1.0, recall=1.0, fmeasure=1.0)) >>> print(results["rouge1"].mid.fmeasure) 1.0 ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { """predictions""": datasets.Value("""string""" , id="""sequence""" ), """references""": datasets.Value("""string""" , id="""sequence""" ), } ) , codebase_urls=["""https://github.com/google-research/google-research/tree/master/rouge"""] , reference_urls=[ """https://en.wikipedia.org/wiki/ROUGE_(metric)""", """https://github.com/google-research/google-research/tree/master/rouge""", ] , ) def _UpperCamelCase ( self : str , snake_case : Optional[Any] , snake_case : List[Any] , snake_case : Union[str, Any]=None , snake_case : Dict=True , snake_case : List[str]=False ): '''simple docstring''' if rouge_types is None: A__ : int = ["""rouge1""", """rouge2""", """rougeL""", """rougeLsum"""] A__ : List[str] = rouge_scorer.RougeScorer(rouge_types=snake_case , use_stemmer=snake_case ) if use_aggregator: A__ : Any = scoring.BootstrapAggregator() else: A__ : str = [] for ref, pred in zip(snake_case , snake_case ): A__ : Tuple = scorer.score(snake_case , snake_case ) if use_aggregator: aggregator.add_scores(snake_case ) else: scores.append(snake_case ) if use_aggregator: A__ : Optional[int] = aggregator.aggregate() else: A__ : Union[str, Any] = {} for key in scores[0]: A__ : int = [score[key] for score in scores] return result
296
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''andreasmadsen/efficient_mlm_m0.40''': ( '''https://huggingface.co/andreasmadsen/efficient_mlm_m0.40/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'roberta-prelayernorm' def __init__( self : Optional[Any] , snake_case : Optional[Any]=5_0265 , snake_case : Union[str, Any]=768 , snake_case : Optional[int]=12 , snake_case : Optional[int]=12 , snake_case : Tuple=3072 , snake_case : str="gelu" , snake_case : Optional[int]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : int=2 , snake_case : Tuple=0.02 , snake_case : Optional[Any]=1e-12 , snake_case : List[str]=1 , snake_case : Tuple=0 , snake_case : Optional[int]=2 , snake_case : List[str]="absolute" , snake_case : Union[str, Any]=True , snake_case : Dict=None , **snake_case : Union[str, Any] , ): '''simple docstring''' super().__init__(pad_token_id=snake_case , bos_token_id=snake_case , eos_token_id=snake_case , **snake_case ) A__ : Dict = vocab_size A__ : Optional[int] = hidden_size A__ : Union[str, Any] = num_hidden_layers A__ : int = num_attention_heads A__ : Union[str, Any] = hidden_act A__ : str = intermediate_size A__ : Union[str, Any] = hidden_dropout_prob A__ : int = attention_probs_dropout_prob A__ : Optional[Any] = max_position_embeddings A__ : int = type_vocab_size A__ : Dict = initializer_range A__ : List[str] = layer_norm_eps A__ : Tuple = position_embedding_type A__ : str = use_cache A__ : int = classifier_dropout class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' if self.task == "multiple-choice": A__ : List[str] = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A__ : List[Any] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
296
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = logging.get_logger(__name__) A_ = Dict[str, Any] A_ = List[Prediction] @add_end_docstrings(UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : str , *snake_case : Tuple , **snake_case : Tuple ): '''simple docstring''' super().__init__(*snake_case , **snake_case ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _UpperCamelCase ( self : List[Any] , **snake_case : Optional[int] ): '''simple docstring''' A__ : Dict = {} if "threshold" in kwargs: A__ : int = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return super().__call__(*snake_case , **snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : List[str] = load_image(snake_case ) A__ : int = torch.IntTensor([[image.height, image.width]] ) A__ : Union[str, Any] = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: A__ : str = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) A__ : List[str] = target_size return inputs def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : str = model_inputs.pop("""target_size""" ) A__ : Dict = self.model(**snake_case ) A__ : Optional[Any] = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: A__ : str = model_inputs["""bbox"""] return model_outputs def _UpperCamelCase ( self : Tuple , snake_case : Optional[int] , snake_case : int=0.9 ): '''simple docstring''' A__ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A__ , A__ : Tuple = target_size[0].tolist() def unnormalize(snake_case : Optional[int] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) A__ , A__ : Optional[int] = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A__ : Optional[Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A__ : List[str] = [unnormalize(snake_case ) for bbox in model_outputs["""bbox"""].squeeze(0 )] A__ : Tuple = ["""score""", """label""", """box"""] A__ : Any = [dict(zip(snake_case , snake_case ) ) for vals in zip(scores.tolist() , snake_case , snake_case ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A__ : Union[str, Any] = self.image_processor.post_process_object_detection(snake_case , snake_case , snake_case ) A__ : str = raw_annotations[0] A__ : str = raw_annotation["""scores"""] A__ : List[Any] = raw_annotation["""labels"""] A__ : int = raw_annotation["""boxes"""] A__ : str = scores.tolist() A__ : Any = [self.model.config.idalabel[label.item()] for label in labels] A__ : int = [self._get_bounding_box(snake_case ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A__ : str = ["""score""", """label""", """box"""] A__ : Dict = [ dict(zip(snake_case , snake_case ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def _UpperCamelCase ( self : Union[str, Any] , snake_case : "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) A__ , A__ , A__ , A__ : Any = box.int().tolist() A__ : Any = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
296
1
"""simple docstring""" # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import argparse from .config import config_command_parser from .config_args import default_config_file, load_config_from_file # noqa: F401 from .default import default_command_parser from .update import update_command_parser def _lowerCAmelCase ( UpperCAmelCase__ : List[str]=None ) ->Optional[Any]: A__ : str = argparse.ArgumentParser(add_help=UpperCAmelCase__, allow_abbrev=UpperCAmelCase__ ) # The main config parser A__ : Dict = config_command_parser(UpperCAmelCase__ ) # The subparser to add commands to A__ : Tuple = config_parser.add_subparsers(title="""subcommands""", dest="""subcommand""" ) # Then add other parsers with the parent parser default_command_parser(UpperCAmelCase__, parents=[parent_parser] ) update_command_parser(UpperCAmelCase__, parents=[parent_parser] ) return config_parser def _lowerCAmelCase ( ) ->int: A__ : int = get_config_parser() A__ : str = config_parser.parse_args() if not hasattr(UpperCAmelCase__, """func""" ): config_parser.print_help() exit(1 ) # Run args.func(UpperCAmelCase__ ) if __name__ == "__main__": main()
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
"""simple docstring""" from __future__ import annotations from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : int = 2, UpperCAmelCase__ : int = 1, UpperCAmelCase__ : int = 3, ) ->int | None: # A value less than 2 can cause an infinite loop in the algorithm. if num < 2: raise ValueError("""The input value cannot be less than 2""" ) # Because of the relationship between ``f(f(x))`` and ``f(x)``, this # algorithm struggles to find factors that are divisible by two. # As a workaround, we specifically check for two and even inputs. # See: https://math.stackexchange.com/a/2856214/165820 if num > 2 and num % 2 == 0: return 2 # Pollard's Rho algorithm requires a function that returns pseudorandom # values between 0 <= X < ``num``. It doesn't need to be random in the # sense that the output value is cryptographically secure or difficult # to calculate, it only needs to be random in the sense that all output # values should be equally likely to appear. # For this reason, Pollard suggested using ``f(x) = (x**2 - 1) % num`` # However, the success of Pollard's algorithm isn't guaranteed and is # determined in part by the initial seed and the chosen random function. # To make retries easier, we will instead use ``f(x) = (x**2 + C) % num`` # where ``C`` is a value that we can modify between each attempt. def rand_fn(UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : int ) -> int: return (pow(UpperCAmelCase__, 2 ) + step) % modulus for _ in range(UpperCAmelCase__ ): # These track the position within the cycle detection logic. A__ : Optional[Any] = seed A__ : int = seed while True: # At each iteration, the tortoise moves one step and the hare moves two. A__ : Union[str, Any] = rand_fn(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ : Union[str, Any] = rand_fn(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ : Optional[int] = rand_fn(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # At some point both the tortoise and the hare will enter a cycle whose # length ``p`` is a divisor of ``num``. Once in that cycle, at some point # the tortoise and hare will end up on the same value modulo ``p``. # We can detect when this happens because the position difference between # the tortoise and the hare will share a common divisor with ``num``. A__ : str = gcd(hare - tortoise, UpperCAmelCase__ ) if divisor == 1: # No common divisor yet, just keep searching. continue else: # We found a common divisor! if divisor == num: # Unfortunately, the divisor is ``num`` itself and is useless. break else: # The divisor is a nontrivial factor of ``num``! return divisor # If we made it here, then this attempt failed. # We need to pick a new starting seed for the tortoise and hare # in addition to a new step value for the random function. # To keep this example implementation deterministic, the # new values will be generated based on currently available # values instead of using something like ``random.randint``. # We can use the hare's position as the new seed. # This is actually what Richard Brent's the "optimized" variant does. A__ : List[Any] = hare # The new step value for the random function can just be incremented. # At first the results will be similar to what the old function would # have produced, but the value will quickly diverge after a bit. step += 1 # We haven't found a divisor within the requested number of attempts. # We were unlucky or ``num`` itself is actually prime. return None if __name__ == "__main__": import argparse A_ = argparse.ArgumentParser() parser.add_argument( '''num''', type=int, help='''The value to find a divisor of''', ) parser.add_argument( '''--attempts''', type=int, default=3, help='''The number of attempts before giving up''', ) A_ = parser.parse_args() A_ = pollard_rho(args.num, attempts=args.attempts) if divisor is None: print(F'{args.num} is probably prime') else: A_ = args.num // divisor print(F'{args.num} = {divisor} * {quotient}')
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_torch_available, ) A_ = { '''configuration_falcon''': ['''FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''FalconConfig'''], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''FALCON_PRETRAINED_MODEL_ARCHIVE_LIST''', '''FalconForCausalLM''', '''FalconModel''', '''FalconPreTrainedModel''', '''FalconForSequenceClassification''', '''FalconForTokenClassification''', '''FalconForQuestionAnswering''', ] if TYPE_CHECKING: from .configuration_falcon import FALCON_PRETRAINED_CONFIG_ARCHIVE_MAP, FalconConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_falcon import ( FALCON_PRETRAINED_MODEL_ARCHIVE_LIST, FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, FalconPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
1
"""simple docstring""" import gc import unittest import numpy as np import torch import torch.nn.functional as F from transformers import ( ClapTextConfig, ClapTextModelWithProjection, RobertaTokenizer, SpeechTaHifiGan, SpeechTaHifiGanConfig, ) from diffusers import ( AudioLDMPipeline, AutoencoderKL, DDIMScheduler, LMSDiscreteScheduler, PNDMScheduler, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_AUDIO_BATCH_PARAMS, TEXT_TO_AUDIO_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = AudioLDMPipeline snake_case_ = TEXT_TO_AUDIO_PARAMS snake_case_ = TEXT_TO_AUDIO_BATCH_PARAMS snake_case_ = frozenset( [ 'num_inference_steps', 'num_waveforms_per_prompt', 'generator', 'latents', 'output_type', 'return_dict', 'callback', 'callback_steps', ] ) def _UpperCamelCase ( self : Any ): '''simple docstring''' torch.manual_seed(0 ) A__ : List[str] = UNetaDConditionModel( block_out_channels=(32, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""DownBlock2D""", """CrossAttnDownBlock2D""") , up_block_types=("""CrossAttnUpBlock2D""", """UpBlock2D""") , cross_attention_dim=(32, 64) , class_embed_type="""simple_projection""" , projection_class_embeddings_input_dim=32 , class_embeddings_concat=snake_case , ) A__ : int = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case , set_alpha_to_one=snake_case , ) torch.manual_seed(0 ) A__ : int = AutoencoderKL( block_out_channels=[32, 64] , in_channels=1 , out_channels=1 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , ) torch.manual_seed(0 ) A__ : List[str] = ClapTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , projection_dim=32 , ) A__ : Dict = ClapTextModelWithProjection(snake_case ) A__ : Optional[int] = RobertaTokenizer.from_pretrained("""hf-internal-testing/tiny-random-roberta""" , model_max_length=77 ) A__ : str = SpeechTaHifiGanConfig( model_in_dim=8 , sampling_rate=1_6000 , upsample_initial_channel=16 , upsample_rates=[2, 2] , upsample_kernel_sizes=[4, 4] , resblock_kernel_sizes=[3, 7] , resblock_dilation_sizes=[[1, 3, 5], [1, 3, 5]] , normalize_before=snake_case , ) A__ : List[Any] = SpeechTaHifiGan(snake_case ) A__ : Union[str, Any] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, """vocoder""": vocoder, } return components def _UpperCamelCase ( self : int , snake_case : Union[str, Any] , snake_case : Any=0 ): '''simple docstring''' if str(snake_case ).startswith("""mps""" ): A__ : Optional[int] = torch.manual_seed(snake_case ) else: A__ : Optional[Any] = torch.Generator(device=snake_case ).manual_seed(snake_case ) A__ : Union[str, Any] = { """prompt""": """A hammer hitting a wooden surface""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, } return inputs def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Tuple = self.get_dummy_components() A__ : Tuple = AudioLDMPipeline(**snake_case ) A__ : Any = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : Dict = self.get_dummy_inputs(snake_case ) A__ : Dict = audioldm_pipe(**snake_case ) A__ : Union[str, Any] = output.audios[0] assert audio.ndim == 1 assert len(snake_case ) == 256 A__ : Union[str, Any] = audio[:10] A__ : Optional[Any] = np.array( [-0.0050, 0.0050, -0.0060, 0.0033, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0033] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.get_dummy_components() A__ : Union[str, Any] = AudioLDMPipeline(**snake_case ) A__ : Tuple = audioldm_pipe.to(snake_case ) A__ : Optional[Any] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : int = self.get_dummy_inputs(snake_case ) A__ : Tuple = 3 * [inputs["""prompt"""]] # forward A__ : Dict = audioldm_pipe(**snake_case ) A__ : List[str] = output.audios[0] A__ : Dict = self.get_dummy_inputs(snake_case ) A__ : Optional[int] = 3 * [inputs.pop("""prompt""" )] A__ : Tuple = audioldm_pipe.tokenizer( snake_case , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=snake_case , return_tensors="""pt""" , ) A__ : Union[str, Any] = text_inputs["""input_ids"""].to(snake_case ) A__ : int = audioldm_pipe.text_encoder( snake_case , ) A__ : List[Any] = prompt_embeds.text_embeds # additional L_2 normalization over each hidden-state A__ : Tuple = F.normalize(snake_case , dim=-1 ) A__ : Dict = prompt_embeds # forward A__ : List[Any] = audioldm_pipe(**snake_case ) A__ : Dict = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : int = self.get_dummy_components() A__ : List[str] = AudioLDMPipeline(**snake_case ) A__ : Tuple = audioldm_pipe.to(snake_case ) A__ : List[str] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : str = self.get_dummy_inputs(snake_case ) A__ : int = 3 * ["""this is a negative prompt"""] A__ : Union[str, Any] = negative_prompt A__ : Tuple = 3 * [inputs["""prompt"""]] # forward A__ : Union[str, Any] = audioldm_pipe(**snake_case ) A__ : Tuple = output.audios[0] A__ : Tuple = self.get_dummy_inputs(snake_case ) A__ : Optional[int] = 3 * [inputs.pop("""prompt""" )] A__ : int = [] for p in [prompt, negative_prompt]: A__ : Union[str, Any] = audioldm_pipe.tokenizer( snake_case , padding="""max_length""" , max_length=audioldm_pipe.tokenizer.model_max_length , truncation=snake_case , return_tensors="""pt""" , ) A__ : int = text_inputs["""input_ids"""].to(snake_case ) A__ : int = audioldm_pipe.text_encoder( snake_case , ) A__ : int = text_embeds.text_embeds # additional L_2 normalization over each hidden-state A__ : Union[str, Any] = F.normalize(snake_case , dim=-1 ) embeds.append(snake_case ) A__ , A__ : Any = embeds # forward A__ : int = audioldm_pipe(**snake_case ) A__ : str = output.audios[0] assert np.abs(audio_a - audio_a ).max() < 1e-2 def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : str = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Optional[Any] = self.get_dummy_components() A__ : List[Any] = PNDMScheduler(skip_prk_steps=snake_case ) A__ : int = AudioLDMPipeline(**snake_case ) A__ : str = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : Union[str, Any] = self.get_dummy_inputs(snake_case ) A__ : List[str] = """egg cracking""" A__ : Dict = audioldm_pipe(**snake_case , negative_prompt=snake_case ) A__ : Union[str, Any] = output.audios[0] assert audio.ndim == 1 assert len(snake_case ) == 256 A__ : Optional[int] = audio[:10] A__ : Tuple = np.array( [-0.0051, 0.0050, -0.0060, 0.0034, -0.0026, 0.0033, -0.0027, 0.0033, -0.0028, 0.0032] ) assert np.abs(audio_slice - expected_slice ).max() < 1e-2 def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Optional[int] = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : int = self.get_dummy_components() A__ : Optional[int] = PNDMScheduler(skip_prk_steps=snake_case ) A__ : Optional[int] = AudioLDMPipeline(**snake_case ) A__ : Optional[Any] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : str = """A hammer hitting a wooden surface""" # test num_waveforms_per_prompt=1 (default) A__ : str = audioldm_pipe(snake_case , num_inference_steps=2 ).audios assert audios.shape == (1, 256) # test num_waveforms_per_prompt=1 (default) for batch of prompts A__ : List[Any] = 2 A__ : Optional[Any] = audioldm_pipe([prompt] * batch_size , num_inference_steps=2 ).audios assert audios.shape == (batch_size, 256) # test num_waveforms_per_prompt for single prompt A__ : int = 2 A__ : Dict = audioldm_pipe(snake_case , num_inference_steps=2 , num_waveforms_per_prompt=snake_case ).audios assert audios.shape == (num_waveforms_per_prompt, 256) # test num_waveforms_per_prompt for batch of prompts A__ : Optional[Any] = 2 A__ : int = audioldm_pipe( [prompt] * batch_size , num_inference_steps=2 , num_waveforms_per_prompt=snake_case ).audios assert audios.shape == (batch_size * num_waveforms_per_prompt, 256) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Optional[Any] = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : int = self.get_dummy_components() A__ : List[Any] = AudioLDMPipeline(**snake_case ) A__ : Optional[Any] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : Union[str, Any] = audioldm_pipe.vocoder.config.sampling_rate A__ : Optional[Any] = self.get_dummy_inputs(snake_case ) A__ : int = audioldm_pipe(audio_length_in_s=0.016 , **snake_case ) A__ : Optional[int] = output.audios[0] assert audio.ndim == 1 assert len(snake_case ) / vocoder_sampling_rate == 0.016 A__ : List[str] = audioldm_pipe(audio_length_in_s=0.032 , **snake_case ) A__ : Union[str, Any] = output.audios[0] assert audio.ndim == 1 assert len(snake_case ) / vocoder_sampling_rate == 0.032 def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : List[str] = self.get_dummy_components() A__ : Tuple = AudioLDMPipeline(**snake_case ) A__ : List[str] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : Union[str, Any] = ["""hey"""] A__ : Optional[int] = audioldm_pipe(snake_case , num_inference_steps=1 ) A__ : Union[str, Any] = output.audios.shape assert audio_shape == (1, 256) A__ : int = audioldm_pipe.vocoder.config config.model_in_dim *= 2 A__ : Optional[Any] = SpeechTaHifiGan(snake_case ).to(snake_case ) A__ : str = audioldm_pipe(snake_case , num_inference_steps=1 ) A__ : Any = output.audios.shape # waveform shape is unchanged, we just have 2x the number of mel channels in the spectrogram assert audio_shape == (1, 256) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' self._test_inference_batch_single_identical(test_mean_pixel_difference=snake_case ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case ) @slow class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' super().tearDown() gc.collect() torch.cuda.empty_cache() def _UpperCamelCase ( self : Optional[Any] , snake_case : List[str] , snake_case : Tuple="cpu" , snake_case : Optional[int]=torch.floataa , snake_case : Optional[Any]=0 ): '''simple docstring''' A__ : Union[str, Any] = torch.Generator(device=snake_case ).manual_seed(snake_case ) A__ : List[str] = np.random.RandomState(snake_case ).standard_normal((1, 8, 128, 16) ) A__ : List[Any] = torch.from_numpy(snake_case ).to(device=snake_case , dtype=snake_case ) A__ : Dict = { """prompt""": """A hammer hitting a wooden surface""", """latents""": latents, """generator""": generator, """num_inference_steps""": 3, """guidance_scale""": 2.5, } return inputs def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Union[str, Any] = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) A__ : Any = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : Optional[int] = self.get_inputs(snake_case ) A__ : Tuple = 25 A__ : Optional[int] = audioldm_pipe(**snake_case ).audios[0] assert audio.ndim == 1 assert len(snake_case ) == 8_1920 A__ : Union[str, Any] = audio[7_7230:7_7240] A__ : List[Any] = np.array( [-0.4884, -0.4607, 0.0023, 0.5007, 0.5896, 0.5151, 0.3813, -0.0208, -0.3687, -0.4315] ) A__ : int = np.abs(expected_slice - audio_slice ).max() assert max_diff < 1e-2 def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : List[Any] = AudioLDMPipeline.from_pretrained("""cvssp/audioldm""" ) A__ : str = LMSDiscreteScheduler.from_config(audioldm_pipe.scheduler.config ) A__ : Optional[int] = audioldm_pipe.to(snake_case ) audioldm_pipe.set_progress_bar_config(disable=snake_case ) A__ : List[Any] = self.get_inputs(snake_case ) A__ : Optional[int] = audioldm_pipe(**snake_case ).audios[0] assert audio.ndim == 1 assert len(snake_case ) == 8_1920 A__ : Any = audio[2_7780:2_7790] A__ : Dict = np.array([-0.2131, -0.0873, -0.0124, -0.0189, 0.0569, 0.1373, 0.1883, 0.2886, 0.3297, 0.2212] ) A__ : List[Any] = np.abs(expected_slice - audio_slice ).max() assert max_diff < 3e-2
296
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[Any]=13 , snake_case : Union[str, Any]=7 , snake_case : Optional[Any]=True , snake_case : str=True , snake_case : Dict=False , snake_case : Union[str, Any]=True , snake_case : Optional[Any]=99 , snake_case : str=32 , snake_case : Tuple=5 , snake_case : List[str]=4 , snake_case : Optional[int]=37 , snake_case : str="gelu" , snake_case : Tuple=0.1 , snake_case : Optional[int]=0.1 , snake_case : int=512 , snake_case : List[str]=16 , snake_case : str=2 , snake_case : Optional[int]=0.02 , snake_case : str=3 , snake_case : Dict=4 , snake_case : Optional[Any]=None , ): '''simple docstring''' A__ : int = parent A__ : Union[str, Any] = batch_size A__ : Optional[int] = seq_length A__ : List[Any] = is_training A__ : List[str] = use_input_mask A__ : Optional[Any] = use_token_type_ids A__ : List[Any] = use_labels A__ : Union[str, Any] = vocab_size A__ : List[Any] = hidden_size A__ : Any = num_hidden_layers A__ : Any = num_attention_heads A__ : Optional[int] = intermediate_size A__ : Any = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Dict = attention_probs_dropout_prob A__ : Optional[int] = max_position_embeddings A__ : Tuple = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[str] = initializer_range A__ : Any = num_labels A__ : Any = num_choices A__ : int = scope def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = None if self.use_input_mask: A__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_token_type_ids: A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : int = None A__ : int = None A__ : List[str] = None if self.use_labels: A__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) A__ : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Any , snake_case : Dict , snake_case : Any , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case ) A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Optional[int] , snake_case : List[str] , snake_case : str , snake_case : Optional[Any] , snake_case : List[str] , snake_case : List[Any] , snake_case : Tuple , snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = BioGptForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Any , snake_case : str , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Any , *snake_case : Dict ): '''simple docstring''' A__ : Union[str, Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() # create attention mask A__ : List[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) A__ : Any = self.seq_length // 2 A__ : str = 0 # first forward pass A__ , A__ : List[Any] = model(snake_case , attention_mask=snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids A__ : int = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A__ : List[str] = ids_tensor((1,) , snake_case ).item() + 1 A__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A__ : int = random_other_next_tokens # append to next input_ids and attn_mask A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case )] , dim=1 , ) # get two different outputs A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Optional[int] = model(snake_case , past_key_values=snake_case , attention_mask=snake_case )["""last_hidden_state"""] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() A__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str , snake_case : int , snake_case : Optional[Any] , *snake_case : str ): '''simple docstring''' A__ : Dict = BioGptModel(config=snake_case ).to(snake_case ).eval() A__ : Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) # first forward pass A__ : Dict = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ , A__ : List[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Optional[int] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A__ : Any = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , past_key_values=snake_case )[ """last_hidden_state""" ] # select random slice A__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Tuple , *snake_case : Union[str, Any] , snake_case : Union[str, Any]=False ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM(snake_case ) model.to(snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() A__ : Optional[Any] = model(snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self : int , snake_case : Optional[Any] , *snake_case : Optional[int] ): '''simple docstring''' A__ : int = BioGptModel(snake_case ) A__ : Union[str, Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _UpperCamelCase ( self : Any , snake_case : Dict , snake_case : Tuple , snake_case : int , snake_case : Union[str, Any] , snake_case : Dict , *snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = self.num_labels A__ : int = BioGptForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : str = config_and_inputs A__ : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case_ = (BioGptForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[str] = BioGptModelTester(self ) A__ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : str = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case , gradient_checkpointing=snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) A__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = """left""" # Define PAD Token = EOS Token = 50256 A__ : Optional[int] = tokenizer.eos_token A__ : Dict = model.config.eos_token_id # use different length sentences to test batching A__ : Union[str, Any] = [ """Hello, my dog is a little""", """Today, I""", ] A__ : List[str] = tokenizer(snake_case , return_tensors="""pt""" , padding=snake_case ) A__ : str = inputs["""input_ids"""].to(snake_case ) A__ : Dict = model.generate( input_ids=snake_case , attention_mask=inputs["""attention_mask"""].to(snake_case ) , ) A__ : Optional[int] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Any = model.generate(input_ids=snake_case ) A__ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A__ : str = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Dict = model.generate(input_ids=snake_case , max_length=model.config.max_length - num_paddings ) A__ : Optional[Any] = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case ) A__ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case ) A__ : Optional[int] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(snake_case , [non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = BioGptModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(snake_case ) A__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Union[str, Any] = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Any = 3 A__ : List[Any] = """multi_label_classification""" A__ : Dict = input_dict["""input_ids"""] A__ : Tuple = input_ids.ne(1 ).to(snake_case ) A__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Tuple = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A__ : str = torch.tensor([[2, 4805, 9, 656, 21]] ) A__ : Dict = model(snake_case )[0] A__ : Tuple = 4_2384 A__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : str = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) torch.manual_seed(0 ) A__ : Tuple = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(snake_case ) A__ : Optional[int] = model.generate( **snake_case , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case , ) A__ : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case ) A__ : List[str] = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" import argparse import math import traceback import dateutil.parser as date_parser import requests def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->str: A__ : str = {} A__ : List[Any] = job["""started_at"""] A__ : List[Any] = job["""completed_at"""] A__ : Union[str, Any] = date_parser.parse(UpperCAmelCase__ ) A__ : Optional[int] = date_parser.parse(UpperCAmelCase__ ) A__ : List[Any] = round((end_datetime - start_datetime).total_seconds() / 60.0 ) A__ : List[Any] = start A__ : Optional[Any] = end A__ : Tuple = duration_in_min return job_info def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[str]=None ) ->List[Any]: A__ : Any = None if token is not None: A__ : int = {"""Accept""": """application/vnd.github+json""", """Authorization""": f'Bearer {token}'} A__ : Optional[Any] = f'https://api.github.com/repos/huggingface/transformers/actions/runs/{workflow_run_id}/jobs?per_page=100' A__ : Optional[int] = requests.get(UpperCAmelCase__, headers=UpperCAmelCase__ ).json() A__ : Optional[int] = {} try: job_time.update({job["""name"""]: extract_time_from_single_job(UpperCAmelCase__ ) for job in result["""jobs"""]} ) A__ : Tuple = math.ceil((result["""total_count"""] - 1_0_0) / 1_0_0 ) for i in range(UpperCAmelCase__ ): A__ : Tuple = requests.get(url + f'&page={i + 2}', headers=UpperCAmelCase__ ).json() job_time.update({job["""name"""]: extract_time_from_single_job(UpperCAmelCase__ ) for job in result["""jobs"""]} ) return job_time except Exception: print(f'Unknown error, could not fetch links:\n{traceback.format_exc()}' ) return {} if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument('''--workflow_run_id''', type=str, required=True, help='''A GitHub Actions workflow run id.''') A_ = parser.parse_args() A_ = get_job_time(args.workflow_run_id) A_ = dict(sorted(job_time.items(), key=lambda item: item[1]["duration"], reverse=True)) for k, v in job_time.items(): print(F'{k}: {v["duration"]}')
296
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
1
"""simple docstring""" import operator as op A_ = '''scaler.pt''' A_ = '''pytorch_model''' A_ = '''random_states''' A_ = '''optimizer''' A_ = '''scheduler''' A_ = '''pytorch_model.bin''' A_ = '''pytorch_model.bin.index.json''' A_ = '''model.safetensors''' A_ = '''model.safetensors.index.json''' A_ = '''1.10.2''' A_ = '''py38''' A_ = '''4.17.0''' A_ = ['''ml.p3.16xlarge''', '''ml.p3dn.24xlarge''', '''ml.p4dn.24xlarge'''] A_ = ['''FULL_SHARD''', '''SHARD_GRAD_OP''', '''NO_SHARD''', '''HYBRID_SHARD''', '''HYBRID_SHARD_ZERO2'''] A_ = ['''TRANSFORMER_BASED_WRAP''', '''SIZE_BASED_WRAP''', '''NO_WRAP'''] A_ = ['''BACKWARD_PRE''', '''BACKWARD_POST''', '''NO_PREFETCH'''] A_ = ['''FULL_STATE_DICT''', '''LOCAL_STATE_DICT''', '''SHARDED_STATE_DICT'''] A_ = '''2.0.1''' A_ = ['''pdsh''', '''standard''', '''openmpi''', '''mvapich'''] A_ = ['''default''', '''reduce-overhead''', '''max-autotune'''] A_ = {'''>''': op.gt, '''>=''': op.ge, '''==''': op.eq, '''!=''': op.ne, '''<=''': op.le, '''<''': op.lt} # These are the args for `torch.distributed.launch` for pytorch < 1.9 A_ = [ '''nnodes''', '''nproc_per_node''', '''rdzv_backend''', '''rdzv_endpoint''', '''rdzv_id''', '''rdzv_conf''', '''standalone''', '''max_restarts''', '''monitor_interval''', '''start_method''', '''role''', '''module''', '''m''', '''no_python''', '''run_path''', '''log_dir''', '''r''', '''redirects''', '''t''', '''tee''', '''node_rank''', '''master_addr''', '''master_port''', ] A_ = ['''DEEPSPEED''', '''MULTI_GPU''', '''FSDP''', '''MEGATRON_LM'''] A_ = ['''DEEPSPEED''', '''MULTI_XPU''', '''FSDP''']
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import Any, Callable, Dict, List, Optional, Union import pandas as pd import pyarrow as pa import datasets import datasets.config from datasets.features.features import require_storage_cast from datasets.table import table_cast from datasets.utils.py_utils import Literal A_ = datasets.utils.logging.get_logger(__name__) A_ = ['''names''', '''prefix'''] A_ = ['''warn_bad_lines''', '''error_bad_lines''', '''mangle_dupe_cols'''] A_ = ['''encoding_errors''', '''on_bad_lines'''] A_ = ['''date_format'''] @dataclass class __SCREAMING_SNAKE_CASE ( datasets.BuilderConfig ): snake_case_ = "," snake_case_ = None snake_case_ = "infer" snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = True snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = False snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = True snake_case_ = True snake_case_ = False snake_case_ = True snake_case_ = None snake_case_ = "." snake_case_ = None snake_case_ = '"' snake_case_ = 0 snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = None snake_case_ = True snake_case_ = True snake_case_ = 0 snake_case_ = True snake_case_ = False snake_case_ = None snake_case_ = 10000 snake_case_ = None snake_case_ = "strict" snake_case_ = "error" snake_case_ = None def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' if self.delimiter is not None: A__ : Tuple = self.delimiter if self.column_names is not None: A__ : str = self.column_names @property def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Optional[Any] = { """sep""": self.sep, """header""": self.header, """names""": self.names, """index_col""": self.index_col, """usecols""": self.usecols, """prefix""": self.prefix, """mangle_dupe_cols""": self.mangle_dupe_cols, """engine""": self.engine, """converters""": self.converters, """true_values""": self.true_values, """false_values""": self.false_values, """skipinitialspace""": self.skipinitialspace, """skiprows""": self.skiprows, """nrows""": self.nrows, """na_values""": self.na_values, """keep_default_na""": self.keep_default_na, """na_filter""": self.na_filter, """verbose""": self.verbose, """skip_blank_lines""": self.skip_blank_lines, """thousands""": self.thousands, """decimal""": self.decimal, """lineterminator""": self.lineterminator, """quotechar""": self.quotechar, """quoting""": self.quoting, """escapechar""": self.escapechar, """comment""": self.comment, """encoding""": self.encoding, """dialect""": self.dialect, """error_bad_lines""": self.error_bad_lines, """warn_bad_lines""": self.warn_bad_lines, """skipfooter""": self.skipfooter, """doublequote""": self.doublequote, """memory_map""": self.memory_map, """float_precision""": self.float_precision, """chunksize""": self.chunksize, """encoding_errors""": self.encoding_errors, """on_bad_lines""": self.on_bad_lines, """date_format""": self.date_format, } # some kwargs must not be passed if they don't have a default value # some others are deprecated and we can also not pass them if they are the default value for pd_read_csv_parameter in _PANDAS_READ_CSV_NO_DEFAULT_PARAMETERS + _PANDAS_READ_CSV_DEPRECATED_PARAMETERS: if pd_read_csv_kwargs[pd_read_csv_parameter] == getattr(CsvConfig() , snake_case ): del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 2.0 new arguments if not (datasets.config.PANDAS_VERSION.major >= 2): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_2_0_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] # Remove 1.3 new arguments if not (datasets.config.PANDAS_VERSION.major >= 1 and datasets.config.PANDAS_VERSION.minor >= 3): for pd_read_csv_parameter in _PANDAS_READ_CSV_NEW_1_3_0_PARAMETERS: del pd_read_csv_kwargs[pd_read_csv_parameter] return pd_read_csv_kwargs class __SCREAMING_SNAKE_CASE ( datasets.ArrowBasedBuilder ): snake_case_ = CsvConfig def _UpperCamelCase ( self : Any ): '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Dict ): '''simple docstring''' if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}' ) A__ : int = dl_manager.download_and_extract(self.config.data_files ) if isinstance(snake_case , (str, list, tuple) ): A__ : Optional[Any] = data_files if isinstance(snake_case , snake_case ): A__ : List[str] = [files] A__ : Union[str, Any] = [dl_manager.iter_files(snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] A__ : str = [] for split_name, files in data_files.items(): if isinstance(snake_case , snake_case ): A__ : List[Any] = [files] A__ : List[str] = [dl_manager.iter_files(snake_case ) for file in files] splits.append(datasets.SplitGenerator(name=snake_case , gen_kwargs={"""files""": files} ) ) return splits def _UpperCamelCase ( self : List[str] , snake_case : pa.Table ): '''simple docstring''' if self.config.features is not None: A__ : int = self.config.features.arrow_schema if all(not require_storage_cast(snake_case ) for feature in self.config.features.values() ): # cheaper cast A__ : int = pa.Table.from_arrays([pa_table[field.name] for field in schema] , schema=snake_case ) else: # more expensive cast; allows str <-> int/float or str to Audio for example A__ : Dict = table_cast(snake_case , snake_case ) return pa_table def _UpperCamelCase ( self : int , snake_case : Dict ): '''simple docstring''' A__ : Optional[int] = self.config.features.arrow_schema if self.config.features else None # dtype allows reading an int column as str A__ : int = ( { name: dtype.to_pandas_dtype() if not require_storage_cast(snake_case ) else object for name, dtype, feature in zip(schema.names , schema.types , self.config.features.values() ) } if schema is not None else None ) for file_idx, file in enumerate(itertools.chain.from_iterable(snake_case ) ): A__ : Optional[int] = pd.read_csv(snake_case , iterator=snake_case , dtype=snake_case , **self.config.pd_read_csv_kwargs ) try: for batch_idx, df in enumerate(snake_case ): A__ : Union[str, Any] = pa.Table.from_pandas(snake_case ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(snake_case ) except ValueError as e: logger.error(F'Failed to read file \'{file}\' with error {type(snake_case )}: {e}' ) raise
296
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A_ = '''src/diffusers''' A_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. A_ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) A_ = spec.loader.load_module() def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any] ) ->Any: return line.startswith(UpperCAmelCase__ ) or len(UpperCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", UpperCAmelCase__ ) is not None def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Union[str, Any]: A__ : Any = object_name.split(""".""" ) A__ : int = 0 # First let's find the module where our object lives. A__ : str = parts[i] while i < len(UpperCAmelCase__ ) and not os.path.isfile(os.path.join(UpperCAmelCase__, f'{module}.py' ) ): i += 1 if i < len(UpperCAmelCase__ ): A__ : Union[str, Any] = os.path.join(UpperCAmelCase__, parts[i] ) if i >= len(UpperCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(UpperCAmelCase__, f'{module}.py' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : List[Any] = f.readlines() # Now let's find the class / func in the code! A__ : Optional[Any] = """""" A__ : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A__ : List[Any] = line_index while line_index < len(UpperCAmelCase__ ) and _should_continue(lines[line_index], UpperCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : List[Any] = lines[start_index:line_index] return "".join(UpperCAmelCase__ ) A_ = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') A_ = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') A_ = re.compile(r'''<FILL\s+[^>]*>''') def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Optional[Any]: A__ : Dict = code.split("""\n""" ) A__ : List[Any] = 0 while idx < len(UpperCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase__ ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->int: A__ : str = len(get_indent(UpperCAmelCase__ ) ) > 0 if has_indent: A__ : Union[str, Any] = f'class Bla:\n{code}' A__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=1_1_9, preview=UpperCAmelCase__ ) A__ : Tuple = black.format_str(UpperCAmelCase__, mode=UpperCAmelCase__ ) A__ , A__ : List[Any] = style_docstrings_in_code(UpperCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : Dict=False ) ->List[Any]: with open(UpperCAmelCase__, """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : int = f.readlines() A__ : Dict = [] A__ : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase__ ): A__ : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A__ , A__ , A__ : Dict = search.groups() A__ : Tuple = find_code_in_diffusers(UpperCAmelCase__ ) A__ : int = get_indent(UpperCAmelCase__ ) A__ : List[str] = line_index + 1 if indent == theoretical_indent else line_index + 2 A__ : Tuple = theoretical_indent A__ : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A__ : Tuple = True while line_index < len(UpperCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase__ ): break A__ : Optional[int] = lines[line_index] A__ : Tuple = _should_continue(UpperCAmelCase__, UpperCAmelCase__ ) and re.search(f'^{indent}# End copy', UpperCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : Dict = lines[start_index:line_index] A__ : Tuple = """""".join(UpperCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies A__ : Optional[int] = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase__ ) is None] A__ : Optional[Any] = """\n""".join(UpperCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase__ ) > 0: A__ : int = replace_pattern.replace("""with""", """""" ).split(""",""" ) A__ : List[Any] = [_re_replace_pattern.search(UpperCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue A__ , A__ , A__ : Union[str, Any] = pattern.groups() A__ : Union[str, Any] = re.sub(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if option.strip() == "all-casing": A__ : List[Any] = re.sub(obja.lower(), obja.lower(), UpperCAmelCase__ ) A__ : Tuple = re.sub(obja.upper(), obja.upper(), UpperCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A__ : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) A__ : List[Any] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A__ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] A__ : Tuple = start_index + 1 if overwrite and len(UpperCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(UpperCAmelCase__, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(UpperCAmelCase__ ) return diffs def _lowerCAmelCase ( UpperCAmelCase__ : bool = False ) ->Any: A__ : Dict = glob.glob(os.path.join(UpperCAmelCase__, """**/*.py""" ), recursive=UpperCAmelCase__ ) A__ : str = [] for filename in all_files: A__ : Any = is_copy_consistent(UpperCAmelCase__, UpperCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(UpperCAmelCase__ ) > 0: A__ : Any = """\n""".join(UpperCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() check_copies(args.fix_and_overwrite)
296
1
"""simple docstring""" import math A_ = 10 A_ = 7 A_ = BALLS_PER_COLOUR * NUM_COLOURS def _lowerCAmelCase ( UpperCAmelCase__ : int = 2_0 ) ->str: A__ : List[Any] = math.comb(UpperCAmelCase__, UpperCAmelCase__ ) A__ : Optional[int] = math.comb(NUM_BALLS - BALLS_PER_COLOUR, UpperCAmelCase__ ) A__ : int = NUM_COLOURS * (1 - missing_colour / total) return f'{result:.9f}' if __name__ == "__main__": print(solution(20))
296
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
1
"""simple docstring""" import warnings from ...utils import logging from .image_processing_layoutlmva import LayoutLMvaImageProcessor A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Union[str, Any] , *snake_case : Tuple , **snake_case : Optional[int] ): '''simple docstring''' warnings.warn( """The class LayoutLMv2FeatureExtractor is deprecated and will be removed in version 5 of Transformers.""" """ Please use LayoutLMv2ImageProcessor instead.""" , snake_case , ) super().__init__(*snake_case , **snake_case )
296
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels A_ = object() # For specifying empty leaf dict `{}` A_ = object() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any] ) ->Dict: A__ : Union[str, Any] = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) + 1 ): A__ : Optional[Any] = [x.match(UpperCAmelCase__ ) for x, y in zip(UpperCAmelCase__, ks[i:] )] if matches and all(UpperCAmelCase__ ): return True return False def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Dict: def replace(UpperCAmelCase__ : int, UpperCAmelCase__ : List[str] ): for rule, replacement in rules: if _match(UpperCAmelCase__, UpperCAmelCase__ ): return replacement return val return replace def _lowerCAmelCase ( ) ->Tuple: return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""", UpperCAmelCase__ )), (("transformer", "wte", "embedding"), P("""mp""", UpperCAmelCase__ )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(UpperCAmelCase__, """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(UpperCAmelCase__, """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : Union[str, Any] = _get_partition_rules() A__ : int = _replacement_rules(UpperCAmelCase__ ) A__ : Tuple = {k: _unmatched for k in flatten_dict(UpperCAmelCase__ )} A__ : Optional[int] = {k: replace(UpperCAmelCase__, UpperCAmelCase__ ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(UpperCAmelCase__ ) )
296
1
"""simple docstring""" import json import os import unittest from transformers.models.biogpt.tokenization_biogpt import VOCAB_FILES_NAMES, BioGptTokenizer from transformers.testing_utils import slow from ...test_tokenization_common import TokenizerTesterMixin class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = BioGptTokenizer snake_case_ = False def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A__ : int = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] A__ : Tuple = dict(zip(snake_case , range(len(snake_case ) ) ) ) A__ : Dict = ["""l o 123""", """lo w 1456""", """e r</w> 1789""", """"""] A__ : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(snake_case ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(snake_case ) ) def _UpperCamelCase ( self : str , snake_case : str ): '''simple docstring''' A__ : List[Any] = """lower newer""" A__ : Tuple = """lower newer""" return input_text, output_text def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Optional[Any] = BioGptTokenizer(self.vocab_file , self.merges_file ) A__ : Tuple = """lower""" A__ : Optional[int] = ["""low""", """er</w>"""] A__ : List[str] = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , snake_case ) A__ : Optional[Any] = tokens + ["""<unk>"""] A__ : List[str] = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Dict = tokenizer.encode("""sequence builders""" , add_special_tokens=snake_case ) A__ : str = tokenizer.encode("""multi-sequence build""" , add_special_tokens=snake_case ) A__ : int = tokenizer.build_inputs_with_special_tokens(snake_case ) A__ : Optional[int] = tokenizer.build_inputs_with_special_tokens(snake_case , snake_case ) self.assertTrue(encoded_sentence == [2] + text ) self.assertTrue(encoded_pair == [2] + text + [2] + text_a )
296
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] , snake_case : Tuple , snake_case : List[str]=2 , snake_case : List[str]=8 , snake_case : List[Any]=True , snake_case : Optional[Any]=True , snake_case : List[Any]=True , snake_case : Dict=True , snake_case : Tuple=99 , snake_case : Dict=16 , snake_case : Dict=5 , snake_case : int=2 , snake_case : Any=36 , snake_case : str="gelu" , snake_case : Dict=0.0 , snake_case : List[Any]=0.0 , snake_case : int=512 , snake_case : List[Any]=16 , snake_case : Tuple=2 , snake_case : Any=0.02 , snake_case : Optional[Any]=3 , snake_case : List[Any]=4 , snake_case : str=None , ): '''simple docstring''' A__ : Union[str, Any] = parent A__ : Optional[Any] = batch_size A__ : Dict = seq_length A__ : str = is_training A__ : Tuple = use_input_mask A__ : Dict = use_token_type_ids A__ : Dict = use_labels A__ : int = vocab_size A__ : List[str] = hidden_size A__ : Union[str, Any] = num_hidden_layers A__ : int = num_attention_heads A__ : List[str] = intermediate_size A__ : int = hidden_act A__ : str = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : Any = max_position_embeddings A__ : Optional[int] = type_vocab_size A__ : int = type_sequence_label_size A__ : Optional[Any] = initializer_range A__ : int = num_labels A__ : Optional[int] = num_choices A__ : Optional[int] = scope def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Any = None if self.use_input_mask: A__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Optional[int] = None if self.use_token_type_ids: A__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Dict = None A__ : List[str] = None A__ : Union[str, Any] = None if self.use_labels: A__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Any = ids_tensor([self.batch_size] , self.num_choices ) A__ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.get_config() A__ : List[str] = 300 return config def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Tuple = self.prepare_config_and_inputs() A__ : List[str] = True A__ : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _UpperCamelCase ( self : Any , snake_case : Any , snake_case : Tuple , snake_case : Any , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Dict ): '''simple docstring''' A__ : List[str] = MraModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A__ : List[str] = model(snake_case , token_type_ids=snake_case ) A__ : Union[str, Any] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : List[Any] , snake_case : Any , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Dict , snake_case : str , snake_case : Dict , snake_case : str , ): '''simple docstring''' A__ : Dict = True A__ : Optional[Any] = MraModel(snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , ) A__ : Optional[int] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : str , snake_case : Union[str, Any] , snake_case : Dict , snake_case : List[str] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Dict , snake_case : Dict , snake_case : Dict , snake_case : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : Dict = MraForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : Tuple , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Optional[int] , snake_case : List[str] , snake_case : Union[str, Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Optional[Any] = MraForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict , snake_case : str , snake_case : List[Any] , snake_case : Any , snake_case : Dict , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Union[str, Any] = MraForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : Tuple , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Dict , snake_case : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.num_choices A__ : str = MraForMultipleChoice(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Dict = config_and_inputs A__ : Optional[int] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = () def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Optional[Any] = MraModelTester(self ) A__ : List[str] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : List[str] = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : str = MraModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip(reason="""MRA does not output attentions""" ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Any = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : List[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , snake_case ) A__ : int = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Dict = 5_0265 A__ : List[str] = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : List[Any] = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) A__ : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Union[str, Any] = 5_0265 A__ : Optional[Any] = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : Optional[int] = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) )
296
1
"""simple docstring""" from argparse import ArgumentParser from datasets.commands.convert import ConvertCommand from datasets.commands.dummy_data import DummyDataCommand from datasets.commands.env import EnvironmentCommand from datasets.commands.run_beam import RunBeamCommand from datasets.commands.test import TestCommand from datasets.utils.logging import set_verbosity_info def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->Optional[int]: return {key.lstrip("""-""" ): value for key, value in zip(unknown_args[::2], unknown_args[1::2] )} def _lowerCAmelCase ( ) ->int: A__ : List[str] = ArgumentParser( """HuggingFace Datasets CLI tool""", usage="""datasets-cli <command> [<args>]""", allow_abbrev=UpperCAmelCase__ ) A__ : Optional[int] = parser.add_subparsers(help="""datasets-cli command helpers""" ) set_verbosity_info() # Register commands ConvertCommand.register_subcommand(UpperCAmelCase__ ) EnvironmentCommand.register_subcommand(UpperCAmelCase__ ) TestCommand.register_subcommand(UpperCAmelCase__ ) RunBeamCommand.register_subcommand(UpperCAmelCase__ ) DummyDataCommand.register_subcommand(UpperCAmelCase__ ) # Parse args A__ , A__ : Dict = parser.parse_known_args() if not hasattr(UpperCAmelCase__, """func""" ): parser.print_help() exit(1 ) A__ : List[str] = parse_unknown_args(UpperCAmelCase__ ) # Run A__ : int = args.func(UpperCAmelCase__, **UpperCAmelCase__ ) service.run() if __name__ == "__main__": main()
296
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets A_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' A_ = '''\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. ''' A_ = ''' Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {"raw_values", "uniform_average"} or array-like of shape (n_outputs,), default="uniform_average" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. "raw_values" : Returns a full set of errors in case of multioutput input. "uniform_average" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric("mse") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {\'mse\': 0.6123724356957945} If you\'re using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric("mse", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput=\'raw_values\') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mse\': array([0.41666667, 1. ])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : Dict ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html""" ] , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("""float""" ) ), "references": datasets.Sequence(datasets.Value("""float""" ) ), } else: return { "predictions": datasets.Value("""float""" ), "references": datasets.Value("""float""" ), } def _UpperCamelCase ( self : List[str] , snake_case : Dict , snake_case : List[Any] , snake_case : List[str]=None , snake_case : List[Any]="uniform_average" , snake_case : int=True ): '''simple docstring''' A__ : Optional[int] = mean_squared_error( snake_case , snake_case , sample_weight=snake_case , multioutput=snake_case , squared=snake_case ) return {"mse": mse}
296
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''EleutherAI/gpt-neox-20b''': '''https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json''', # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'gpt_neox' def __init__( self : Union[str, Any] , snake_case : List[str]=5_0432 , snake_case : int=6144 , snake_case : List[Any]=44 , snake_case : str=64 , snake_case : Optional[int]=2_4576 , snake_case : List[Any]="gelu" , snake_case : Optional[Any]=0.25 , snake_case : Optional[int]=1_0000 , snake_case : Union[str, Any]=0.0 , snake_case : str=0.0 , snake_case : Tuple=0.1 , snake_case : int=2048 , snake_case : Dict=0.02 , snake_case : Optional[int]=1e-5 , snake_case : Any=True , snake_case : int=0 , snake_case : str=2 , snake_case : Tuple=False , snake_case : Union[str, Any]=True , snake_case : List[Any]=None , **snake_case : List[str] , ): '''simple docstring''' super().__init__(bos_token_id=snake_case , eos_token_id=snake_case , **snake_case ) A__ : int = vocab_size A__ : Any = max_position_embeddings A__ : int = hidden_size A__ : int = num_hidden_layers A__ : Tuple = num_attention_heads A__ : Any = intermediate_size A__ : List[Any] = hidden_act A__ : List[Any] = rotary_pct A__ : Dict = rotary_emb_base A__ : Tuple = attention_dropout A__ : Optional[Any] = hidden_dropout A__ : Tuple = classifier_dropout A__ : List[Any] = initializer_range A__ : Union[str, Any] = layer_norm_eps A__ : Union[str, Any] = use_cache A__ : int = tie_word_embeddings A__ : List[str] = use_parallel_residual A__ : Optional[Any] = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( """The hidden size is not divisble by the number of attention heads! Make sure to update them!""" ) def _UpperCamelCase ( self : str ): '''simple docstring''' if self.rope_scaling is None: return if not isinstance(self.rope_scaling , snake_case ) or len(self.rope_scaling ) != 2: raise ValueError( """`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, """ F'got {self.rope_scaling}' ) A__ : List[Any] = self.rope_scaling.get("""type""" , snake_case ) A__ : Union[str, Any] = self.rope_scaling.get("""factor""" , snake_case ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}' ) if rope_scaling_factor is None or not isinstance(snake_case , snake_case ) or rope_scaling_factor <= 1.0: raise ValueError(F'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}' )
296
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Optional[int] , snake_case : List[str]=None , **snake_case : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , snake_case , ) super().__init__(args=snake_case , **snake_case )
296
1
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Optional[int] , snake_case : List[str]=None , **snake_case : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , snake_case , ) super().__init__(args=snake_case , **snake_case )
296
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch A_ = random.Random() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Tuple=1.0, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : str=None ) ->Union[str, Any]: if rng is None: A__ : Optional[int] = global_rng A__ : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[str]=7 , snake_case : str=400 , snake_case : Optional[Any]=2000 , snake_case : Union[str, Any]=10 , snake_case : str=160 , snake_case : List[str]=8 , snake_case : List[Any]=0.0 , snake_case : Optional[Any]=4000 , snake_case : Any=False , snake_case : int=True , ): '''simple docstring''' A__ : Any = parent A__ : str = batch_size A__ : List[str] = min_seq_length A__ : Dict = max_seq_length A__ : str = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) A__ : Dict = padding_value A__ : Optional[Any] = sampling_rate A__ : Any = return_attention_mask A__ : Optional[int] = do_normalize A__ : Tuple = feature_size A__ : Optional[Any] = chunk_length A__ : Union[str, Any] = hop_length def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict=False , snake_case : Optional[Any]=False ): '''simple docstring''' def _flatten(snake_case : Dict ): return list(itertools.chain(*snake_case ) ) if equal_length: A__ : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size A__ : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: A__ : List[str] = [np.asarray(snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = WhisperFeatureExtractor if is_speech_available() else None def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : str = WhisperFeatureExtractionTester(self ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] = feat_extract_first.save_pretrained(snake_case )[0] check_json_file_has_correct_format(snake_case ) A__ : Union[str, Any] = self.feature_extraction_class.from_pretrained(snake_case ) A__ : str = feat_extract_first.to_dict() A__ : Union[str, Any] = feat_extract_second.to_dict() A__ : List[Any] = feat_extract_first.mel_filters A__ : Optional[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : Any = os.path.join(snake_case , """feat_extract.json""" ) feat_extract_first.to_json_file(snake_case ) A__ : int = self.feature_extraction_class.from_json_file(snake_case ) A__ : Dict = feat_extract_first.to_dict() A__ : str = feat_extract_second.to_dict() A__ : str = feat_extract_first.mel_filters A__ : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 A__ : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] # Test feature size A__ : Dict = feature_extractor(snake_case , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input A__ : str = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test batched A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. A__ : Tuple = [floats_list((1, x) )[0] for x in (800, 800, 800)] A__ : str = np.asarray(snake_case ) A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test truncation required A__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] A__ : Union[str, Any] = [x[: feature_extractor.n_samples] for x in speech_inputs] A__ : str = [np.asarray(snake_case ) for speech_input in speech_inputs_truncated] A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : str = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' import torch A__ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : List[str] = np.random.rand(100 , 32 ).astype(np.floataa ) A__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: A__ : Optional[Any] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) A__ : Optional[int] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' A__ : int = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech A__ : Union[str, Any] = ds.sort("""id""" ).select(range(snake_case ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on A__ : Optional[Any] = self._load_datasamples(1 ) A__ : Union[str, Any] = WhisperFeatureExtractor() A__ : List[str] = feature_extractor(snake_case , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , snake_case , atol=1e-4 ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : Union[str, Any] = self._load_datasamples(1 )[0] A__ : Any = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue A__ : str = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=snake_case )[0] self.assertTrue(np.all(np.mean(snake_case ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case ) - 1 ) < 1e-3 ) )
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_0_0_0_0_0_0 ) ->int: A__ : List[str] = limit + 1 A__ : Tuple = [0] * limit for first_term in range(1, UpperCAmelCase__ ): for n in range(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ): A__ : List[str] = first_term + n / first_term if common_difference % 4: # d must be divisble by 4 continue else: common_difference /= 4 if ( first_term > common_difference and first_term < 4 * common_difference ): # since x,y,z are positive integers frequency[n] += 1 # so z>0 and a>d ,also 4d<a A__ : Any = sum(1 for x in frequency[1:limit] if x == 1_0 ) return count if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] ): '''simple docstring''' A__ : Optional[int] = (0, 0) A__ : Dict = None A__ : int = 0 A__ : str = 0 A__ : Optional[Any] = 0 def __eq__( self : str , snake_case : Optional[int] ): '''simple docstring''' return self.position == cell.position def _UpperCamelCase ( self : List[str] ): '''simple docstring''' print(self.position ) class __SCREAMING_SNAKE_CASE : def __init__( self : int , snake_case : Any=(5, 5) ): '''simple docstring''' A__ : Optional[int] = np.zeros(snake_case ) A__ : List[Any] = world_size[0] A__ : Dict = world_size[1] def _UpperCamelCase ( self : Any ): '''simple docstring''' print(self.w ) def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : int = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] A__ : int = cell.position[0] A__ : str = cell.position[1] A__ : Any = [] for n in neughbour_cord: A__ : List[Any] = current_x + n[0] A__ : Tuple = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: A__ : List[Any] = Cell() A__ : str = (x, y) A__ : Optional[Any] = cell neighbours.append(snake_case ) return neighbours def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Dict ) ->Dict: A__ : Union[str, Any] = [] A__ : Optional[int] = [] _open.append(UpperCAmelCase__ ) while _open: A__ : List[Any] = np.argmin([n.f for n in _open] ) A__ : Union[str, Any] = _open[min_f] _closed.append(_open.pop(UpperCAmelCase__ ) ) if current == goal: break for n in world.get_neigbours(UpperCAmelCase__ ): for c in _closed: if c == n: continue A__ : Dict = current.g + 1 A__ , A__ : int = n.position A__ , A__ : Optional[int] = goal.position A__ : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 A__ : Optional[int] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(UpperCAmelCase__ ) A__ : List[str] = [] while current.parent is not None: path.append(current.position ) A__ : Union[str, Any] = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": A_ = Gridworld() # Start position and goal A_ = Cell() A_ = (0, 0) A_ = Cell() A_ = (4, 4) print(F'path from {start.position} to {goal.position}') A_ = astar(world, start, goal) # Just for visual reasons. for i in s: A_ = 1 print(world.w)
296
1
"""simple docstring""" from collections import OrderedDict from ...utils import logging from .auto_factory import _BaseAutoModelClass, _LazyAutoMapping, auto_class_update from .configuration_auto import CONFIG_MAPPING_NAMES A_ = logging.get_logger(__name__) A_ = OrderedDict( [ # Base model mapping ('''albert''', '''FlaxAlbertModel'''), ('''bart''', '''FlaxBartModel'''), ('''beit''', '''FlaxBeitModel'''), ('''bert''', '''FlaxBertModel'''), ('''big_bird''', '''FlaxBigBirdModel'''), ('''blenderbot''', '''FlaxBlenderbotModel'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallModel'''), ('''clip''', '''FlaxCLIPModel'''), ('''distilbert''', '''FlaxDistilBertModel'''), ('''electra''', '''FlaxElectraModel'''), ('''gpt-sw3''', '''FlaxGPT2Model'''), ('''gpt2''', '''FlaxGPT2Model'''), ('''gpt_neo''', '''FlaxGPTNeoModel'''), ('''gptj''', '''FlaxGPTJModel'''), ('''longt5''', '''FlaxLongT5Model'''), ('''marian''', '''FlaxMarianModel'''), ('''mbart''', '''FlaxMBartModel'''), ('''mt5''', '''FlaxMT5Model'''), ('''opt''', '''FlaxOPTModel'''), ('''pegasus''', '''FlaxPegasusModel'''), ('''regnet''', '''FlaxRegNetModel'''), ('''resnet''', '''FlaxResNetModel'''), ('''roberta''', '''FlaxRobertaModel'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormModel'''), ('''roformer''', '''FlaxRoFormerModel'''), ('''t5''', '''FlaxT5Model'''), ('''vision-text-dual-encoder''', '''FlaxVisionTextDualEncoderModel'''), ('''vit''', '''FlaxViTModel'''), ('''wav2vec2''', '''FlaxWav2Vec2Model'''), ('''whisper''', '''FlaxWhisperModel'''), ('''xglm''', '''FlaxXGLMModel'''), ('''xlm-roberta''', '''FlaxXLMRobertaModel'''), ] ) A_ = OrderedDict( [ # Model for pre-training mapping ('''albert''', '''FlaxAlbertForPreTraining'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForPreTraining'''), ('''big_bird''', '''FlaxBigBirdForPreTraining'''), ('''electra''', '''FlaxElectraForPreTraining'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ('''wav2vec2''', '''FlaxWav2Vec2ForPreTraining'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) A_ = OrderedDict( [ # Model for Masked LM mapping ('''albert''', '''FlaxAlbertForMaskedLM'''), ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''bert''', '''FlaxBertForMaskedLM'''), ('''big_bird''', '''FlaxBigBirdForMaskedLM'''), ('''distilbert''', '''FlaxDistilBertForMaskedLM'''), ('''electra''', '''FlaxElectraForMaskedLM'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''roberta''', '''FlaxRobertaForMaskedLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMaskedLM'''), ('''roformer''', '''FlaxRoFormerForMaskedLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMaskedLM'''), ] ) A_ = OrderedDict( [ # Model for Seq2Seq Causal LM mapping ('''bart''', '''FlaxBartForConditionalGeneration'''), ('''blenderbot''', '''FlaxBlenderbotForConditionalGeneration'''), ('''blenderbot-small''', '''FlaxBlenderbotSmallForConditionalGeneration'''), ('''encoder-decoder''', '''FlaxEncoderDecoderModel'''), ('''longt5''', '''FlaxLongT5ForConditionalGeneration'''), ('''marian''', '''FlaxMarianMTModel'''), ('''mbart''', '''FlaxMBartForConditionalGeneration'''), ('''mt5''', '''FlaxMT5ForConditionalGeneration'''), ('''pegasus''', '''FlaxPegasusForConditionalGeneration'''), ('''t5''', '''FlaxT5ForConditionalGeneration'''), ] ) A_ = OrderedDict( [ # Model for Image-classsification ('''beit''', '''FlaxBeitForImageClassification'''), ('''regnet''', '''FlaxRegNetForImageClassification'''), ('''resnet''', '''FlaxResNetForImageClassification'''), ('''vit''', '''FlaxViTForImageClassification'''), ] ) A_ = OrderedDict( [ ('''vision-encoder-decoder''', '''FlaxVisionEncoderDecoderModel'''), ] ) A_ = OrderedDict( [ # Model for Causal LM mapping ('''bart''', '''FlaxBartForCausalLM'''), ('''bert''', '''FlaxBertForCausalLM'''), ('''big_bird''', '''FlaxBigBirdForCausalLM'''), ('''electra''', '''FlaxElectraForCausalLM'''), ('''gpt-sw3''', '''FlaxGPT2LMHeadModel'''), ('''gpt2''', '''FlaxGPT2LMHeadModel'''), ('''gpt_neo''', '''FlaxGPTNeoForCausalLM'''), ('''gptj''', '''FlaxGPTJForCausalLM'''), ('''opt''', '''FlaxOPTForCausalLM'''), ('''roberta''', '''FlaxRobertaForCausalLM'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForCausalLM'''), ('''xglm''', '''FlaxXGLMForCausalLM'''), ('''xlm-roberta''', '''FlaxXLMRobertaForCausalLM'''), ] ) A_ = OrderedDict( [ # Model for Sequence Classification mapping ('''albert''', '''FlaxAlbertForSequenceClassification'''), ('''bart''', '''FlaxBartForSequenceClassification'''), ('''bert''', '''FlaxBertForSequenceClassification'''), ('''big_bird''', '''FlaxBigBirdForSequenceClassification'''), ('''distilbert''', '''FlaxDistilBertForSequenceClassification'''), ('''electra''', '''FlaxElectraForSequenceClassification'''), ('''mbart''', '''FlaxMBartForSequenceClassification'''), ('''roberta''', '''FlaxRobertaForSequenceClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForSequenceClassification'''), ('''roformer''', '''FlaxRoFormerForSequenceClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForSequenceClassification'''), ] ) A_ = OrderedDict( [ # Model for Question Answering mapping ('''albert''', '''FlaxAlbertForQuestionAnswering'''), ('''bart''', '''FlaxBartForQuestionAnswering'''), ('''bert''', '''FlaxBertForQuestionAnswering'''), ('''big_bird''', '''FlaxBigBirdForQuestionAnswering'''), ('''distilbert''', '''FlaxDistilBertForQuestionAnswering'''), ('''electra''', '''FlaxElectraForQuestionAnswering'''), ('''mbart''', '''FlaxMBartForQuestionAnswering'''), ('''roberta''', '''FlaxRobertaForQuestionAnswering'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForQuestionAnswering'''), ('''roformer''', '''FlaxRoFormerForQuestionAnswering'''), ('''xlm-roberta''', '''FlaxXLMRobertaForQuestionAnswering'''), ] ) A_ = OrderedDict( [ # Model for Token Classification mapping ('''albert''', '''FlaxAlbertForTokenClassification'''), ('''bert''', '''FlaxBertForTokenClassification'''), ('''big_bird''', '''FlaxBigBirdForTokenClassification'''), ('''distilbert''', '''FlaxDistilBertForTokenClassification'''), ('''electra''', '''FlaxElectraForTokenClassification'''), ('''roberta''', '''FlaxRobertaForTokenClassification'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForTokenClassification'''), ('''roformer''', '''FlaxRoFormerForTokenClassification'''), ('''xlm-roberta''', '''FlaxXLMRobertaForTokenClassification'''), ] ) A_ = OrderedDict( [ # Model for Multiple Choice mapping ('''albert''', '''FlaxAlbertForMultipleChoice'''), ('''bert''', '''FlaxBertForMultipleChoice'''), ('''big_bird''', '''FlaxBigBirdForMultipleChoice'''), ('''distilbert''', '''FlaxDistilBertForMultipleChoice'''), ('''electra''', '''FlaxElectraForMultipleChoice'''), ('''roberta''', '''FlaxRobertaForMultipleChoice'''), ('''roberta-prelayernorm''', '''FlaxRobertaPreLayerNormForMultipleChoice'''), ('''roformer''', '''FlaxRoFormerForMultipleChoice'''), ('''xlm-roberta''', '''FlaxXLMRobertaForMultipleChoice'''), ] ) A_ = OrderedDict( [ ('''bert''', '''FlaxBertForNextSentencePrediction'''), ] ) A_ = OrderedDict( [ ('''speech-encoder-decoder''', '''FlaxSpeechEncoderDecoderModel'''), ('''whisper''', '''FlaxWhisperForConditionalGeneration'''), ] ) A_ = OrderedDict( [ ('''whisper''', '''FlaxWhisperForAudioClassification'''), ] ) A_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_MAPPING_NAMES) A_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_PRETRAINING_MAPPING_NAMES) A_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MASKED_LM_MAPPING_NAMES) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING_NAMES ) A_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING_NAMES) A_ = _LazyAutoMapping(CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_CAUSAL_LM_MAPPING_NAMES) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING_NAMES ) A_ = _LazyAutoMapping( CONFIG_MAPPING_NAMES, FLAX_MODEL_FOR_AUDIO_CLASSIFICATION_MAPPING_NAMES ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_MAPPING A_ = auto_class_update(FlaxAutoModel) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_PRETRAINING_MAPPING A_ = auto_class_update(FlaxAutoModelForPreTraining, head_doc='''pretraining''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_CAUSAL_LM_MAPPING A_ = auto_class_update(FlaxAutoModelForCausalLM, head_doc='''causal language modeling''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_MASKED_LM_MAPPING A_ = auto_class_update(FlaxAutoModelForMaskedLM, head_doc='''masked language modeling''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_SEQ_TO_SEQ_CAUSAL_LM_MAPPING A_ = auto_class_update( FlaxAutoModelForSeqaSeqLM, head_doc='''sequence-to-sequence language modeling''', checkpoint_for_example='''t5-base''' ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING A_ = auto_class_update( FlaxAutoModelForSequenceClassification, head_doc='''sequence classification''' ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_QUESTION_ANSWERING_MAPPING A_ = auto_class_update(FlaxAutoModelForQuestionAnswering, head_doc='''question answering''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = auto_class_update( FlaxAutoModelForTokenClassification, head_doc='''token classification''' ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_MULTIPLE_CHOICE_MAPPING A_ = auto_class_update(FlaxAutoModelForMultipleChoice, head_doc='''multiple choice''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_NEXT_SENTENCE_PREDICTION_MAPPING A_ = auto_class_update( FlaxAutoModelForNextSentencePrediction, head_doc='''next sentence prediction''' ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_IMAGE_CLASSIFICATION_MAPPING A_ = auto_class_update( FlaxAutoModelForImageClassification, head_doc='''image classification''' ) class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_VISION_2_SEQ_MAPPING A_ = auto_class_update(FlaxAutoModelForVisionaSeq, head_doc='''vision-to-text modeling''') class __SCREAMING_SNAKE_CASE ( _BaseAutoModelClass ): snake_case_ = FLAX_MODEL_FOR_SPEECH_SEQ_2_SEQ_MAPPING A_ = auto_class_update( FlaxAutoModelForSpeechSeqaSeq, head_doc='''sequence-to-sequence speech-to-text modeling''' )
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple=False ) ->str: A__ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ : Optional[int] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]=False ) ->str: for i in range(config.num_hidden_layers ): if base_model: A__ : Any = """""" else: A__ : Tuple = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A__ : str = in_proj_bias[: config.hidden_size] A__ : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A__ : Any = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : int = dct.pop(UpperCAmelCase__ ) A__ : Tuple = val def _lowerCAmelCase ( ) ->List[Any]: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Any ) ->Tuple: A__ : List[Any] = DeiTConfig() # all deit models have fine-tuned heads A__ : Tuple = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ : str = 1_0_0_0 A__ : List[str] = """huggingface/label-files""" A__ : Dict = """imagenet-1k-id2label.json""" A__ : List[str] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Dict = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[int] = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} A__ : List[str] = int(deit_name[-6:-4] ) A__ : str = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): A__ : List[str] = 1_9_2 A__ : int = 7_6_8 A__ : List[Any] = 1_2 A__ : Dict = 3 elif deit_name[9:].startswith("""small""" ): A__ : List[Any] = 3_8_4 A__ : List[str] = 1_5_3_6 A__ : Any = 1_2 A__ : Union[str, Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): A__ : int = 1_0_2_4 A__ : str = 4_0_9_6 A__ : Any = 2_4 A__ : int = 1_6 # load original model from timm A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ : Tuple = timm_model.state_dict() A__ : str = create_rename_keys(UpperCAmelCase__, UpperCAmelCase__ ) for src, dest in rename_keys: rename_key(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : str = DeiTForImageClassificationWithTeacher(UpperCAmelCase__ ).eval() model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor A__ : int = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ : Any = DeiTImageProcessor(size=UpperCAmelCase__, crop_size=config.image_size ) A__ : Union[str, Any] = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : Optional[Any] = encoding["""pixel_values"""] A__ : Union[str, Any] = model(UpperCAmelCase__ ) A__ : Union[str, Any] = timm_model(UpperCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCAmelCase__, outputs.logits, atol=1e-3 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
296
1
"""simple docstring""" A_ = ''' # Transformers installation ! pip install transformers datasets # To install from source instead of the last release, comment the command above and uncomment the following one. # ! pip install git+https://github.com/huggingface/transformers.git ''' A_ = [{'''type''': '''code''', '''content''': INSTALL_CONTENT}] A_ = { '''{processor_class}''': '''FakeProcessorClass''', '''{model_class}''': '''FakeModelClass''', '''{object_class}''': '''FakeObjectClass''', }
296
"""simple docstring""" from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] A__ : Optional[int] = (low + high) // 2 A__ , A__ , A__ : List[Any] = max_subarray(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_subarray(UpperCAmelCase__, mid + 1, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_cross_sum(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int, int, float]: A__ , A__ : Dict = float("""-inf""" ), -1 A__ , A__ : Optional[Any] = float("""-inf""" ), -1 A__ : int | float = 0 for i in range(UpperCAmelCase__, low - 1, -1 ): summ += arr[i] if summ > left_sum: A__ : Optional[int] = summ A__ : Union[str, Any] = i A__ : Optional[Any] = 0 for i in range(mid + 1, high + 1 ): summ += arr[i] if summ > right_sum: A__ : int = summ A__ : Union[str, Any] = i return max_left, max_right, (left_sum + right_sum) def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->float: A__ : Union[str, Any] = [randint(1, UpperCAmelCase__ ) for _ in range(UpperCAmelCase__ )] A__ : Any = time.time() max_subarray(UpperCAmelCase__, 0, input_size - 1 ) A__ : List[Any] = time.time() return end - start def _lowerCAmelCase ( ) ->None: A__ : List[Any] = [1_0, 1_0_0, 1_0_0_0, 1_0_0_0_0, 5_0_0_0_0, 1_0_0_0_0_0, 2_0_0_0_0_0, 3_0_0_0_0_0, 4_0_0_0_0_0, 5_0_0_0_0_0] A__ : Any = [time_max_subarray(UpperCAmelCase__ ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(UpperCAmelCase__, UpperCAmelCase__ ): print(UpperCAmelCase__, """\t\t""", UpperCAmelCase__ ) plt.plot(UpperCAmelCase__, UpperCAmelCase__ ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
296
1
"""simple docstring""" import os def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : Any = len(grid[0] ) A__ : Optional[Any] = len(UpperCAmelCase__ ) A__ : str = 0 A__ : List[str] = 0 A__ : Optional[Any] = 0 # Check vertically, horizontally, diagonally at the same time (only works # for nxn grid) for i in range(UpperCAmelCase__ ): for j in range(n_rows - 3 ): A__ : Any = grid[j][i] * grid[j + 1][i] * grid[j + 2][i] * grid[j + 3][i] A__ : List[str] = grid[i][j] * grid[i][j + 1] * grid[i][j + 2] * grid[i][j + 3] # Left-to-right diagonal (\) product if i < n_columns - 3: A__ : Optional[int] = ( grid[i][j] * grid[i + 1][j + 1] * grid[i + 2][j + 2] * grid[i + 3][j + 3] ) # Right-to-left diagonal(/) product if i > 2: A__ : str = ( grid[i][j] * grid[i - 1][j + 1] * grid[i - 2][j + 2] * grid[i - 3][j + 3] ) A__ : List[str] = max( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if max_product > largest: A__ : Dict = max_product return largest def _lowerCAmelCase ( ) ->List[str]: A__ : List[Any] = [] with open(os.path.dirname(UpperCAmelCase__ ) + """/grid.txt""" ) as file: for line in file: grid.append(line.strip("""\n""" ).split(""" """ ) ) A__ : Optional[int] = [[int(UpperCAmelCase__ ) for i in grid[j]] for j in range(len(UpperCAmelCase__ ) )] return largest_product(UpperCAmelCase__ ) if __name__ == "__main__": print(solution())
296
"""simple docstring""" from __future__ import annotations class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , snake_case : int ): '''simple docstring''' A__ : List[Any] = order # a_{0} ... a_{k} A__ : List[Any] = [1.0] + [0.0] * order # b_{0} ... b_{k} A__ : str = [1.0] + [0.0] * order # x[n-1] ... x[n-k] A__ : Union[str, Any] = [0.0] * self.order # y[n-1] ... y[n-k] A__ : List[str] = [0.0] * self.order def _UpperCamelCase ( self : Optional[int] , snake_case : list[float] , snake_case : list[float] ): '''simple docstring''' if len(snake_case ) < self.order: A__ : Any = [1.0, *a_coeffs] if len(snake_case ) != self.order + 1: A__ : str = ( F'Expected a_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) if len(snake_case ) != self.order + 1: A__ : Union[str, Any] = ( F'Expected b_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) A__ : Dict = a_coeffs A__ : Any = b_coeffs def _UpperCamelCase ( self : List[str] , snake_case : float ): '''simple docstring''' A__ : str = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) A__ : Dict = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] A__ : Tuple = self.input_history[:-1] A__ : int = self.output_history[:-1] A__ : Dict = sample A__ : Tuple = result return result
296
1
"""simple docstring""" # limitations under the License. # NOTE: This file is deprecated and will be removed in a future version. # It only exists so that temporarely `from diffusers.pipelines import DiffusionPipeline` works from .pipelines import DiffusionPipeline, ImagePipelineOutput # noqa: F401 from .utils import deprecate deprecate( '''pipelines_utils''', '''0.22.0''', '''Importing `DiffusionPipeline` or `ImagePipelineOutput` from diffusers.pipeline_utils is deprecated. Please import from diffusers.pipelines.pipeline_utils instead.''', standard_warn=False, stacklevel=3, )
296
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , snake_case : Optional[Any] , snake_case : Tuple=13 , snake_case : Dict=7 , snake_case : Optional[int]=True , snake_case : Union[str, Any]=True , snake_case : Dict=True , snake_case : Any=True , snake_case : List[str]=99 , snake_case : str=64 , snake_case : Optional[int]=5 , snake_case : str=4 , snake_case : List[Any]=37 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : Dict=16 , snake_case : List[Any]=2 , snake_case : Optional[int]=0.02 , snake_case : Any=3 , snake_case : Union[str, Any]=4 , snake_case : Dict=None , ): '''simple docstring''' A__ : Tuple = parent A__ : Union[str, Any] = batch_size A__ : List[str] = seq_length A__ : Optional[int] = is_training A__ : Dict = use_input_mask A__ : Any = use_token_type_ids A__ : Optional[Any] = use_labels A__ : List[str] = vocab_size A__ : Optional[int] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : Any = num_attention_heads A__ : List[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Optional[int] = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[Any] = initializer_range A__ : Optional[int] = num_labels A__ : Dict = num_choices A__ : Dict = scope A__ : List[Any] = vocab_size - 1 def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[Any] = None if self.use_input_mask: A__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_labels: A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Tuple = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ , A__ , A__ : str = self.prepare_config_and_inputs() A__ : Union[str, Any] = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : List[str] , snake_case : int ): '''simple docstring''' A__ : Any = GPTNeoXModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case ) A__ : Optional[int] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : str , snake_case : Any , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = True A__ : str = GPTNeoXModel(snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Dict , snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : int = self.num_labels A__ : int = GPTNeoXForQuestionAnswering(snake_case ) model.to(snake_case ) model.eval() A__ : Optional[Any] = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : str , snake_case : Tuple , snake_case : int , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : List[Any] = self.num_labels A__ : Tuple = GPTNeoXForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Any , snake_case : Union[str, Any] , snake_case : int , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : Any = GPTNeoXForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Optional[int] = True A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass A__ : Tuple = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ : str = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Any = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Tuple = model(snake_case , attention_mask=snake_case , output_hidden_states=snake_case ) A__ : List[Any] = output_from_no_past["""hidden_states"""][0] A__ : List[str] = model( snake_case , attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )["""hidden_states"""][0] # select random slice A__ : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Dict = config_and_inputs A__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) snake_case_ = (GPTNeoXForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = GPTNeoXModelTester(self ) A__ : Any = ConfigTester(self , config_class=snake_case , hidden_size=64 , num_attention_heads=8 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ , A__ , A__ , A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() A__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ , A__ , A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] = ids_tensor([1, 10] , config.vocab_size ) A__ : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Union[str, Any] = GPTNeoXModel(snake_case ) original_model.to(snake_case ) original_model.eval() A__ : Optional[int] = original_model(snake_case ).last_hidden_state A__ : List[str] = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Optional[int] = {"""type""": scaling_type, """factor""": 10.0} A__ : Optional[int] = GPTNeoXModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() A__ : List[str] = scaled_model(snake_case ).last_hidden_state A__ : Tuple = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: A__ : Optional[Any] = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(snake_case ) A__ : Optional[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" A__ : Tuple = model.generate(**snake_case , do_sample=snake_case , max_new_tokens=20 ) A__ : Tuple = tokenizer.batch_decode(snake_case )[0] self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 42 class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): @register_to_config def __init__( self : int , snake_case : int = 6_5536 , snake_case : Optional[int] = None , snake_case : int = 2 , snake_case : int = 2 , snake_case : int = 0 , snake_case : str = "fourier" , snake_case : bool = True , snake_case : bool = False , snake_case : float = 0.0 , snake_case : Tuple[str] = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , snake_case : Tuple[str] = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , snake_case : Tuple[str] = "UNetMidBlock1D" , snake_case : str = None , snake_case : Tuple[int] = (32, 32, 64) , snake_case : str = None , snake_case : int = 8 , snake_case : int = 1 , snake_case : bool = False , ): '''simple docstring''' super().__init__() A__ : Dict = sample_size # time if time_embedding_type == "fourier": A__ : Any = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=snake_case , log=snake_case , flip_sin_to_cos=snake_case ) A__ : int = 2 * block_out_channels[0] elif time_embedding_type == "positional": A__ : Dict = Timesteps( block_out_channels[0] , flip_sin_to_cos=snake_case , downscale_freq_shift=snake_case ) A__ : Tuple = block_out_channels[0] if use_timestep_embedding: A__ : List[str] = block_out_channels[0] * 4 A__ : List[str] = TimestepEmbedding( in_channels=snake_case , time_embed_dim=snake_case , act_fn=snake_case , out_dim=block_out_channels[0] , ) A__ : List[str] = nn.ModuleList([] ) A__ : Tuple = None A__ : Optional[Any] = nn.ModuleList([] ) A__ : Optional[int] = None # down A__ : List[str] = in_channels for i, down_block_type in enumerate(snake_case ): A__ : Optional[Any] = output_channel A__ : Dict = block_out_channels[i] if i == 0: input_channel += extra_in_channels A__ : Any = i == len(snake_case ) - 1 A__ : str = get_down_block( snake_case , num_layers=snake_case , in_channels=snake_case , out_channels=snake_case , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(snake_case ) # mid A__ : int = get_mid_block( snake_case , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=snake_case , add_downsample=snake_case , ) # up A__ : Union[str, Any] = list(reversed(snake_case ) ) A__ : List[str] = reversed_block_out_channels[0] if out_block_type is None: A__ : Dict = out_channels else: A__ : Dict = block_out_channels[0] for i, up_block_type in enumerate(snake_case ): A__ : List[Any] = output_channel A__ : Any = ( reversed_block_out_channels[i + 1] if i < len(snake_case ) - 1 else final_upsample_channels ) A__ : Dict = i == len(snake_case ) - 1 A__ : Tuple = get_up_block( snake_case , num_layers=snake_case , in_channels=snake_case , out_channels=snake_case , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(snake_case ) A__ : str = output_channel # out A__ : str = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) A__ : Union[str, Any] = get_out_block( out_block_type=snake_case , num_groups_out=snake_case , embed_dim=block_out_channels[0] , out_channels=snake_case , act_fn=snake_case , fc_dim=block_out_channels[-1] // 4 , ) def _UpperCamelCase ( self : List[Any] , snake_case : torch.FloatTensor , snake_case : Union[torch.Tensor, float, int] , snake_case : bool = True , ): '''simple docstring''' A__ : List[Any] = timestep if not torch.is_tensor(snake_case ): A__ : Optional[Any] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(snake_case ) and len(timesteps.shape ) == 0: A__ : Optional[Any] = timesteps[None].to(sample.device ) A__ : Optional[Any] = self.time_proj(snake_case ) if self.config.use_timestep_embedding: A__ : Optional[int] = self.time_mlp(snake_case ) else: A__ : int = timestep_embed[..., None] A__ : Dict = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) A__ : Union[str, Any] = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down A__ : List[str] = () for downsample_block in self.down_blocks: A__ , A__ : int = downsample_block(hidden_states=snake_case , temb=snake_case ) down_block_res_samples += res_samples # 3. mid if self.mid_block: A__ : List[Any] = self.mid_block(snake_case , snake_case ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): A__ : Optional[Any] = down_block_res_samples[-1:] A__ : str = down_block_res_samples[:-1] A__ : Optional[int] = upsample_block(snake_case , res_hidden_states_tuple=snake_case , temb=snake_case ) # 5. post-process if self.out_block: A__ : List[Any] = self.out_block(snake_case , snake_case ) if not return_dict: return (sample,) return UNetaDOutput(sample=snake_case )
296
"""simple docstring""" from collections import defaultdict from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_5_0_0_0_0_0 ) ->int: A__ : defaultdict = defaultdict(UpperCAmelCase__ ) A__ : Any = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, UpperCAmelCase__, 2 ): if gcd(UpperCAmelCase__, UpperCAmelCase__ ) > 1: continue A__ : str = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'{solution() = }')
296
1
"""simple docstring""" import os A_ = {'''I''': 1, '''V''': 5, '''X''': 10, '''L''': 50, '''C''': 100, '''D''': 500, '''M''': 1000} def _lowerCAmelCase ( UpperCAmelCase__ : str ) ->int: A__ : Optional[int] = 0 A__ : Optional[Any] = 0 while index < len(UpperCAmelCase__ ) - 1: A__ : Any = SYMBOLS[numerals[index]] A__ : List[str] = SYMBOLS[numerals[index + 1]] if current_value < next_value: total_value -= current_value else: total_value += current_value index += 1 total_value += SYMBOLS[numerals[index]] return total_value def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->str: A__ : Union[str, Any] = """""" A__ : Dict = num // 1_0_0_0 numerals += m_count * "M" num %= 1_0_0_0 A__ : Optional[Any] = num // 1_0_0 if c_count == 9: numerals += "CM" c_count -= 9 elif c_count == 4: numerals += "CD" c_count -= 4 if c_count >= 5: numerals += "D" c_count -= 5 numerals += c_count * "C" num %= 1_0_0 A__ : Dict = num // 1_0 if x_count == 9: numerals += "XC" x_count -= 9 elif x_count == 4: numerals += "XL" x_count -= 4 if x_count >= 5: numerals += "L" x_count -= 5 numerals += x_count * "X" num %= 1_0 if num == 9: numerals += "IX" num -= 9 elif num == 4: numerals += "IV" num -= 4 if num >= 5: numerals += "V" num -= 5 numerals += num * "I" return numerals def _lowerCAmelCase ( UpperCAmelCase__ : str = "/p089_roman.txt" ) ->int: A__ : str = 0 with open(os.path.dirname(UpperCAmelCase__ ) + roman_numerals_filename ) as filea: A__ : List[Any] = filea.readlines() for line in lines: A__ : Optional[Any] = line.strip() A__ : Tuple = parse_roman_numerals(UpperCAmelCase__ ) A__ : Tuple = generate_roman_numerals(UpperCAmelCase__ ) savings += len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) return savings if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" import os from distutils.util import strtobool def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any] ) ->List[str]: for e in env_keys: A__ : List[Any] = int(os.environ.get(UpperCAmelCase__, -1 ) ) if val >= 0: return val return default def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : str=False ) ->List[str]: A__ : List[Any] = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return strtobool(UpperCAmelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]="no" ) ->int: A__ : str = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return value
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, ) A_ = { '''configuration_resnet''': ['''RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ResNetConfig''', '''ResNetOnnxConfig'''] } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ResNetForImageClassification''', '''ResNetModel''', '''ResNetPreTrainedModel''', '''ResNetBackbone''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFResNetForImageClassification''', '''TFResNetModel''', '''TFResNetPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''FlaxResNetForImageClassification''', '''FlaxResNetModel''', '''FlaxResNetPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_resnet import RESNET_PRETRAINED_CONFIG_ARCHIVE_MAP, ResNetConfig, ResNetOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_resnet import ( RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, ResNetBackbone, ResNetForImageClassification, ResNetModel, ResNetPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_resnet import ( TF_RESNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFResNetForImageClassification, TFResNetModel, TFResNetPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_resnet import FlaxResNetForImageClassification, FlaxResNetModel, FlaxResNetPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure)
296
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
1
"""simple docstring""" from math import ceil, sqrt def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_0_0_0_0_0_0 ) ->int: A__ : List[Any] = 0 for outer_width in range(3, (limit // 4) + 2 ): if outer_width**2 > limit: A__ : Dict = max(ceil(sqrt(outer_width**2 - limit ) ), 1 ) else: A__ : List[str] = 1 if (outer_width - hole_width_lower_bound) % 2: hole_width_lower_bound += 1 answer += (outer_width - hole_width_lower_bound - 2) // 2 + 1 return answer if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = logging.get_logger(__name__) A_ = Dict[str, Any] A_ = List[Prediction] @add_end_docstrings(UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : str , *snake_case : Tuple , **snake_case : Tuple ): '''simple docstring''' super().__init__(*snake_case , **snake_case ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _UpperCamelCase ( self : List[Any] , **snake_case : Optional[int] ): '''simple docstring''' A__ : Dict = {} if "threshold" in kwargs: A__ : int = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return super().__call__(*snake_case , **snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : List[str] = load_image(snake_case ) A__ : int = torch.IntTensor([[image.height, image.width]] ) A__ : Union[str, Any] = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: A__ : str = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) A__ : List[str] = target_size return inputs def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : str = model_inputs.pop("""target_size""" ) A__ : Dict = self.model(**snake_case ) A__ : Optional[Any] = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: A__ : str = model_inputs["""bbox"""] return model_outputs def _UpperCamelCase ( self : Tuple , snake_case : Optional[int] , snake_case : int=0.9 ): '''simple docstring''' A__ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A__ , A__ : Tuple = target_size[0].tolist() def unnormalize(snake_case : Optional[int] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) A__ , A__ : Optional[int] = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A__ : Optional[Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A__ : List[str] = [unnormalize(snake_case ) for bbox in model_outputs["""bbox"""].squeeze(0 )] A__ : Tuple = ["""score""", """label""", """box"""] A__ : Any = [dict(zip(snake_case , snake_case ) ) for vals in zip(scores.tolist() , snake_case , snake_case ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A__ : Union[str, Any] = self.image_processor.post_process_object_detection(snake_case , snake_case , snake_case ) A__ : str = raw_annotations[0] A__ : str = raw_annotation["""scores"""] A__ : List[Any] = raw_annotation["""labels"""] A__ : int = raw_annotation["""boxes"""] A__ : str = scores.tolist() A__ : Any = [self.model.config.idalabel[label.item()] for label in labels] A__ : int = [self._get_bounding_box(snake_case ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A__ : str = ["""score""", """label""", """box"""] A__ : Dict = [ dict(zip(snake_case , snake_case ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def _UpperCamelCase ( self : Union[str, Any] , snake_case : "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) A__ , A__ , A__ , A__ : Any = box.int().tolist() A__ : Any = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
"""simple docstring""" def _lowerCAmelCase ( ) ->List[Any]: for n in range(1, 1_0_0_0_0_0_0 ): yield n * (n + 1) // 2 def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->str: A__ : Union[str, Any] = 1 A__ : Tuple = 2 while i * i <= n: A__ : List[Any] = 0 while n % i == 0: n //= i multiplicity += 1 divisors_count *= multiplicity + 1 i += 1 if n > 1: divisors_count *= 2 return divisors_count def _lowerCAmelCase ( ) ->Any: return next(i for i in triangle_number_generator() if count_divisors(UpperCAmelCase__ ) > 5_0_0 ) if __name__ == "__main__": print(solution())
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
"""simple docstring""" import argparse from transformers import TaConfig, TaForConditionalGeneration, load_tf_weights_in_ta from transformers.utils import logging logging.set_verbosity_info() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : Tuple ) ->int: # Initialise PyTorch model A__ : Optional[int] = TaConfig.from_json_file(UpperCAmelCase__ ) print(f'Building PyTorch model from configuration: {config}' ) A__ : int = TaForConditionalGeneration(UpperCAmelCase__ ) # Load weights from tf checkpoint load_tf_weights_in_ta(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--tf_checkpoint_path''', default=None, type=str, required=True, help='''Path to the TensorFlow checkpoint path.''' ) parser.add_argument( '''--config_file''', default=None, type=str, required=True, help=( '''The config json file corresponding to the pre-trained T5 model. \nThis specifies the model architecture.''' ), ) parser.add_argument( '''--pytorch_dump_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model.''' ) A_ = parser.parse_args() convert_tf_checkpoint_to_pytorch(args.tf_checkpoint_path, args.config_file, args.pytorch_dump_path)
296
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
1
"""simple docstring""" def _lowerCAmelCase ( ) ->int: return [ a * b * (1_0_0_0 - a - b) for a in range(1, 9_9_9 ) for b in range(UpperCAmelCase__, 9_9_9 ) if (a * a + b * b == (1_0_0_0 - a - b) ** 2) ][0] if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[Any]=13 , snake_case : Union[str, Any]=7 , snake_case : Optional[Any]=True , snake_case : str=True , snake_case : Dict=False , snake_case : Union[str, Any]=True , snake_case : Optional[Any]=99 , snake_case : str=32 , snake_case : Tuple=5 , snake_case : List[str]=4 , snake_case : Optional[int]=37 , snake_case : str="gelu" , snake_case : Tuple=0.1 , snake_case : Optional[int]=0.1 , snake_case : int=512 , snake_case : List[str]=16 , snake_case : str=2 , snake_case : Optional[int]=0.02 , snake_case : str=3 , snake_case : Dict=4 , snake_case : Optional[Any]=None , ): '''simple docstring''' A__ : int = parent A__ : Union[str, Any] = batch_size A__ : Optional[int] = seq_length A__ : List[Any] = is_training A__ : List[str] = use_input_mask A__ : Optional[Any] = use_token_type_ids A__ : List[Any] = use_labels A__ : Union[str, Any] = vocab_size A__ : List[Any] = hidden_size A__ : Any = num_hidden_layers A__ : Any = num_attention_heads A__ : Optional[int] = intermediate_size A__ : Any = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Dict = attention_probs_dropout_prob A__ : Optional[int] = max_position_embeddings A__ : Tuple = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[str] = initializer_range A__ : Any = num_labels A__ : Any = num_choices A__ : int = scope def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = None if self.use_input_mask: A__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_token_type_ids: A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : int = None A__ : int = None A__ : List[str] = None if self.use_labels: A__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) A__ : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Any , snake_case : Dict , snake_case : Any , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case ) A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Optional[int] , snake_case : List[str] , snake_case : str , snake_case : Optional[Any] , snake_case : List[str] , snake_case : List[Any] , snake_case : Tuple , snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = BioGptForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Any , snake_case : str , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Any , *snake_case : Dict ): '''simple docstring''' A__ : Union[str, Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() # create attention mask A__ : List[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) A__ : Any = self.seq_length // 2 A__ : str = 0 # first forward pass A__ , A__ : List[Any] = model(snake_case , attention_mask=snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids A__ : int = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A__ : List[str] = ids_tensor((1,) , snake_case ).item() + 1 A__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A__ : int = random_other_next_tokens # append to next input_ids and attn_mask A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case )] , dim=1 , ) # get two different outputs A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Optional[int] = model(snake_case , past_key_values=snake_case , attention_mask=snake_case )["""last_hidden_state"""] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() A__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str , snake_case : int , snake_case : Optional[Any] , *snake_case : str ): '''simple docstring''' A__ : Dict = BioGptModel(config=snake_case ).to(snake_case ).eval() A__ : Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) # first forward pass A__ : Dict = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ , A__ : List[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Optional[int] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A__ : Any = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , past_key_values=snake_case )[ """last_hidden_state""" ] # select random slice A__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Tuple , *snake_case : Union[str, Any] , snake_case : Union[str, Any]=False ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM(snake_case ) model.to(snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() A__ : Optional[Any] = model(snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self : int , snake_case : Optional[Any] , *snake_case : Optional[int] ): '''simple docstring''' A__ : int = BioGptModel(snake_case ) A__ : Union[str, Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _UpperCamelCase ( self : Any , snake_case : Dict , snake_case : Tuple , snake_case : int , snake_case : Union[str, Any] , snake_case : Dict , *snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = self.num_labels A__ : int = BioGptForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : str = config_and_inputs A__ : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case_ = (BioGptForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[str] = BioGptModelTester(self ) A__ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : str = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case , gradient_checkpointing=snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) A__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = """left""" # Define PAD Token = EOS Token = 50256 A__ : Optional[int] = tokenizer.eos_token A__ : Dict = model.config.eos_token_id # use different length sentences to test batching A__ : Union[str, Any] = [ """Hello, my dog is a little""", """Today, I""", ] A__ : List[str] = tokenizer(snake_case , return_tensors="""pt""" , padding=snake_case ) A__ : str = inputs["""input_ids"""].to(snake_case ) A__ : Dict = model.generate( input_ids=snake_case , attention_mask=inputs["""attention_mask"""].to(snake_case ) , ) A__ : Optional[int] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Any = model.generate(input_ids=snake_case ) A__ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A__ : str = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Dict = model.generate(input_ids=snake_case , max_length=model.config.max_length - num_paddings ) A__ : Optional[Any] = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case ) A__ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case ) A__ : Optional[int] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(snake_case , [non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = BioGptModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(snake_case ) A__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Union[str, Any] = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Any = 3 A__ : List[Any] = """multi_label_classification""" A__ : Dict = input_dict["""input_ids"""] A__ : Tuple = input_ids.ne(1 ).to(snake_case ) A__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Tuple = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A__ : str = torch.tensor([[2, 4805, 9, 656, 21]] ) A__ : Dict = model(snake_case )[0] A__ : Tuple = 4_2384 A__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : str = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) torch.manual_seed(0 ) A__ : Tuple = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(snake_case ) A__ : Optional[int] = model.generate( **snake_case , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case , ) A__ : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case ) A__ : List[str] = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" import json import os import pickle import shutil import tempfile from unittest import TestCase from unittest.mock import patch import numpy as np from datasets import Dataset from transformers import is_faiss_available from transformers.models.bart.configuration_bart import BartConfig from transformers.models.bart.tokenization_bart import BartTokenizer from transformers.models.bert.tokenization_bert import VOCAB_FILES_NAMES as DPR_VOCAB_FILES_NAMES from transformers.models.dpr.configuration_dpr import DPRConfig from transformers.models.dpr.tokenization_dpr import DPRContextEncoderTokenizer, DPRQuestionEncoderTokenizer from transformers.models.rag.configuration_rag import RagConfig from transformers.models.rag.retrieval_rag import CustomHFIndex, RagRetriever from transformers.models.roberta.tokenization_roberta import VOCAB_FILES_NAMES as BART_VOCAB_FILES_NAMES from transformers.testing_utils import require_faiss, require_sentencepiece, require_tokenizers, require_torch if is_faiss_available(): import faiss @require_faiss class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Tuple = tempfile.mkdtemp() A__ : List[Any] = 8 # DPR tok A__ : str = [ """[UNK]""", """[CLS]""", """[SEP]""", """[PAD]""", """[MASK]""", """want""", """##want""", """##ed""", """wa""", """un""", """runn""", """##ing""", """,""", """low""", """lowest""", ] A__ : Union[str, Any] = os.path.join(self.tmpdirname , """dpr_tokenizer""" ) os.makedirs(snake_case , exist_ok=snake_case ) A__ : str = os.path.join(snake_case , DPR_VOCAB_FILES_NAMES["""vocab_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as vocab_writer: vocab_writer.write("""""".join([x + """\n""" for x in vocab_tokens] ) ) # BART tok A__ : Optional[int] = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """\u0120""", """\u0120l""", """\u0120n""", """\u0120lo""", """\u0120low""", """er""", """\u0120lowest""", """\u0120newer""", """\u0120wider""", """<unk>""", ] A__ : Tuple = dict(zip(snake_case , range(len(snake_case ) ) ) ) A__ : Optional[Any] = ["""#version: 0.2""", """\u0120 l""", """\u0120l o""", """\u0120lo w""", """e r""", """"""] A__ : Any = {"""unk_token""": """<unk>"""} A__ : List[Any] = os.path.join(self.tmpdirname , """bart_tokenizer""" ) os.makedirs(snake_case , exist_ok=snake_case ) A__ : List[Any] = os.path.join(snake_case , BART_VOCAB_FILES_NAMES["""vocab_file"""] ) A__ : List[Any] = os.path.join(snake_case , BART_VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case ) + """\n""" ) with open(self.merges_file , """w""" , encoding="""utf-8""" ) as fp: fp.write("""\n""".join(snake_case ) ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' return DPRQuestionEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , """dpr_tokenizer""" ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return DPRContextEncoderTokenizer.from_pretrained(os.path.join(self.tmpdirname , """dpr_tokenizer""" ) ) def _UpperCamelCase ( self : int ): '''simple docstring''' return BartTokenizer.from_pretrained(os.path.join(self.tmpdirname , """bart_tokenizer""" ) ) def _UpperCamelCase ( self : int ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : List[str] = Dataset.from_dict( { """id""": ["""0""", """1"""], """text""": ["""foo""", """bar"""], """title""": ["""Foo""", """Bar"""], """embeddings""": [np.ones(self.retrieval_vector_size ), 2 * np.ones(self.retrieval_vector_size )], } ) dataset.add_faiss_index("""embeddings""" , string_factory="""Flat""" , metric_type=faiss.METRIC_INNER_PRODUCT ) return dataset def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = self.get_dummy_dataset() A__ : int = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , ) with patch("""transformers.models.rag.retrieval_rag.load_dataset""" ) as mock_load_dataset: A__ : Any = dataset A__ : Any = RagRetriever( snake_case , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) return retriever def _UpperCamelCase ( self : List[str] , snake_case : bool ): '''simple docstring''' A__ : Dict = self.get_dummy_dataset() A__ : Union[str, Any] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="""custom""" , ) if from_disk: A__ : Any = os.path.join(self.tmpdirname , """dataset""" ) A__ : Tuple = os.path.join(self.tmpdirname , """index.faiss""" ) dataset.get_index("""embeddings""" ).save(os.path.join(self.tmpdirname , """index.faiss""" ) ) dataset.drop_index("""embeddings""" ) dataset.save_to_disk(os.path.join(self.tmpdirname , """dataset""" ) ) del dataset A__ : Tuple = RagRetriever( snake_case , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , ) else: A__ : int = RagRetriever( snake_case , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() , index=CustomHFIndex(config.retrieval_vector_size , snake_case ) , ) return retriever def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = Dataset.from_dict( { """id""": ["""0""", """1"""], """text""": ["""foo""", """bar"""], """title""": ["""Foo""", """Bar"""], """embeddings""": [np.ones(self.retrieval_vector_size + 1 ), 2 * np.ones(self.retrieval_vector_size + 1 )], } ) dataset.add_faiss_index("""embeddings""" , string_factory="""Flat""" , metric_type=faiss.METRIC_INNER_PRODUCT ) A__ : List[str] = os.path.join(self.tmpdirname , """hf_bert_base.hnswSQ8_correct_phi_128.c_index""" ) dataset.save_faiss_index("""embeddings""" , index_file_name + """.index.dpr""" ) pickle.dump(dataset["""id"""] , open(index_file_name + """.index_meta.dpr""" , """wb""" ) ) A__ : int = os.path.join(self.tmpdirname , """psgs_w100.tsv.pkl""" ) A__ : Dict = {sample["""id"""]: [sample["""text"""], sample["""title"""]] for sample in dataset} pickle.dump(snake_case , open(snake_case , """wb""" ) ) A__ : Optional[int] = RagConfig( retrieval_vector_size=self.retrieval_vector_size , question_encoder=DPRConfig().to_dict() , generator=BartConfig().to_dict() , index_name="""legacy""" , index_path=self.tmpdirname , ) A__ : List[Any] = RagRetriever( snake_case , question_encoder_tokenizer=self.get_dpr_tokenizer() , generator_tokenizer=self.get_bart_tokenizer() ) return retriever def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Dict = 1 A__ : str = self.get_dummy_canonical_hf_index_retriever() A__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ , A__ , A__ : Optional[Any] = retriever.retrieve(snake_case , n_docs=snake_case ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(snake_case ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , snake_case ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.get_dummy_canonical_hf_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: with patch("""transformers.models.rag.retrieval_rag.load_dataset""" ) as mock_load_dataset: A__ : Dict = self.get_dummy_dataset() retriever.save_pretrained(snake_case ) A__ : List[str] = RagRetriever.from_pretrained(snake_case ) self.assertIsInstance(snake_case , snake_case ) A__ : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : Union[str, Any] = retriever.retrieve(snake_case , n_docs=1 ) self.assertTrue(out is not None ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Union[str, Any] = 1 A__ : int = self.get_dummy_custom_hf_index_retriever(from_disk=snake_case ) A__ : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ , A__ , A__ : Optional[int] = retriever.retrieve(snake_case , n_docs=snake_case ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(snake_case ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , snake_case ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=snake_case ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(snake_case ) A__ : Dict = RagRetriever.from_pretrained(snake_case ) self.assertIsInstance(snake_case , snake_case ) A__ : List[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : List[Any] = retriever.retrieve(snake_case , n_docs=1 ) self.assertTrue(out is not None ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Tuple = 1 A__ : Dict = self.get_dummy_custom_hf_index_retriever(from_disk=snake_case ) A__ : Any = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ , A__ , A__ : List[Any] = retriever.retrieve(snake_case , n_docs=snake_case ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(snake_case ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""embeddings""", """id""", """text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""id"""] ) , snake_case ) self.assertEqual(doc_dicts[0]["""id"""][0] , """1""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""id"""][0] , """0""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : int = self.get_dummy_custom_hf_index_retriever(from_disk=snake_case ) with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(snake_case ) A__ : Optional[int] = RagRetriever.from_pretrained(snake_case ) self.assertIsInstance(snake_case , snake_case ) A__ : Tuple = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : Any = retriever.retrieve(snake_case , n_docs=1 ) self.assertTrue(out is not None ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : List[str] = 1 A__ : Tuple = self.get_dummy_legacy_index_retriever() A__ : Optional[int] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ , A__ , A__ : int = retriever.retrieve(snake_case , n_docs=snake_case ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertEqual(len(snake_case ) , 2 ) self.assertEqual(sorted(doc_dicts[0] ) , ["""text""", """title"""] ) self.assertEqual(len(doc_dicts[0]["""text"""] ) , snake_case ) self.assertEqual(doc_dicts[0]["""text"""][0] , """bar""" ) # max inner product is reached with second doc self.assertEqual(doc_dicts[1]["""text"""][0] , """foo""" ) # max inner product is reached with first doc self.assertListEqual(doc_ids.tolist() , [[1], [0]] ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.get_dummy_legacy_index_retriever() with tempfile.TemporaryDirectory() as tmp_dirname: retriever.save_pretrained(snake_case ) A__ : Union[str, Any] = RagRetriever.from_pretrained(snake_case ) self.assertIsInstance(snake_case , snake_case ) A__ : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : List[str] = retriever.retrieve(snake_case , n_docs=1 ) self.assertTrue(out is not None ) @require_torch @require_tokenizers @require_sentencepiece def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' import torch A__ : Tuple = 1 A__ : Union[str, Any] = self.get_dummy_canonical_hf_index_retriever() A__ : Union[str, Any] = [[5, 7], [10, 11]] A__ : Optional[Any] = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : Optional[Any] = retriever(snake_case , snake_case , prefix=retriever.config.generator.prefix , n_docs=snake_case ) A__ , A__ , A__ : List[str] = ( out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(snake_case , snake_case ) self.assertIsInstance(snake_case , snake_case ) self.assertIsInstance(snake_case , np.ndarray ) A__ : Union[str, Any] = retriever( snake_case , snake_case , prefix=retriever.config.generator.prefix , n_docs=snake_case , return_tensors="""pt""" , ) A__ , A__ , A__ , A__ : List[Any] = ( # noqa: F841 out["""context_input_ids"""], out["""context_attention_mask"""], out["""retrieved_doc_embeds"""], out["""doc_ids"""], ) self.assertEqual(retrieved_doc_embeds.shape , (2, n_docs, self.retrieval_vector_size) ) self.assertIsInstance(snake_case , torch.Tensor ) self.assertIsInstance(snake_case , torch.Tensor ) self.assertIsInstance(snake_case , torch.Tensor ) @require_torch @require_tokenizers @require_sentencepiece def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = self.get_dpr_ctx_encoder_tokenizer() A__ : str = 1 A__ : Union[str, Any] = self.get_dummy_custom_hf_index_retriever(from_disk=snake_case ) retriever.set_ctx_encoder_tokenizer(snake_case ) A__ : List[str] = [[5, 7], [10, 11]] A__ : Dict = np.array( [np.ones(self.retrieval_vector_size ), -np.ones(self.retrieval_vector_size )] , dtype=np.floataa ) A__ : Dict = retriever(snake_case , snake_case , prefix=retriever.config.generator.prefix , n_docs=snake_case ) self.assertEqual( len(snake_case ) , 6 ) # check whether the retriever output consist of 6 attributes including tokenized docs self.assertEqual( all(k in out for k in ("""tokenized_doc_ids""", """tokenized_doc_attention_mask""") ) , snake_case ) # check for doc token related keys in dictionary.
296
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
1
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
1
"""simple docstring""" from __future__ import annotations import inspect import unittest from math import floor import numpy as np from transformers import CvtConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import TFCvtForImageClassification, TFCvtModel from transformers.models.cvt.modeling_tf_cvt import TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[int] = self.config_class(**self.inputs_dict ) self.parent.assertTrue(hasattr(snake_case , """embed_dim""" ) ) self.parent.assertTrue(hasattr(snake_case , """num_heads""" ) ) class __SCREAMING_SNAKE_CASE : def __init__( self : Any , snake_case : Tuple , snake_case : Dict=13 , snake_case : List[Any]=64 , snake_case : str=3 , snake_case : Dict=[16, 48, 96] , snake_case : Optional[Any]=[1, 3, 6] , snake_case : str=[1, 2, 10] , snake_case : Optional[Any]=[7, 3, 3] , snake_case : Dict=[4, 2, 2] , snake_case : List[str]=[2, 1, 1] , snake_case : Optional[int]=[2, 2, 2] , snake_case : Optional[Any]=[False, False, True] , snake_case : Any=[0.0, 0.0, 0.0] , snake_case : Optional[int]=0.02 , snake_case : Union[str, Any]=1e-12 , snake_case : Any=True , snake_case : Tuple=True , snake_case : Optional[int]=2 , ): '''simple docstring''' A__ : int = parent A__ : str = batch_size A__ : Union[str, Any] = image_size A__ : Dict = patch_sizes A__ : Any = patch_stride A__ : Dict = patch_padding A__ : str = is_training A__ : Optional[Any] = use_labels A__ : Tuple = num_labels A__ : Any = num_channels A__ : Any = embed_dim A__ : Optional[Any] = num_heads A__ : int = stride_kv A__ : Optional[Any] = depth A__ : List[str] = cls_token A__ : Any = attention_drop_rate A__ : List[str] = initializer_range A__ : List[Any] = layer_norm_eps def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[Any] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Dict = None if self.use_labels: # create a random int32 tensor of given shape A__ : int = ids_tensor([self.batch_size] , self.num_labels ) A__ : Any = self.get_config() return config, pixel_values, labels def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return CvtConfig( image_size=self.image_size , num_labels=self.num_labels , num_channels=self.num_channels , embed_dim=self.embed_dim , num_heads=self.num_heads , patch_sizes=self.patch_sizes , patch_padding=self.patch_padding , patch_stride=self.patch_stride , stride_kv=self.stride_kv , depth=self.depth , cls_token=self.cls_token , attention_drop_rate=self.attention_drop_rate , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Dict , snake_case : Union[str, Any] , snake_case : Dict , snake_case : Optional[Any] ): '''simple docstring''' A__ : List[Any] = TFCvtModel(config=snake_case ) A__ : List[Any] = model(snake_case , training=snake_case ) A__ : Dict = (self.image_size, self.image_size) A__ , A__ : List[Any] = image_size[0], image_size[1] for i in range(len(self.depth ) ): A__ : Tuple = floor(((height + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) A__ : Any = floor(((width + 2 * self.patch_padding[i] - self.patch_sizes[i]) / self.patch_stride[i]) + 1 ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.embed_dim[-1], height, width) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : Optional[int] = self.num_labels A__ : Any = TFCvtForImageClassification(snake_case ) A__ : List[str] = model(snake_case , labels=snake_case , training=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Tuple = self.prepare_config_and_inputs() A__ , A__ , A__ : Dict = config_and_inputs A__ : Tuple = {"""pixel_values""": pixel_values} return config, inputs_dict @require_tf class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = (TFCvtModel, TFCvtForImageClassification) if is_tf_available() else () snake_case_ = ( {'feature-extraction': TFCvtModel, 'image-classification': TFCvtForImageClassification} if is_tf_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Union[str, Any] = TFCvtModelTester(self ) A__ : int = TFCvtConfigTester(self , config_class=snake_case , has_text_modality=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() @unittest.skip(reason="""Cvt does not output attentions""" ) def _UpperCamelCase ( self : str ): '''simple docstring''' pass @unittest.skip(reason="""Cvt does not use inputs_embeds""" ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' pass @unittest.skip(reason="""Cvt does not support input and output embeddings""" ) def _UpperCamelCase ( self : int ): '''simple docstring''' pass @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) def _UpperCamelCase ( self : int ): '''simple docstring''' super().test_dataset_conversion() @unittest.skipIf( not is_tf_available() or len(tf.config.list_physical_devices("""GPU""" ) ) == 0 , reason="""TF does not support backprop for grouped convolutions on CPU.""" , ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' super().test_keras_fit() @unittest.skip(reason="""Get `Failed to determine best cudnn convolution algo.` error after using TF 2.12+cuda 11.8""" ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Tuple = tf.keras.mixed_precision.Policy("""mixed_float16""" ) tf.keras.mixed_precision.set_global_policy(snake_case ) super().test_keras_fit() tf.keras.mixed_precision.set_global_policy("""float32""" ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : str = model_class(snake_case ) A__ : List[Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : str = [*signature.parameters.keys()] A__ : Tuple = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' def check_hidden_states_output(snake_case : str , snake_case : Tuple , snake_case : int ): A__ : Dict = model_class(snake_case ) A__ : List[Any] = model(**self._prepare_for_class(snake_case , snake_case ) ) A__ : Optional[Any] = outputs.hidden_states A__ : List[str] = len(self.model_tester.depth ) self.assertEqual(len(snake_case ) , snake_case ) # verify the first hidden states (first block) self.assertListEqual( list(hidden_states[0].shape[-3:] ) , [ self.model_tester.embed_dim[0], self.model_tester.image_size // 4, self.model_tester.image_size // 4, ] , ) A__ , A__ : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Union[str, Any] = True check_hidden_states_output(snake_case , snake_case , snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : List[Any] = True check_hidden_states_output(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case ) @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' for model_name in TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[int] = TFCvtModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _lowerCAmelCase ( ) ->Tuple: A__ : Optional[Any] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_tf @require_vision class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _UpperCamelCase ( self : int ): '''simple docstring''' return AutoImageProcessor.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Any = TFCvtForImageClassification.from_pretrained(TF_CVT_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) A__ : Tuple = self.default_image_processor A__ : str = prepare_img() A__ : Union[str, Any] = image_processor(images=snake_case , return_tensors="""tf""" ) # forward pass A__ : Any = model(**snake_case ) # verify the logits A__ : Optional[int] = tf.TensorShape((1, 1000) ) self.assertEqual(outputs.logits.shape , snake_case ) A__ : Optional[int] = tf.constant([0.9285, 0.9015, -0.3150] ) self.assertTrue(np.allclose(outputs.logits[0, :3].numpy() , snake_case , atol=1e-4 ) )
296
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A_ = '''src/diffusers''' A_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. A_ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) A_ = spec.loader.load_module() def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any] ) ->Any: return line.startswith(UpperCAmelCase__ ) or len(UpperCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", UpperCAmelCase__ ) is not None def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Union[str, Any]: A__ : Any = object_name.split(""".""" ) A__ : int = 0 # First let's find the module where our object lives. A__ : str = parts[i] while i < len(UpperCAmelCase__ ) and not os.path.isfile(os.path.join(UpperCAmelCase__, f'{module}.py' ) ): i += 1 if i < len(UpperCAmelCase__ ): A__ : Union[str, Any] = os.path.join(UpperCAmelCase__, parts[i] ) if i >= len(UpperCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(UpperCAmelCase__, f'{module}.py' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : List[Any] = f.readlines() # Now let's find the class / func in the code! A__ : Optional[Any] = """""" A__ : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A__ : List[Any] = line_index while line_index < len(UpperCAmelCase__ ) and _should_continue(lines[line_index], UpperCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : List[Any] = lines[start_index:line_index] return "".join(UpperCAmelCase__ ) A_ = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') A_ = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') A_ = re.compile(r'''<FILL\s+[^>]*>''') def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Optional[Any]: A__ : Dict = code.split("""\n""" ) A__ : List[Any] = 0 while idx < len(UpperCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase__ ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->int: A__ : str = len(get_indent(UpperCAmelCase__ ) ) > 0 if has_indent: A__ : Union[str, Any] = f'class Bla:\n{code}' A__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=1_1_9, preview=UpperCAmelCase__ ) A__ : Tuple = black.format_str(UpperCAmelCase__, mode=UpperCAmelCase__ ) A__ , A__ : List[Any] = style_docstrings_in_code(UpperCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : Dict=False ) ->List[Any]: with open(UpperCAmelCase__, """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : int = f.readlines() A__ : Dict = [] A__ : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase__ ): A__ : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A__ , A__ , A__ : Dict = search.groups() A__ : Tuple = find_code_in_diffusers(UpperCAmelCase__ ) A__ : int = get_indent(UpperCAmelCase__ ) A__ : List[str] = line_index + 1 if indent == theoretical_indent else line_index + 2 A__ : Tuple = theoretical_indent A__ : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A__ : Tuple = True while line_index < len(UpperCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase__ ): break A__ : Optional[int] = lines[line_index] A__ : Tuple = _should_continue(UpperCAmelCase__, UpperCAmelCase__ ) and re.search(f'^{indent}# End copy', UpperCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : Dict = lines[start_index:line_index] A__ : Tuple = """""".join(UpperCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies A__ : Optional[int] = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase__ ) is None] A__ : Optional[Any] = """\n""".join(UpperCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase__ ) > 0: A__ : int = replace_pattern.replace("""with""", """""" ).split(""",""" ) A__ : List[Any] = [_re_replace_pattern.search(UpperCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue A__ , A__ , A__ : Union[str, Any] = pattern.groups() A__ : Union[str, Any] = re.sub(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if option.strip() == "all-casing": A__ : List[Any] = re.sub(obja.lower(), obja.lower(), UpperCAmelCase__ ) A__ : Tuple = re.sub(obja.upper(), obja.upper(), UpperCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A__ : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) A__ : List[Any] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A__ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] A__ : Tuple = start_index + 1 if overwrite and len(UpperCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(UpperCAmelCase__, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(UpperCAmelCase__ ) return diffs def _lowerCAmelCase ( UpperCAmelCase__ : bool = False ) ->Any: A__ : Dict = glob.glob(os.path.join(UpperCAmelCase__, """**/*.py""" ), recursive=UpperCAmelCase__ ) A__ : str = [] for filename in all_files: A__ : Any = is_copy_consistent(UpperCAmelCase__, UpperCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(UpperCAmelCase__ ) > 0: A__ : Any = """\n""".join(UpperCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() check_copies(args.fix_and_overwrite)
296
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from transformers import BatchEncoding, CanineTokenizer from transformers.testing_utils import require_tokenizers, require_torch from transformers.tokenization_utils import AddedToken from transformers.utils import cached_property from ...test_tokenization_common import TokenizerTesterMixin class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = CanineTokenizer snake_case_ = False def _UpperCamelCase ( self : Dict ): '''simple docstring''' super().setUp() A__ : List[str] = CanineTokenizer() tokenizer.save_pretrained(self.tmpdirname ) @cached_property def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return CanineTokenizer.from_pretrained("""google/canine-s""" ) def _UpperCamelCase ( self : Optional[int] , **snake_case : List[str] ): '''simple docstring''' A__ : Union[str, Any] = self.tokenizer_class.from_pretrained(self.tmpdirname , **snake_case ) A__ : str = 1024 return tokenizer @require_torch def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = self.canine_tokenizer A__ : Optional[int] = ["""Life is like a box of chocolates.""", """You never know what you're gonna get."""] # fmt: off A__ : Tuple = [5_7344, 76, 105, 102, 101, 32, 105, 115, 32, 108, 105, 107, 101, 32, 97, 32, 98, 111, 120, 32, 111, 102, 32, 99, 104, 111, 99, 111, 108, 97, 116, 101, 115, 46, 5_7345, 0, 0, 0, 0] # fmt: on A__ : str = tokenizer(snake_case , padding=snake_case , return_tensors="""pt""" ) self.assertIsInstance(snake_case , snake_case ) A__ : List[Any] = list(batch.input_ids.numpy()[0] ) self.assertListEqual(snake_case , snake_case ) self.assertEqual((2, 39) , batch.input_ids.shape ) self.assertEqual((2, 39) , batch.attention_mask.shape ) @require_torch def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Dict = self.canine_tokenizer A__ : Optional[int] = ["""Once there was a man.""", """He wrote a test in HuggingFace Tranformers."""] A__ : List[str] = tokenizer(snake_case , padding=snake_case , return_tensors="""pt""" ) # check if input_ids, attention_mask and token_type_ids are returned self.assertIn("""input_ids""" , snake_case ) self.assertIn("""attention_mask""" , snake_case ) self.assertIn("""token_type_ids""" , snake_case ) @require_torch def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = self.canine_tokenizer A__ : Dict = [ """What's the weater?""", """It's about 25 degrees.""", ] A__ : Dict = tokenizer( text_target=snake_case , max_length=32 , padding="""max_length""" , truncation=snake_case , return_tensors="""pt""" ) self.assertEqual(32 , targets["""input_ids"""].shape[1] ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): self.assertNotEqual(tokenizer.model_max_length , 42 ) # Now let's start the test A__ : Tuple = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # Isolate this from the other tests because we save additional tokens/etc A__ : Tuple = tempfile.mkdtemp() A__ : Optional[int] = """ He is very happy, UNwant\u00E9d,running""" A__ : Tuple = tokenizer.encode(snake_case , add_special_tokens=snake_case ) tokenizer.save_pretrained(snake_case ) A__ : Optional[int] = tokenizer.__class__.from_pretrained(snake_case ) A__ : Tuple = after_tokenizer.encode(snake_case , add_special_tokens=snake_case ) self.assertListEqual(snake_case , snake_case ) shutil.rmtree(snake_case ) A__ : int = self.get_tokenizers(model_max_length=42 ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # Isolate this from the other tests because we save additional tokens/etc A__ : List[Any] = tempfile.mkdtemp() A__ : Optional[Any] = """ He is very happy, UNwant\u00E9d,running""" A__ : Optional[int] = tokenizer.additional_special_tokens # We can add a new special token for Canine as follows: A__ : Dict = chr(0xE007 ) additional_special_tokens.append(snake_case ) tokenizer.add_special_tokens({"""additional_special_tokens""": additional_special_tokens} ) A__ : int = tokenizer.encode(snake_case , add_special_tokens=snake_case ) tokenizer.save_pretrained(snake_case ) A__ : Optional[int] = tokenizer.__class__.from_pretrained(snake_case ) A__ : List[Any] = after_tokenizer.encode(snake_case , add_special_tokens=snake_case ) self.assertListEqual(snake_case , snake_case ) self.assertIn(snake_case , after_tokenizer.additional_special_tokens ) self.assertEqual(after_tokenizer.model_max_length , 42 ) A__ : Union[str, Any] = tokenizer.__class__.from_pretrained(snake_case , model_max_length=43 ) self.assertEqual(tokenizer.model_max_length , 43 ) shutil.rmtree(snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = self.get_tokenizers(do_lower_case=snake_case ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): A__ , A__ : Any = self.get_clean_sequence(snake_case ) # a special token for Canine can be defined as follows: A__ : int = 0xE005 A__ : Any = chr(snake_case ) tokenizer.add_special_tokens({"""cls_token""": special_token} ) A__ : str = tokenizer.encode(snake_case , add_special_tokens=snake_case ) self.assertEqual(len(snake_case ) , 1 ) A__ : Optional[int] = tokenizer.decode(ids + encoded_special_token , clean_up_tokenization_spaces=snake_case ) A__ : Optional[Any] = tokenizer.encode(snake_case , add_special_tokens=snake_case ) A__ : Dict = tokenizer.encode(snake_case , add_special_tokens=snake_case ) A__ : List[Any] = tokenizer.encode(snake_case , add_special_tokens=snake_case ) self.assertEqual(snake_case , input_encoded + special_token_id ) A__ : Dict = tokenizer.decode(snake_case , skip_special_tokens=snake_case ) self.assertTrue(special_token not in decoded ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.get_tokenizers(do_lower_case=snake_case ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): A__ : Optional[int] = chr(0xE005 ) A__ : Any = chr(0xE006 ) # `add_tokens` method stores special tokens only in `tokenizer.unique_no_split_tokens`. (in tokenization_utils.py) tokenizer.add_tokens([SPECIAL_TOKEN_1] , special_tokens=snake_case ) # `add_special_tokens` method stores special tokens in `tokenizer.additional_special_tokens`, # which also occur in `tokenizer.all_special_tokens`. (in tokenization_utils_base.py) tokenizer.add_special_tokens({"""additional_special_tokens""": [SPECIAL_TOKEN_2]} ) A__ : Optional[int] = tokenizer.tokenize(snake_case ) A__ : Union[str, Any] = tokenizer.tokenize(snake_case ) self.assertEqual(len(snake_case ) , 1 ) self.assertEqual(len(snake_case ) , 1 ) self.assertEqual(token_a[0] , snake_case ) self.assertEqual(token_a[0] , snake_case ) @require_tokenizers def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.get_tokenizers(do_lower_case=snake_case ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): # a special token for Canine can be defined as follows: A__ : List[str] = 0xE006 A__ : int = chr(snake_case ) A__ : str = AddedToken(snake_case , lstrip=snake_case ) tokenizer.add_special_tokens({"""additional_special_tokens""": [new_token]} ) with tempfile.TemporaryDirectory() as tmp_dir_name: tokenizer.save_pretrained(snake_case ) tokenizer.from_pretrained(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = [] if self.test_slow_tokenizer: tokenizer_list.append((self.tokenizer_class, self.get_tokenizer()) ) if self.test_rust_tokenizer: tokenizer_list.append((self.rust_tokenizer_class, self.get_rust_tokenizer()) ) for tokenizer_class, tokenizer_utils in tokenizer_list: with tempfile.TemporaryDirectory() as tmp_dir: tokenizer_utils.save_pretrained(snake_case ) with open(os.path.join(snake_case , """special_tokens_map.json""" ) , encoding="""utf-8""" ) as json_file: A__ : Union[str, Any] = json.load(snake_case ) with open(os.path.join(snake_case , """tokenizer_config.json""" ) , encoding="""utf-8""" ) as json_file: A__ : Dict = json.load(snake_case ) # a special token for Canine can be defined as follows: A__ : Dict = 0xE006 A__ : Dict = chr(snake_case ) A__ : Dict = [new_token_a] A__ : List[str] = [new_token_a] with open(os.path.join(snake_case , """special_tokens_map.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(snake_case , snake_case ) with open(os.path.join(snake_case , """tokenizer_config.json""" ) , """w""" , encoding="""utf-8""" ) as outfile: json.dump(snake_case , snake_case ) # the following checks allow us to verify that our test works as expected, i.e. that the tokenizer takes # into account the new value of additional_special_tokens given in the "tokenizer_config.json" and # "special_tokens_map.json" files A__ : Optional[int] = tokenizer_class.from_pretrained(snake_case , extra_ids=0 ) self.assertIn(snake_case , tokenizer_without_change_in_init.additional_special_tokens ) # self.assertIn("an_additional_special_token",tokenizer_without_change_in_init.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer_without_change_in_init.convert_ids_to_tokens( tokenizer_without_change_in_init.convert_tokens_to_ids([new_token_a] ) ) , ) A__ : Any = 0xE007 A__ : List[Any] = chr(snake_case ) # Now we test that we can change the value of additional_special_tokens in the from_pretrained A__ : List[str] = [AddedToken(snake_case , lstrip=snake_case )] A__ : List[str] = tokenizer_class.from_pretrained( snake_case , additional_special_tokens=snake_case , extra_ids=0 ) self.assertIn(snake_case , tokenizer.additional_special_tokens ) # self.assertIn(new_token_2,tokenizer.get_vocab()) # ByT5Tokenization no vocab self.assertEqual( [new_token_a] , tokenizer.convert_ids_to_tokens(tokenizer.convert_tokens_to_ids([new_token_a] ) ) ) @require_tokenizers def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Optional[int] = self.get_tokenizers(do_lower_case=snake_case ) for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): A__ : Optional[int] = """hello world""" if self.space_between_special_tokens: A__ : List[Any] = """[CLS] hello world [SEP]""" else: A__ : Any = input A__ : int = tokenizer.encode(snake_case , add_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(snake_case , spaces_between_special_tokens=self.space_between_special_tokens ) self.assertIn(snake_case , [output, output.lower()] ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Tuple = self.get_tokenizers() for tokenizer in tokenizers: with self.subTest(F'{tokenizer.__class__.__name__}' ): A__ : Union[str, Any] = [ """bos_token""", """eos_token""", """unk_token""", """sep_token""", """pad_token""", """cls_token""", """mask_token""", ] A__ : Tuple = """a""" A__ : List[Any] = ord(snake_case ) for attr in attributes_list: setattr(snake_case , attr + """_id""" , snake_case ) self.assertEqual(getattr(snake_case , snake_case ) , snake_case ) self.assertEqual(getattr(snake_case , attr + """_id""" ) , snake_case ) setattr(snake_case , attr + """_id""" , snake_case ) self.assertEqual(getattr(snake_case , snake_case ) , snake_case ) self.assertEqual(getattr(snake_case , attr + """_id""" ) , snake_case ) setattr(snake_case , """additional_special_tokens_ids""" , [] ) self.assertListEqual(getattr(snake_case , """additional_special_tokens""" ) , [] ) self.assertListEqual(getattr(snake_case , """additional_special_tokens_ids""" ) , [] ) A__ : Tuple = 0xE006 A__ : Dict = chr(snake_case ) setattr(snake_case , """additional_special_tokens_ids""" , [additional_special_token_id] ) self.assertListEqual(getattr(snake_case , """additional_special_tokens""" ) , [additional_special_token] ) self.assertListEqual(getattr(snake_case , """additional_special_tokens_ids""" ) , [additional_special_token_id] ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _UpperCamelCase ( self : Dict ): '''simple docstring''' pass def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass def _UpperCamelCase ( self : Any ): '''simple docstring''' pass def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass def _UpperCamelCase ( self : Any ): '''simple docstring''' pass def _UpperCamelCase ( self : int ): '''simple docstring''' pass def _UpperCamelCase ( self : Any ): '''simple docstring''' pass
296
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
1
"""simple docstring""" import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class __SCREAMING_SNAKE_CASE : def __init__( self : str , snake_case : List[Any] , snake_case : int=13 , snake_case : List[str]=2 , snake_case : List[Any]=24 , snake_case : List[Any]=16 , snake_case : List[Any]=True , snake_case : Dict=True , snake_case : Optional[int]=32 , snake_case : int=5 , snake_case : Optional[int]=4 , snake_case : Any=37 , snake_case : Optional[Any]="gelu" , snake_case : Optional[int]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=10 , snake_case : Optional[Any]=0.02 , snake_case : Any=None , snake_case : Tuple=2 , snake_case : Any=2 , ): '''simple docstring''' A__ : Optional[Any] = parent A__ : List[Any] = batch_size A__ : Tuple = patch_size A__ : int = max_length A__ : Union[str, Any] = num_mel_bins A__ : List[str] = is_training A__ : int = use_labels A__ : str = hidden_size A__ : int = num_hidden_layers A__ : List[Any] = num_attention_heads A__ : Any = intermediate_size A__ : Tuple = hidden_act A__ : Dict = hidden_dropout_prob A__ : List[Any] = attention_probs_dropout_prob A__ : Dict = type_sequence_label_size A__ : int = initializer_range A__ : str = scope A__ : Optional[Any] = frequency_stride A__ : Tuple = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) A__ : Tuple = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 A__ : Optional[int] = (self.max_length - self.patch_size) // self.time_stride + 1 A__ : Optional[int] = frequency_out_dimension * time_out_dimension A__ : int = num_patches + 2 def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Union[str, Any] = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) A__ : List[str] = None if self.use_labels: A__ : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Any = self.get_config() return config, input_values, labels def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[int] , snake_case : Any , snake_case : List[Any] ): '''simple docstring''' A__ : Optional[Any] = ASTModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Any = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ) : Dict = config_and_inputs A__ : List[str] = {"""input_values""": input_values} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) snake_case_ = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Tuple , snake_case : Any , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Optional[Any] = ASTModelTester(self ) A__ : Tuple = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' self.config_tester.run_common_tests() @unittest.skip(reason="""AST does not use inputs_embeds""" ) def _UpperCamelCase ( self : int ): '''simple docstring''' pass def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Tuple = model_class(snake_case ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) A__ : Dict = model.get_output_embeddings() self.assertTrue(x is None or isinstance(snake_case , nn.Linear ) ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Any = model_class(snake_case ) A__ : Union[str, Any] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Dict = [*signature.parameters.keys()] A__ : str = ["""input_values"""] self.assertListEqual(arg_names[:1] , snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) @slow def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = ASTModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _lowerCAmelCase ( ) ->Optional[int]: A__ : str = hf_hub_download( repo_id="""nielsr/audio-spectogram-transformer-checkpoint""", filename="""sample_audio.flac""", repo_type="""dataset""" ) A__ , A__ : List[Any] = torchaudio.load(UpperCAmelCase__ ) return audio, sampling_rate @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _UpperCamelCase ( self : int ): '''simple docstring''' return ( ASTFeatureExtractor.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ) if is_torchaudio_available() else None ) @slow def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[Any] = self.default_feature_extractor A__ : Any = ASTForAudioClassification.from_pretrained("""MIT/ast-finetuned-audioset-10-10-0.4593""" ).to(snake_case ) A__ : List[Any] = self.default_feature_extractor A__ , A__ : Optional[Any] = prepare_audio() A__ : str = audio.squeeze().numpy() A__ : Any = feature_extractor(snake_case , sampling_rate=snake_case , return_tensors="""pt""" ).to(snake_case ) # forward pass with torch.no_grad(): A__ : Any = model(**snake_case ) # verify the logits A__ : Tuple = torch.Size((1, 527) ) self.assertEqual(outputs.logits.shape , snake_case ) A__ : Dict = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(snake_case ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case , atol=1e-4 ) )
296
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels A_ = object() # For specifying empty leaf dict `{}` A_ = object() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any] ) ->Dict: A__ : Union[str, Any] = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) + 1 ): A__ : Optional[Any] = [x.match(UpperCAmelCase__ ) for x, y in zip(UpperCAmelCase__, ks[i:] )] if matches and all(UpperCAmelCase__ ): return True return False def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Dict: def replace(UpperCAmelCase__ : int, UpperCAmelCase__ : List[str] ): for rule, replacement in rules: if _match(UpperCAmelCase__, UpperCAmelCase__ ): return replacement return val return replace def _lowerCAmelCase ( ) ->Tuple: return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""", UpperCAmelCase__ )), (("transformer", "wte", "embedding"), P("""mp""", UpperCAmelCase__ )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(UpperCAmelCase__, """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(UpperCAmelCase__, """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : Union[str, Any] = _get_partition_rules() A__ : int = _replacement_rules(UpperCAmelCase__ ) A__ : Tuple = {k: _unmatched for k in flatten_dict(UpperCAmelCase__ )} A__ : Optional[int] = {k: replace(UpperCAmelCase__, UpperCAmelCase__ ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(UpperCAmelCase__ ) )
296
1
"""simple docstring""" from typing import List from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''snap-research/efficientformer-l1-300''': ( '''https://huggingface.co/snap-research/efficientformer-l1-300/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'efficientformer' def __init__( self : Optional[int] , snake_case : List[int] = [3, 2, 6, 4] , snake_case : List[int] = [48, 96, 224, 448] , snake_case : List[bool] = [True, True, True, True] , snake_case : int = 448 , snake_case : int = 32 , snake_case : int = 4 , snake_case : int = 7 , snake_case : int = 5 , snake_case : int = 8 , snake_case : int = 4 , snake_case : float = 0.0 , snake_case : int = 16 , snake_case : int = 3 , snake_case : int = 3 , snake_case : int = 3 , snake_case : int = 2 , snake_case : int = 1 , snake_case : float = 0.0 , snake_case : int = 1 , snake_case : bool = True , snake_case : bool = True , snake_case : float = 1e-5 , snake_case : str = "gelu" , snake_case : float = 0.02 , snake_case : float = 1e-12 , snake_case : int = 224 , snake_case : float = 1e-05 , **snake_case : Any , ): '''simple docstring''' super().__init__(**snake_case ) A__ : Optional[Any] = hidden_act A__ : Dict = hidden_dropout_prob A__ : int = hidden_sizes A__ : Tuple = num_hidden_layers A__ : int = num_attention_heads A__ : Tuple = initializer_range A__ : Tuple = layer_norm_eps A__ : Tuple = patch_size A__ : List[Any] = num_channels A__ : Optional[int] = depths A__ : List[str] = mlp_expansion_ratio A__ : str = downsamples A__ : Tuple = dim A__ : Tuple = key_dim A__ : str = attention_ratio A__ : Optional[int] = resolution A__ : str = pool_size A__ : Any = downsample_patch_size A__ : Tuple = downsample_stride A__ : Tuple = downsample_pad A__ : List[Any] = drop_path_rate A__ : List[Any] = num_metaad_blocks A__ : Optional[int] = distillation A__ : List[Any] = use_layer_scale A__ : str = layer_scale_init_value A__ : Optional[Any] = image_size A__ : List[Any] = batch_norm_eps
296
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] , snake_case : Tuple , snake_case : List[str]=2 , snake_case : List[str]=8 , snake_case : List[Any]=True , snake_case : Optional[Any]=True , snake_case : List[Any]=True , snake_case : Dict=True , snake_case : Tuple=99 , snake_case : Dict=16 , snake_case : Dict=5 , snake_case : int=2 , snake_case : Any=36 , snake_case : str="gelu" , snake_case : Dict=0.0 , snake_case : List[Any]=0.0 , snake_case : int=512 , snake_case : List[Any]=16 , snake_case : Tuple=2 , snake_case : Any=0.02 , snake_case : Optional[Any]=3 , snake_case : List[Any]=4 , snake_case : str=None , ): '''simple docstring''' A__ : Union[str, Any] = parent A__ : Optional[Any] = batch_size A__ : Dict = seq_length A__ : str = is_training A__ : Tuple = use_input_mask A__ : Dict = use_token_type_ids A__ : Dict = use_labels A__ : int = vocab_size A__ : List[str] = hidden_size A__ : Union[str, Any] = num_hidden_layers A__ : int = num_attention_heads A__ : List[str] = intermediate_size A__ : int = hidden_act A__ : str = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : Any = max_position_embeddings A__ : Optional[int] = type_vocab_size A__ : int = type_sequence_label_size A__ : Optional[Any] = initializer_range A__ : int = num_labels A__ : Optional[int] = num_choices A__ : Optional[int] = scope def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Any = None if self.use_input_mask: A__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Optional[int] = None if self.use_token_type_ids: A__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Dict = None A__ : List[str] = None A__ : Union[str, Any] = None if self.use_labels: A__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Any = ids_tensor([self.batch_size] , self.num_choices ) A__ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.get_config() A__ : List[str] = 300 return config def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Tuple = self.prepare_config_and_inputs() A__ : List[str] = True A__ : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _UpperCamelCase ( self : Any , snake_case : Any , snake_case : Tuple , snake_case : Any , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Dict ): '''simple docstring''' A__ : List[str] = MraModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A__ : List[str] = model(snake_case , token_type_ids=snake_case ) A__ : Union[str, Any] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : List[Any] , snake_case : Any , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Dict , snake_case : str , snake_case : Dict , snake_case : str , ): '''simple docstring''' A__ : Dict = True A__ : Optional[Any] = MraModel(snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , ) A__ : Optional[int] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : str , snake_case : Union[str, Any] , snake_case : Dict , snake_case : List[str] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Dict , snake_case : Dict , snake_case : Dict , snake_case : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : Dict = MraForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : Tuple , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Optional[int] , snake_case : List[str] , snake_case : Union[str, Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Optional[Any] = MraForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict , snake_case : str , snake_case : List[Any] , snake_case : Any , snake_case : Dict , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Union[str, Any] = MraForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : Tuple , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Dict , snake_case : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.num_choices A__ : str = MraForMultipleChoice(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Dict = config_and_inputs A__ : Optional[int] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = () def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Optional[Any] = MraModelTester(self ) A__ : List[str] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : List[str] = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : str = MraModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip(reason="""MRA does not output attentions""" ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Any = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : List[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , snake_case ) A__ : int = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Dict = 5_0265 A__ : List[str] = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : List[Any] = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) A__ : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Union[str, Any] = 5_0265 A__ : Optional[Any] = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : Optional[int] = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) )
296
1
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig, is_flax_available from transformers.testing_utils import require_flax, slow from transformers.utils import cached_property, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_flax_common import FlaxModelTesterMixin, floats_tensor if is_flax_available(): import jax import jax.numpy as jnp from transformers.models.regnet.modeling_flax_regnet import FlaxRegNetForImageClassification, FlaxRegNetModel if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Any , snake_case : Tuple , snake_case : Dict=3 , snake_case : Optional[Any]=32 , snake_case : str=3 , snake_case : Dict=10 , snake_case : int=[10, 20, 30, 40] , snake_case : Dict=[1, 1, 2, 1] , snake_case : Tuple=True , snake_case : Tuple=True , snake_case : List[str]="relu" , snake_case : Optional[Any]=3 , snake_case : Any=None , ): '''simple docstring''' A__ : List[Any] = parent A__ : Any = batch_size A__ : int = image_size A__ : Union[str, Any] = num_channels A__ : str = embeddings_size A__ : Optional[int] = hidden_sizes A__ : Any = depths A__ : str = is_training A__ : Dict = use_labels A__ : Optional[int] = hidden_act A__ : Optional[Any] = num_labels A__ : Optional[Any] = scope A__ : str = len(snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) A__ : Optional[Any] = self.get_config() return config, pixel_values def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , image_size=self.image_size , ) def _UpperCamelCase ( self : Tuple , snake_case : Optional[Any] , snake_case : int ): '''simple docstring''' A__ : List[str] = FlaxRegNetModel(config=snake_case ) A__ : List[Any] = model(snake_case ) # Output shape (b, c, h, w) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : List[Any] , snake_case : Dict ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : List[Any] = FlaxRegNetForImageClassification(config=snake_case ) A__ : List[Any] = model(snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Tuple = self.prepare_config_and_inputs() A__ , A__ : List[str] = config_and_inputs A__ : List[Any] = {"""pixel_values""": pixel_values} return config, inputs_dict @require_flax class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = (FlaxRegNetModel, FlaxRegNetForImageClassification) if is_flax_available() else () snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Dict = FlaxRegNetModelTester(self ) A__ : Optional[Any] = ConfigTester(self , config_class=snake_case , has_text_modality=snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def _UpperCamelCase ( self : Dict ): '''simple docstring''' return def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case ) @unittest.skip(reason="""RegNet does not use inputs_embeds""" ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' pass @unittest.skip(reason="""RegNet does not support input and output embeddings""" ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Any = model_class(snake_case ) A__ : str = inspect.signature(model.__call__ ) # signature.parameters is an OrderedDict => so arg_names order is deterministic A__ : Union[str, Any] = [*signature.parameters.keys()] A__ : Any = ["""pixel_values"""] self.assertListEqual(arg_names[:1] , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' def check_hidden_states_output(snake_case : str , snake_case : Dict , snake_case : str ): A__ : Optional[int] = model_class(snake_case ) A__ : Optional[Any] = model(**self._prepare_for_class(snake_case , snake_case ) ) A__ : int = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states A__ : List[str] = self.model_tester.num_stages self.assertEqual(len(snake_case ) , expected_num_stages + 1 ) A__ , A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: A__ : Dict = True check_hidden_states_output(snake_case , snake_case , snake_case ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] A__ : Tuple = True check_hidden_states_output(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: with self.subTest(model_class.__name__ ): A__ : str = self._prepare_for_class(snake_case , snake_case ) A__ : Optional[int] = model_class(snake_case ) @jax.jit def model_jitted(snake_case : List[str] , **snake_case : Dict ): return model(pixel_values=snake_case , **snake_case ) with self.subTest("""JIT Enabled""" ): A__ : str = model_jitted(**snake_case ).to_tuple() with self.subTest("""JIT Disabled""" ): with jax.disable_jit(): A__ : Optional[int] = model_jitted(**snake_case ).to_tuple() self.assertEqual(len(snake_case ) , len(snake_case ) ) for jitted_output, output in zip(snake_case , snake_case ): self.assertEqual(jitted_output.shape , output.shape ) def _lowerCAmelCase ( ) ->Dict: A__ : List[str] = Image.open("""./tests/fixtures/tests_samples/COCO/000000039769.png""" ) return image @require_flax class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @cached_property def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return AutoImageProcessor.from_pretrained("""facebook/regnet-y-040""" ) if is_vision_available() else None @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Union[str, Any] = FlaxRegNetForImageClassification.from_pretrained("""facebook/regnet-y-040""" ) A__ : int = self.default_image_processor A__ : Union[str, Any] = prepare_img() A__ : Optional[int] = image_processor(images=snake_case , return_tensors="""np""" ) A__ : Dict = model(**snake_case ) # verify the logits A__ : str = (1, 1000) self.assertEqual(outputs.logits.shape , snake_case ) A__ : int = jnp.array([-0.4180, -1.5051, -3.4836] ) self.assertTrue(jnp.allclose(outputs.logits[0, :3] , snake_case , atol=1e-4 ) )
296
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets A_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' A_ = '''\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. ''' A_ = ''' Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {"raw_values", "uniform_average"} or array-like of shape (n_outputs,), default="uniform_average" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. "raw_values" : Returns a full set of errors in case of multioutput input. "uniform_average" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric("mse") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {\'mse\': 0.6123724356957945} If you\'re using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric("mse", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput=\'raw_values\') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mse\': array([0.41666667, 1. ])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : Dict ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html""" ] , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("""float""" ) ), "references": datasets.Sequence(datasets.Value("""float""" ) ), } else: return { "predictions": datasets.Value("""float""" ), "references": datasets.Value("""float""" ), } def _UpperCamelCase ( self : List[str] , snake_case : Dict , snake_case : List[Any] , snake_case : List[str]=None , snake_case : List[Any]="uniform_average" , snake_case : int=True ): '''simple docstring''' A__ : Optional[int] = mean_squared_error( snake_case , snake_case , sample_weight=snake_case , multioutput=snake_case , squared=snake_case ) return {"mse": mse}
296
1
"""simple docstring""" import copy from typing import TYPE_CHECKING, Any, Mapping, Optional, OrderedDict from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto.configuration_auto import AutoConfig if TYPE_CHECKING: from ... import PreTrainedTokenizerBase, TensorType A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'vision-encoder-decoder' snake_case_ = True def __init__( self : Optional[int] , **snake_case : Tuple ): '''simple docstring''' super().__init__(**snake_case ) if "encoder" not in kwargs or "decoder" not in kwargs: raise ValueError( F'A configuraton of type {self.model_type} cannot be instantiated because ' F'not both `encoder` and `decoder` sub-configurations are passed, but only {kwargs}' ) A__ : Union[str, Any] = kwargs.pop("""encoder""" ) A__ : Tuple = encoder_config.pop("""model_type""" ) A__ : Optional[Any] = kwargs.pop("""decoder""" ) A__ : List[str] = decoder_config.pop("""model_type""" ) A__ : str = AutoConfig.for_model(snake_case , **snake_case ) A__ : Optional[Any] = AutoConfig.for_model(snake_case , **snake_case ) A__ : int = True @classmethod def _UpperCamelCase ( cls : str , snake_case : PretrainedConfig , snake_case : PretrainedConfig , **snake_case : str ): '''simple docstring''' logger.info("""Setting `config.is_decoder=True` and `config.add_cross_attention=True` for decoder_config""" ) A__ : Tuple = True A__ : Tuple = True return cls(encoder=encoder_config.to_dict() , decoder=decoder_config.to_dict() , **snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = copy.deepcopy(self.__dict__ ) A__ : Union[str, Any] = self.encoder.to_dict() A__ : Tuple = self.decoder.to_dict() A__ : str = self.__class__.model_type return output class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ] ) @property def _UpperCamelCase ( self : str ): '''simple docstring''' return 1e-4 @property def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return OrderedDict({"""last_hidden_state""": {0: """batch""", 1: """encoder_sequence"""}} ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): @property def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = OrderedDict() A__ : str = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A__ : Dict = {0: """batch""", 1: """past_decoder_sequence + sequence"""} A__ : List[str] = {0: """batch""", 1: """encoder_sequence"""} return common_inputs def _UpperCamelCase ( self : Tuple , snake_case : "PreTrainedTokenizerBase" , snake_case : int = -1 , snake_case : int = -1 , snake_case : bool = False , snake_case : Optional["TensorType"] = None , ): '''simple docstring''' import torch A__ : List[Any] = OrderedDict() A__ : int = super().generate_dummy_inputs( snake_case , batch_size=snake_case , seq_length=snake_case , is_pair=snake_case , framework=snake_case ) A__ , A__ : int = dummy_input["""input_ids"""].shape A__ : str = (batch, encoder_sequence, self._config.encoder_hidden_size) A__ : Union[str, Any] = dummy_input.pop("""input_ids""" ) A__ : Any = dummy_input.pop("""attention_mask""" ) A__ : int = torch.zeros(snake_case ) return common_inputs class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass def _UpperCamelCase ( self : str , snake_case : PretrainedConfig ): '''simple docstring''' return VisionEncoderDecoderEncoderOnnxConfig(snake_case ) def _UpperCamelCase ( self : Optional[Any] , snake_case : PretrainedConfig , snake_case : PretrainedConfig , snake_case : str = "default" ): '''simple docstring''' A__ : List[str] = encoder_config.hidden_size return VisionEncoderDecoderDecoderOnnxConfig(snake_case , snake_case )
296
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Optional[int] , snake_case : List[str]=None , **snake_case : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , snake_case , ) super().__init__(args=snake_case , **snake_case )
296
1
"""simple docstring""" from PIL import Image def _lowerCAmelCase ( UpperCAmelCase__ : Image, UpperCAmelCase__ : float ) ->Image: def brightness(UpperCAmelCase__ : int ) -> float: return 1_2_8 + level + (c - 1_2_8) if not -255.0 <= level <= 255.0: raise ValueError("""level must be between -255.0 (black) and 255.0 (white)""" ) return img.point(UpperCAmelCase__ ) if __name__ == "__main__": # Load image with Image.open('''image_data/lena.jpg''') as img: # Change brightness to 100 A_ = change_brightness(img, 100) brigt_img.save('''image_data/lena_brightness.png''', format='''png''')
296
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch A_ = random.Random() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Tuple=1.0, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : str=None ) ->Union[str, Any]: if rng is None: A__ : Optional[int] = global_rng A__ : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[str]=7 , snake_case : str=400 , snake_case : Optional[Any]=2000 , snake_case : Union[str, Any]=10 , snake_case : str=160 , snake_case : List[str]=8 , snake_case : List[Any]=0.0 , snake_case : Optional[Any]=4000 , snake_case : Any=False , snake_case : int=True , ): '''simple docstring''' A__ : Any = parent A__ : str = batch_size A__ : List[str] = min_seq_length A__ : Dict = max_seq_length A__ : str = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) A__ : Dict = padding_value A__ : Optional[Any] = sampling_rate A__ : Any = return_attention_mask A__ : Optional[int] = do_normalize A__ : Tuple = feature_size A__ : Optional[Any] = chunk_length A__ : Union[str, Any] = hop_length def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict=False , snake_case : Optional[Any]=False ): '''simple docstring''' def _flatten(snake_case : Dict ): return list(itertools.chain(*snake_case ) ) if equal_length: A__ : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size A__ : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: A__ : List[str] = [np.asarray(snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = WhisperFeatureExtractor if is_speech_available() else None def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : str = WhisperFeatureExtractionTester(self ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] = feat_extract_first.save_pretrained(snake_case )[0] check_json_file_has_correct_format(snake_case ) A__ : Union[str, Any] = self.feature_extraction_class.from_pretrained(snake_case ) A__ : str = feat_extract_first.to_dict() A__ : Union[str, Any] = feat_extract_second.to_dict() A__ : List[Any] = feat_extract_first.mel_filters A__ : Optional[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : Any = os.path.join(snake_case , """feat_extract.json""" ) feat_extract_first.to_json_file(snake_case ) A__ : int = self.feature_extraction_class.from_json_file(snake_case ) A__ : Dict = feat_extract_first.to_dict() A__ : str = feat_extract_second.to_dict() A__ : str = feat_extract_first.mel_filters A__ : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 A__ : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] # Test feature size A__ : Dict = feature_extractor(snake_case , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input A__ : str = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test batched A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. A__ : Tuple = [floats_list((1, x) )[0] for x in (800, 800, 800)] A__ : str = np.asarray(snake_case ) A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test truncation required A__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] A__ : Union[str, Any] = [x[: feature_extractor.n_samples] for x in speech_inputs] A__ : str = [np.asarray(snake_case ) for speech_input in speech_inputs_truncated] A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : str = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' import torch A__ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : List[str] = np.random.rand(100 , 32 ).astype(np.floataa ) A__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: A__ : Optional[Any] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) A__ : Optional[int] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' A__ : int = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech A__ : Union[str, Any] = ds.sort("""id""" ).select(range(snake_case ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on A__ : Optional[Any] = self._load_datasamples(1 ) A__ : Union[str, Any] = WhisperFeatureExtractor() A__ : List[str] = feature_extractor(snake_case , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , snake_case , atol=1e-4 ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : Union[str, Any] = self._load_datasamples(1 )[0] A__ : Any = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue A__ : str = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=snake_case )[0] self.assertTrue(np.all(np.mean(snake_case ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case ) - 1 ) < 1e-3 ) )
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_0**1_2 ) ->int: A__ : int = 1 A__ : Union[str, Any] = 0 A__ : List[str] = 1 A__ : Any = 1 while numerator <= 2 * min_total - 1: prev_numerator += 2 * numerator numerator += 2 * prev_numerator prev_denominator += 2 * denominator denominator += 2 * prev_denominator return (denominator + 1) // 2 if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] ): '''simple docstring''' A__ : Optional[int] = (0, 0) A__ : Dict = None A__ : int = 0 A__ : str = 0 A__ : Optional[Any] = 0 def __eq__( self : str , snake_case : Optional[int] ): '''simple docstring''' return self.position == cell.position def _UpperCamelCase ( self : List[str] ): '''simple docstring''' print(self.position ) class __SCREAMING_SNAKE_CASE : def __init__( self : int , snake_case : Any=(5, 5) ): '''simple docstring''' A__ : Optional[int] = np.zeros(snake_case ) A__ : List[Any] = world_size[0] A__ : Dict = world_size[1] def _UpperCamelCase ( self : Any ): '''simple docstring''' print(self.w ) def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : int = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] A__ : int = cell.position[0] A__ : str = cell.position[1] A__ : Any = [] for n in neughbour_cord: A__ : List[Any] = current_x + n[0] A__ : Tuple = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: A__ : List[Any] = Cell() A__ : str = (x, y) A__ : Optional[Any] = cell neighbours.append(snake_case ) return neighbours def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Dict ) ->Dict: A__ : Union[str, Any] = [] A__ : Optional[int] = [] _open.append(UpperCAmelCase__ ) while _open: A__ : List[Any] = np.argmin([n.f for n in _open] ) A__ : Union[str, Any] = _open[min_f] _closed.append(_open.pop(UpperCAmelCase__ ) ) if current == goal: break for n in world.get_neigbours(UpperCAmelCase__ ): for c in _closed: if c == n: continue A__ : Dict = current.g + 1 A__ , A__ : int = n.position A__ , A__ : Optional[int] = goal.position A__ : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 A__ : Optional[int] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(UpperCAmelCase__ ) A__ : List[str] = [] while current.parent is not None: path.append(current.position ) A__ : Union[str, Any] = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": A_ = Gridworld() # Start position and goal A_ = Cell() A_ = (0, 0) A_ = Cell() A_ = (4, 4) print(F'path from {start.position} to {goal.position}') A_ = astar(world, start, goal) # Just for visual reasons. for i in s: A_ = 1 print(world.w)
296
1
"""simple docstring""" import logging from dataclasses import dataclass, field from typing import Optional from seqaseq_trainer import arg_to_scheduler from transformers import TrainingArguments A_ = logging.getLogger(__name__) @dataclass class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = field( default=0.0 , metadata={'help': 'The label smoothing epsilon to apply (if not zero).'} ) snake_case_ = field(default=UpperCamelCase , metadata={'help': 'Whether to SortishSamler or not.'} ) snake_case_ = field( default=UpperCamelCase , metadata={'help': 'Whether to use generate to calculate generative metrics (ROUGE, BLEU).'} ) snake_case_ = field(default=UpperCamelCase , metadata={'help': 'whether to use adafactor'} ) snake_case_ = field( default=UpperCamelCase , metadata={'help': 'Encoder layer dropout probability. Goes into model.config.'} ) snake_case_ = field( default=UpperCamelCase , metadata={'help': 'Decoder layer dropout probability. Goes into model.config.'} ) snake_case_ = field(default=UpperCamelCase , metadata={'help': 'Dropout probability. Goes into model.config.'} ) snake_case_ = field( default=UpperCamelCase , metadata={'help': 'Attention dropout probability. Goes into model.config.'} ) snake_case_ = field( default='linear' , metadata={'help': F"Which lr scheduler to use. Selected in {sorted(arg_to_scheduler.keys() )}"} , )
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple=False ) ->str: A__ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ : Optional[int] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]=False ) ->str: for i in range(config.num_hidden_layers ): if base_model: A__ : Any = """""" else: A__ : Tuple = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A__ : str = in_proj_bias[: config.hidden_size] A__ : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A__ : Any = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : int = dct.pop(UpperCAmelCase__ ) A__ : Tuple = val def _lowerCAmelCase ( ) ->List[Any]: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Any ) ->Tuple: A__ : List[Any] = DeiTConfig() # all deit models have fine-tuned heads A__ : Tuple = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ : str = 1_0_0_0 A__ : List[str] = """huggingface/label-files""" A__ : Dict = """imagenet-1k-id2label.json""" A__ : List[str] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Dict = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[int] = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} A__ : List[str] = int(deit_name[-6:-4] ) A__ : str = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): A__ : List[str] = 1_9_2 A__ : int = 7_6_8 A__ : List[Any] = 1_2 A__ : Dict = 3 elif deit_name[9:].startswith("""small""" ): A__ : List[Any] = 3_8_4 A__ : List[str] = 1_5_3_6 A__ : Any = 1_2 A__ : Union[str, Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): A__ : int = 1_0_2_4 A__ : str = 4_0_9_6 A__ : Any = 2_4 A__ : int = 1_6 # load original model from timm A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ : Tuple = timm_model.state_dict() A__ : str = create_rename_keys(UpperCAmelCase__, UpperCAmelCase__ ) for src, dest in rename_keys: rename_key(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : str = DeiTForImageClassificationWithTeacher(UpperCAmelCase__ ).eval() model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor A__ : int = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ : Any = DeiTImageProcessor(size=UpperCAmelCase__, crop_size=config.image_size ) A__ : Union[str, Any] = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : Optional[Any] = encoding["""pixel_values"""] A__ : Union[str, Any] = model(UpperCAmelCase__ ) A__ : Union[str, Any] = timm_model(UpperCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCAmelCase__, outputs.logits, atol=1e-3 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
296
1
"""simple docstring""" from collections.abc import Callable class __SCREAMING_SNAKE_CASE : def __init__( self : List[Any] , snake_case : Callable | None = None ): '''simple docstring''' A__ : list = [] # Stores indexes of each item for supporting updates and deletion. A__ : dict = {} # Stores current size of heap. A__ : Optional[int] = 0 # Stores function used to evaluate the score of an item on which basis ordering # will be done. A__ : List[str] = key or (lambda snake_case : x) def _UpperCamelCase ( self : Union[str, Any] , snake_case : int ): '''simple docstring''' return int((i - 1) / 2 ) if i > 0 else None def _UpperCamelCase ( self : List[Any] , snake_case : int ): '''simple docstring''' A__ : List[Any] = int(2 * i + 1 ) return left if 0 < left < self.size else None def _UpperCamelCase ( self : Any , snake_case : int ): '''simple docstring''' A__ : Dict = int(2 * i + 2 ) return right if 0 < right < self.size else None def _UpperCamelCase ( self : List[Any] , snake_case : int , snake_case : int ): '''simple docstring''' A__ , A__ : Tuple = ( self.pos_map[self.arr[j][0]], self.pos_map[self.arr[i][0]], ) # Then swap the items in the list. A__ , A__ : Optional[Any] = self.arr[j], self.arr[i] def _UpperCamelCase ( self : Tuple , snake_case : int , snake_case : int ): '''simple docstring''' return self.arr[i][1] < self.arr[j][1] def _UpperCamelCase ( self : List[str] , snake_case : int ): '''simple docstring''' A__ : Optional[int] = self._left(snake_case ) A__ : List[Any] = self._right(snake_case ) A__ : Dict = i if left is not None and not self._cmp(snake_case , snake_case ): A__ : List[str] = left if right is not None and not self._cmp(snake_case , snake_case ): A__ : Tuple = right return valid_parent def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : Dict = self._parent(snake_case ) while parent is not None and not self._cmp(snake_case , snake_case ): self._swap(snake_case , snake_case ) A__ , A__ : Optional[Any] = parent, self._parent(snake_case ) def _UpperCamelCase ( self : Dict , snake_case : int ): '''simple docstring''' A__ : str = self._get_valid_parent(snake_case ) while valid_parent != index: self._swap(snake_case , snake_case ) A__ , A__ : Dict = valid_parent, self._get_valid_parent(snake_case ) def _UpperCamelCase ( self : Tuple , snake_case : int , snake_case : int ): '''simple docstring''' if item not in self.pos_map: return A__ : Optional[int] = self.pos_map[item] A__ : Optional[int] = [item, self.key(snake_case )] # Make sure heap is right in both up and down direction. # Ideally only one of them will make any change. self._heapify_up(snake_case ) self._heapify_down(snake_case ) def _UpperCamelCase ( self : Dict , snake_case : int ): '''simple docstring''' if item not in self.pos_map: return A__ : List[str] = self.pos_map[item] del self.pos_map[item] A__ : List[str] = self.arr[self.size - 1] A__ : Union[str, Any] = index self.size -= 1 # Make sure heap is right in both up and down direction. Ideally only one # of them will make any change- so no performance loss in calling both. if self.size > index: self._heapify_up(snake_case ) self._heapify_down(snake_case ) def _UpperCamelCase ( self : Dict , snake_case : int , snake_case : int ): '''simple docstring''' A__ : List[Any] = len(self.arr ) if arr_len == self.size: self.arr.append([item, self.key(snake_case )] ) else: A__ : int = [item, self.key(snake_case )] A__ : Optional[int] = self.size self.size += 1 self._heapify_up(self.size - 1 ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' return self.arr[0] if self.size else None def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.get_top() if top_item_tuple: self.delete_item(top_item_tuple[0] ) return top_item_tuple def _lowerCAmelCase ( ) ->None: pass if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] A__ : Optional[int] = (low + high) // 2 A__ , A__ , A__ : List[Any] = max_subarray(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_subarray(UpperCAmelCase__, mid + 1, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_cross_sum(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int, int, float]: A__ , A__ : Dict = float("""-inf""" ), -1 A__ , A__ : Optional[Any] = float("""-inf""" ), -1 A__ : int | float = 0 for i in range(UpperCAmelCase__, low - 1, -1 ): summ += arr[i] if summ > left_sum: A__ : Optional[int] = summ A__ : Union[str, Any] = i A__ : Optional[Any] = 0 for i in range(mid + 1, high + 1 ): summ += arr[i] if summ > right_sum: A__ : int = summ A__ : Union[str, Any] = i return max_left, max_right, (left_sum + right_sum) def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->float: A__ : Union[str, Any] = [randint(1, UpperCAmelCase__ ) for _ in range(UpperCAmelCase__ )] A__ : Any = time.time() max_subarray(UpperCAmelCase__, 0, input_size - 1 ) A__ : List[Any] = time.time() return end - start def _lowerCAmelCase ( ) ->None: A__ : List[Any] = [1_0, 1_0_0, 1_0_0_0, 1_0_0_0_0, 5_0_0_0_0, 1_0_0_0_0_0, 2_0_0_0_0_0, 3_0_0_0_0_0, 4_0_0_0_0_0, 5_0_0_0_0_0] A__ : Any = [time_max_subarray(UpperCAmelCase__ ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(UpperCAmelCase__, UpperCAmelCase__ ): print(UpperCAmelCase__, """\t\t""", UpperCAmelCase__ ) plt.plot(UpperCAmelCase__, UpperCAmelCase__ ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
296
1
"""simple docstring""" import math from collections.abc import Callable def _lowerCAmelCase ( UpperCAmelCase__ : Callable[[float], float], UpperCAmelCase__ : float, UpperCAmelCase__ : float ) ->float: A__ : float = xa A__ : float = xa while True: if x_n == x_na or function(UpperCAmelCase__ ) == function(UpperCAmelCase__ ): raise ZeroDivisionError("""float division by zero, could not find root""" ) A__ : float = x_na - ( function(UpperCAmelCase__ ) / ((function(UpperCAmelCase__ ) - function(UpperCAmelCase__ )) / (x_na - x_n)) ) if abs(x_na - x_na ) < 1_0**-5: return x_na A__ : Union[str, Any] = x_na A__ : Any = x_na def _lowerCAmelCase ( UpperCAmelCase__ : float ) ->float: return math.pow(UpperCAmelCase__, 3 ) - (2 * x) - 5 if __name__ == "__main__": print(intersection(f, 3, 3.5))
296
"""simple docstring""" from __future__ import annotations class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , snake_case : int ): '''simple docstring''' A__ : List[Any] = order # a_{0} ... a_{k} A__ : List[Any] = [1.0] + [0.0] * order # b_{0} ... b_{k} A__ : str = [1.0] + [0.0] * order # x[n-1] ... x[n-k] A__ : Union[str, Any] = [0.0] * self.order # y[n-1] ... y[n-k] A__ : List[str] = [0.0] * self.order def _UpperCamelCase ( self : Optional[int] , snake_case : list[float] , snake_case : list[float] ): '''simple docstring''' if len(snake_case ) < self.order: A__ : Any = [1.0, *a_coeffs] if len(snake_case ) != self.order + 1: A__ : str = ( F'Expected a_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) if len(snake_case ) != self.order + 1: A__ : Union[str, Any] = ( F'Expected b_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) A__ : Dict = a_coeffs A__ : Any = b_coeffs def _UpperCamelCase ( self : List[str] , snake_case : float ): '''simple docstring''' A__ : str = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) A__ : Dict = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] A__ : Tuple = self.input_history[:-1] A__ : int = self.output_history[:-1] A__ : Dict = sample A__ : Tuple = result return result
296
1
"""simple docstring""" # Author: OMKAR PATHAK, Nwachukwu Chidiebere # Use a Python dictionary to construct the graph. from __future__ import annotations from pprint import pformat from typing import Generic, TypeVar A_ = TypeVar('''T''') class __SCREAMING_SNAKE_CASE ( Generic[T] ): def __init__( self : Any , snake_case : bool = True ): '''simple docstring''' A__ : dict[T, list[T]] = {} # dictionary of lists A__ : Union[str, Any] = directed def _UpperCamelCase ( self : Optional[int] , snake_case : T , snake_case : T ): '''simple docstring''' if not self.directed: # For undirected graphs # if both source vertex and destination vertex are both present in the # adjacency list, add destination vertex to source vertex list of adjacent # vertices and add source vertex to destination vertex list of adjacent # vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case ) self.adj_list[destination_vertex].append(snake_case ) # if only source vertex is present in adjacency list, add destination vertex # to source vertex list of adjacent vertices, then create a new vertex with # destination vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case ) A__ : List[Any] = [source_vertex] # if only destination vertex is present in adjacency list, add source vertex # to destination vertex list of adjacent vertices, then create a new vertex # with source vertex as key and assign a list containing the source vertex # as it's first adjacent vertex. elif destination_vertex in self.adj_list: self.adj_list[destination_vertex].append(snake_case ) A__ : List[str] = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and assign a list # containing the destination vertex as it's first adjacent vertex also # create a new vertex with destination vertex as key and assign a list # containing the source vertex as it's first adjacent vertex. else: A__ : str = [destination_vertex] A__ : Union[str, Any] = [source_vertex] else: # For directed graphs # if both source vertex and destination vertex are present in adjacency # list, add destination vertex to source vertex list of adjacent vertices. if source_vertex in self.adj_list and destination_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case ) # if only source vertex is present in adjacency list, add destination # vertex to source vertex list of adjacent vertices and create a new vertex # with destination vertex as key, which has no adjacent vertex elif source_vertex in self.adj_list: self.adj_list[source_vertex].append(snake_case ) A__ : Dict = [] # if only destination vertex is present in adjacency list, create a new # vertex with source vertex as key and assign a list containing destination # vertex as first adjacent vertex elif destination_vertex in self.adj_list: A__ : Optional[int] = [destination_vertex] # if both source vertex and destination vertex are not present in adjacency # list, create a new vertex with source vertex as key and a list containing # destination vertex as it's first adjacent vertex. Then create a new vertex # with destination vertex as key, which has no adjacent vertex else: A__ : Union[str, Any] = [destination_vertex] A__ : Tuple = [] return self def __repr__( self : Any ): '''simple docstring''' return pformat(self.adj_list )
296
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , snake_case : Optional[Any] , snake_case : Tuple=13 , snake_case : Dict=7 , snake_case : Optional[int]=True , snake_case : Union[str, Any]=True , snake_case : Dict=True , snake_case : Any=True , snake_case : List[str]=99 , snake_case : str=64 , snake_case : Optional[int]=5 , snake_case : str=4 , snake_case : List[Any]=37 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : Dict=16 , snake_case : List[Any]=2 , snake_case : Optional[int]=0.02 , snake_case : Any=3 , snake_case : Union[str, Any]=4 , snake_case : Dict=None , ): '''simple docstring''' A__ : Tuple = parent A__ : Union[str, Any] = batch_size A__ : List[str] = seq_length A__ : Optional[int] = is_training A__ : Dict = use_input_mask A__ : Any = use_token_type_ids A__ : Optional[Any] = use_labels A__ : List[str] = vocab_size A__ : Optional[int] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : Any = num_attention_heads A__ : List[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Optional[int] = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[Any] = initializer_range A__ : Optional[int] = num_labels A__ : Dict = num_choices A__ : Dict = scope A__ : List[Any] = vocab_size - 1 def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[Any] = None if self.use_input_mask: A__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_labels: A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Tuple = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ , A__ , A__ : str = self.prepare_config_and_inputs() A__ : Union[str, Any] = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : List[str] , snake_case : int ): '''simple docstring''' A__ : Any = GPTNeoXModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case ) A__ : Optional[int] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : str , snake_case : Any , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = True A__ : str = GPTNeoXModel(snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Dict , snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : int = self.num_labels A__ : int = GPTNeoXForQuestionAnswering(snake_case ) model.to(snake_case ) model.eval() A__ : Optional[Any] = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : str , snake_case : Tuple , snake_case : int , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : List[Any] = self.num_labels A__ : Tuple = GPTNeoXForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Any , snake_case : Union[str, Any] , snake_case : int , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : Any = GPTNeoXForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Optional[int] = True A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass A__ : Tuple = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ : str = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Any = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Tuple = model(snake_case , attention_mask=snake_case , output_hidden_states=snake_case ) A__ : List[Any] = output_from_no_past["""hidden_states"""][0] A__ : List[str] = model( snake_case , attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )["""hidden_states"""][0] # select random slice A__ : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Dict = config_and_inputs A__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) snake_case_ = (GPTNeoXForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = GPTNeoXModelTester(self ) A__ : Any = ConfigTester(self , config_class=snake_case , hidden_size=64 , num_attention_heads=8 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ , A__ , A__ , A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() A__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ , A__ , A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] = ids_tensor([1, 10] , config.vocab_size ) A__ : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Union[str, Any] = GPTNeoXModel(snake_case ) original_model.to(snake_case ) original_model.eval() A__ : Optional[int] = original_model(snake_case ).last_hidden_state A__ : List[str] = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Optional[int] = {"""type""": scaling_type, """factor""": 10.0} A__ : Optional[int] = GPTNeoXModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() A__ : List[str] = scaled_model(snake_case ).last_hidden_state A__ : Tuple = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: A__ : Optional[Any] = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(snake_case ) A__ : Optional[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" A__ : Tuple = model.generate(**snake_case , do_sample=snake_case , max_new_tokens=20 ) A__ : Tuple = tokenizer.batch_decode(snake_case )[0] self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" from math import factorial, radians def _lowerCAmelCase ( UpperCAmelCase__ : float, UpperCAmelCase__ : int = 1_8, UpperCAmelCase__ : int = 1_0 ) ->float: A__ : Optional[Any] = angle_in_degrees - ((angle_in_degrees // 360.0) * 360.0) # Converting from degrees to radians A__ : str = radians(UpperCAmelCase__ ) A__ : List[Any] = angle_in_radians A__ : int = 3 A__ : Any = -1 for _ in range(UpperCAmelCase__ ): result += (b * (angle_in_radians**a)) / factorial(UpperCAmelCase__ ) A__ : Optional[Any] = -b # One positive term and the next will be negative and so on... a += 2 # Increased by 2 for every term. return round(UpperCAmelCase__, UpperCAmelCase__ ) if __name__ == "__main__": __import__('''doctest''').testmod()
296
"""simple docstring""" from collections import defaultdict from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_5_0_0_0_0_0 ) ->int: A__ : defaultdict = defaultdict(UpperCAmelCase__ ) A__ : Any = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, UpperCAmelCase__, 2 ): if gcd(UpperCAmelCase__, UpperCAmelCase__ ) > 1: continue A__ : str = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'{solution() = }')
296
1
"""simple docstring""" import json import os import shutil import tempfile import unittest from multiprocessing import get_context from pathlib import Path import datasets import numpy as np from datasets import load_dataset from parameterized import parameterized from transformers import AutoProcessor from transformers.models.wavaveca import WavaVecaCTCTokenizer, WavaVecaFeatureExtractor from transformers.models.wavaveca.tokenization_wavaveca import VOCAB_FILES_NAMES from transformers.testing_utils import require_pyctcdecode, require_torch, require_torchaudio, slow from transformers.utils import FEATURE_EXTRACTOR_NAME, is_pyctcdecode_available, is_torch_available from ..wavaveca.test_feature_extraction_wavaveca import floats_list if is_pyctcdecode_available(): from huggingface_hub import snapshot_download from pyctcdecode import BeamSearchDecoderCTC from transformers.models.wavaveca_with_lm import WavaVecaProcessorWithLM from transformers.models.wavaveca_with_lm.processing_wavaveca_with_lm import WavaVecaDecoderWithLMOutput if is_torch_available(): from transformers import WavaVecaForCTC @require_pyctcdecode class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Union[str, Any] = """| <pad> <unk> <s> </s> a b c d e f g h i j k""".split() A__ : Optional[int] = dict(zip(snake_case , range(len(snake_case ) ) ) ) A__ : int = { """unk_token""": """<unk>""", """bos_token""": """<s>""", """eos_token""": """</s>""", } A__ : Union[str, Any] = { """feature_size""": 1, """padding_value""": 0.0, """sampling_rate""": 1_6000, """return_attention_mask""": False, """do_normalize""": True, } A__ : List[str] = tempfile.mkdtemp() A__ : Union[str, Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A__ : Any = os.path.join(self.tmpdirname , snake_case ) with open(self.vocab_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case ) + """\n""" ) with open(self.feature_extraction_file , """w""" , encoding="""utf-8""" ) as fp: fp.write(json.dumps(snake_case ) + """\n""" ) # load decoder from hub A__ : Tuple = """hf-internal-testing/ngram-beam-search-decoder""" def _UpperCamelCase ( self : Dict , **snake_case : str ): '''simple docstring''' A__ : int = self.add_kwargs_tokens_map.copy() kwargs.update(snake_case ) return WavaVecaCTCTokenizer.from_pretrained(self.tmpdirname , **snake_case ) def _UpperCamelCase ( self : str , **snake_case : Optional[int] ): '''simple docstring''' return WavaVecaFeatureExtractor.from_pretrained(self.tmpdirname , **snake_case ) def _UpperCamelCase ( self : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return BeamSearchDecoderCTC.load_from_hf_hub(self.decoder_name , **snake_case ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' shutil.rmtree(self.tmpdirname ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = self.get_tokenizer() A__ : Union[str, Any] = self.get_feature_extractor() A__ : List[Any] = self.get_decoder() A__ : Optional[int] = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) processor.save_pretrained(self.tmpdirname ) A__ : str = WavaVecaProcessorWithLM.from_pretrained(self.tmpdirname ) # tokenizer self.assertEqual(processor.tokenizer.get_vocab() , tokenizer.get_vocab() ) self.assertIsInstance(processor.tokenizer , snake_case ) # feature extractor self.assertEqual(processor.feature_extractor.to_json_string() , feature_extractor.to_json_string() ) self.assertIsInstance(processor.feature_extractor , snake_case ) # decoder self.assertEqual(processor.decoder._alphabet.labels , decoder._alphabet.labels ) self.assertEqual( processor.decoder.model_container[decoder._model_key]._unigram_set , decoder.model_container[decoder._model_key]._unigram_set , ) self.assertIsInstance(processor.decoder , snake_case ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Dict = WavaVecaProcessorWithLM( tokenizer=self.get_tokenizer() , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) processor.save_pretrained(self.tmpdirname ) # make sure that error is thrown when decoder alphabet doesn't match A__ : int = WavaVecaProcessorWithLM.from_pretrained( self.tmpdirname , alpha=5.0 , beta=3.0 , score_boundary=-7.0 , unk_score_offset=3 ) # decoder self.assertEqual(processor.language_model.alpha , 5.0 ) self.assertEqual(processor.language_model.beta , 3.0 ) self.assertEqual(processor.language_model.score_boundary , -7.0 ) self.assertEqual(processor.language_model.unk_score_offset , 3 ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = self.get_tokenizer() # add token to trigger raise tokenizer.add_tokens(["""xx"""] ) with self.assertRaisesRegex(snake_case , """include""" ): WavaVecaProcessorWithLM( tokenizer=snake_case , feature_extractor=self.get_feature_extractor() , decoder=self.get_decoder() ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = self.get_feature_extractor() A__ : List[str] = self.get_tokenizer() A__ : str = self.get_decoder() A__ : Any = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : Tuple = floats_list((3, 1000) ) A__ : List[Any] = feature_extractor(snake_case , return_tensors="""np""" ) A__ : Optional[Any] = processor(snake_case , return_tensors="""np""" ) for key in input_feat_extract.keys(): self.assertAlmostEqual(input_feat_extract[key].sum() , input_processor[key].sum() , delta=1e-2 ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = self.get_feature_extractor() A__ : Dict = self.get_tokenizer() A__ : Optional[Any] = self.get_decoder() A__ : Tuple = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : Optional[Any] = """This is a test string""" A__ : Dict = processor(text=snake_case ) A__ : int = tokenizer(snake_case ) for key in encoded_tok.keys(): self.assertListEqual(encoded_tok[key] , encoded_processor[key] ) def _UpperCamelCase ( self : Tuple , snake_case : int=(2, 10, 16) , snake_case : Union[str, Any]=77 ): '''simple docstring''' np.random.seed(snake_case ) return np.random.rand(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Any = self.get_feature_extractor() A__ : List[Any] = self.get_tokenizer() A__ : List[str] = self.get_decoder() A__ : str = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : Optional[int] = self._get_dummy_logits(shape=(10, 16) , seed=13 ) A__ : List[str] = processor.decode(snake_case ) A__ : List[str] = decoder.decode_beams(snake_case )[0] self.assertEqual(decoded_decoder[0] , decoded_processor.text ) self.assertEqual("""</s> <s> </s>""" , decoded_processor.text ) self.assertEqual(decoded_decoder[-2] , decoded_processor.logit_score ) self.assertEqual(decoded_decoder[-1] , decoded_processor.lm_score ) @parameterized.expand([[None], ["""fork"""], ["""spawn"""]] ) def _UpperCamelCase ( self : List[Any] , snake_case : List[str] ): '''simple docstring''' A__ : List[Any] = self.get_feature_extractor() A__ : List[Any] = self.get_tokenizer() A__ : Optional[int] = self.get_decoder() A__ : List[Any] = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : Tuple = self._get_dummy_logits() # note: pool should be instantiated *after* Wav2Vec2ProcessorWithLM. # otherwise, the LM won't be available to the pool's sub-processes. # manual logic used to allow parameterized test for both pool=None and pool=Pool(...) if pool_context is None: A__ : int = processor.batch_decode(snake_case ) else: with get_context(snake_case ).Pool() as pool: A__ : Optional[Any] = processor.batch_decode(snake_case , snake_case ) A__ : Tuple = list(snake_case ) with get_context("""fork""" ).Pool() as p: A__ : Union[str, Any] = decoder.decode_beams_batch(snake_case , snake_case ) A__ , A__ , A__ : int = [], [], [] for beams in decoded_beams: texts_decoder.append(beams[0][0] ) logit_scores_decoder.append(beams[0][-2] ) lm_scores_decoder.append(beams[0][-1] ) self.assertListEqual(snake_case , decoded_processor.text ) self.assertListEqual(["""<s> <s> </s>""", """<s> <s> <s>"""] , decoded_processor.text ) self.assertListEqual(snake_case , decoded_processor.logit_score ) self.assertListEqual(snake_case , decoded_processor.lm_score ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : List[Any] = self.get_feature_extractor() A__ : str = self.get_tokenizer() A__ : Dict = self.get_decoder() A__ : Union[str, Any] = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : List[str] = self._get_dummy_logits() A__ : List[Any] = 15 A__ : Any = -20.0 A__ : Dict = -4.0 A__ : Dict = processor.batch_decode( snake_case , beam_width=snake_case , beam_prune_logp=snake_case , token_min_logp=snake_case , ) A__ : Optional[int] = decoded_processor_out.text A__ : List[str] = list(snake_case ) with get_context("""fork""" ).Pool() as pool: A__ : Any = decoder.decode_beams_batch( snake_case , snake_case , beam_width=snake_case , beam_prune_logp=snake_case , token_min_logp=snake_case , ) A__ : Optional[int] = [d[0][0] for d in decoded_decoder_out] A__ : int = [d[0][2] for d in decoded_decoder_out] A__ : Tuple = [d[0][3] for d in decoded_decoder_out] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(["""</s> <s> <s>""", """<s> <s> <s>"""] , snake_case ) self.assertTrue(np.array_equal(snake_case , decoded_processor_out.logit_score ) ) self.assertTrue(np.allclose([-20.054, -18.447] , snake_case , atol=1e-3 ) ) self.assertTrue(np.array_equal(snake_case , decoded_processor_out.lm_score ) ) self.assertTrue(np.allclose([-15.554, -13.9474] , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Dict = self.get_feature_extractor() A__ : List[str] = self.get_tokenizer() A__ : List[Any] = self.get_decoder() A__ : int = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) A__ : List[str] = self._get_dummy_logits() A__ : Union[str, Any] = 2.0 A__ : Any = 5.0 A__ : int = -20.0 A__ : int = True A__ : List[Any] = processor.batch_decode( snake_case , alpha=snake_case , beta=snake_case , unk_score_offset=snake_case , lm_score_boundary=snake_case , ) A__ : Optional[Any] = decoded_processor_out.text A__ : Union[str, Any] = list(snake_case ) decoder.reset_params( alpha=snake_case , beta=snake_case , unk_score_offset=snake_case , lm_score_boundary=snake_case , ) with get_context("""fork""" ).Pool() as pool: A__ : Optional[int] = decoder.decode_beams_batch( snake_case , snake_case , ) A__ : Tuple = [d[0][0] for d in decoded_decoder_out] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(["""<s> </s> <s> </s> </s>""", """</s> </s> <s> </s> </s>"""] , snake_case ) A__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] self.assertEqual(lm_model.alpha , 2.0 ) self.assertEqual(lm_model.beta , 5.0 ) self.assertEqual(lm_model.unk_score_offset , -20.0 ) self.assertEqual(lm_model.score_boundary , snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) A__ : Union[str, Any] = processor.decoder.model_container[processor.decoder._model_key] A__ : Tuple = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() A__ : Tuple = os.listdir(snake_case ) A__ : int = ["""alphabet.json""", """language_model"""] downloaded_decoder_files.sort() expected_decoder_files.sort() # test that only decoder relevant files from # https://huggingface.co/hf-internal-testing/processor_with_lm/tree/main # are downloaded and none of the rest (e.g. README.md, ...) self.assertListEqual(snake_case , snake_case ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = snapshot_download("""hf-internal-testing/processor_with_lm""" ) A__ : Tuple = WavaVecaProcessorWithLM.from_pretrained(snake_case ) A__ : List[Any] = processor.decoder.model_container[processor.decoder._model_key] A__ : List[str] = Path(language_model._kenlm_model.path.decode("""utf-8""" ) ).parent.parent.absolute() A__ : Optional[int] = os.listdir(snake_case ) A__ : Any = os.listdir(snake_case ) local_decoder_files.sort() expected_decoder_files.sort() # test that both decoder form hub and local files in cache are the same self.assertListEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : str = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) A__ : Optional[Any] = AutoProcessor.from_pretrained("""hf-internal-testing/processor_with_lm""" ) A__ : Union[str, Any] = floats_list((3, 1000) ) A__ : Tuple = processor_wavaveca(snake_case , return_tensors="""np""" ) A__ : Any = processor_auto(snake_case , return_tensors="""np""" ) for key in input_wavaveca.keys(): self.assertAlmostEqual(input_wavaveca[key].sum() , input_auto[key].sum() , delta=1e-2 ) A__ : List[str] = self._get_dummy_logits() A__ : Dict = processor_wavaveca.batch_decode(snake_case ) A__ : str = processor_auto.batch_decode(snake_case ) self.assertListEqual(decoded_wavaveca.text , decoded_auto.text ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.get_feature_extractor() A__ : Union[str, Any] = self.get_tokenizer() A__ : Union[str, Any] = self.get_decoder() A__ : str = WavaVecaProcessorWithLM(tokenizer=snake_case , feature_extractor=snake_case , decoder=snake_case ) self.assertListEqual( processor.model_input_names , feature_extractor.model_input_names , msg="""`processor` and `feature_extractor` model input names do not match""" , ) @staticmethod def _UpperCamelCase ( snake_case : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' A__ : List[str] = [d[key] for d in offsets] return retrieved_list def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) A__ : str = self._get_dummy_logits()[0] A__ : Optional[int] = processor.decode(snake_case , output_word_offsets=snake_case ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(snake_case , snake_case ) ) self.assertEqual(""" """.join(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) ) , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""] , """end_offset""" ) , [1, 3, 5] ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Optional[Any] = WavaVecaProcessorWithLM.from_pretrained("""hf-internal-testing/processor_with_lm""" ) A__ : Optional[int] = self._get_dummy_logits() A__ : List[str] = processor.batch_decode(snake_case , output_word_offsets=snake_case ) # check Wav2Vec2CTCTokenizerOutput keys for word self.assertEqual(len(outputs.keys() ) , 4 ) self.assertTrue("""text""" in outputs ) self.assertTrue("""word_offsets""" in outputs ) self.assertTrue(isinstance(snake_case , snake_case ) ) self.assertListEqual( [""" """.join(self.get_from_offsets(snake_case , """word""" ) ) for o in outputs["""word_offsets"""]] , outputs.text ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """word""" ) , ["""<s>""", """<s>""", """</s>"""] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """start_offset""" ) , [0, 2, 4] ) self.assertListEqual(self.get_from_offsets(outputs["""word_offsets"""][0] , """end_offset""" ) , [1, 3, 5] ) @slow @require_torch @require_torchaudio def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' import torch A__ : Optional[int] = load_dataset("""common_voice""" , """en""" , split="""train""" , streaming=snake_case ) A__ : Dict = ds.cast_column("""audio""" , datasets.Audio(sampling_rate=1_6000 ) ) A__ : int = iter(snake_case ) A__ : List[str] = next(snake_case ) A__ : Tuple = AutoProcessor.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) A__ : Union[str, Any] = WavaVecaForCTC.from_pretrained("""patrickvonplaten/wav2vec2-base-100h-with-lm""" ) # compare to filename `common_voice_en_100038.mp3` of dataset viewer on https://huggingface.co/datasets/common_voice/viewer/en/train A__ : Optional[int] = processor(sample["""audio"""]["""array"""] , return_tensors="""pt""" ).input_values with torch.no_grad(): A__ : List[str] = model(snake_case ).logits.cpu().numpy() A__ : Tuple = processor.decode(logits[0] , output_word_offsets=snake_case ) A__ : Union[str, Any] = model.config.inputs_to_logits_ratio / processor.feature_extractor.sampling_rate A__ : Union[str, Any] = [ { """start_time""": d["""start_offset"""] * time_offset, """end_time""": d["""end_offset"""] * time_offset, """word""": d["""word"""], } for d in output["""word_offsets"""] ] A__ : Optional[int] = """WHY DOES MILISANDRA LOOK LIKE SHE WANTS TO CONSUME JOHN SNOW ON THE RIVER AT THE WALL""" # output words self.assertEqual(""" """.join(self.get_from_offsets(snake_case , """word""" ) ) , snake_case ) self.assertEqual(""" """.join(self.get_from_offsets(snake_case , """word""" ) ) , output.text ) # output times A__ : Optional[int] = torch.tensor(self.get_from_offsets(snake_case , """start_time""" ) ) A__ : Dict = torch.tensor(self.get_from_offsets(snake_case , """end_time""" ) ) # fmt: off A__ : Optional[int] = torch.tensor([1.4199, 1.6599, 2.2599, 3.0, 3.24, 3.5999, 3.7999, 4.0999, 4.26, 4.94, 5.28, 5.6599, 5.78, 5.94, 6.32, 6.5399, 6.6599] ) A__ : Optional[int] = torch.tensor([1.5399, 1.8999, 2.9, 3.16, 3.5399, 3.72, 4.0199, 4.1799, 4.76, 5.1599, 5.5599, 5.6999, 5.86, 6.1999, 6.38, 6.6199, 6.94] ) # fmt: on self.assertTrue(torch.allclose(snake_case , snake_case , atol=0.01 ) ) self.assertTrue(torch.allclose(snake_case , snake_case , atol=0.01 ) )
296
"""simple docstring""" import os from distutils.util import strtobool def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any] ) ->List[str]: for e in env_keys: A__ : List[Any] = int(os.environ.get(UpperCAmelCase__, -1 ) ) if val >= 0: return val return default def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : str=False ) ->List[str]: A__ : List[Any] = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return strtobool(UpperCAmelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]="no" ) ->int: A__ : str = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return value
296
1
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging A_ = logging.get_logger(__name__) A_ = '''▁''' A_ = {'''vocab_file''': '''sentencepiece.bpe.model'''} A_ = { '''vocab_file''': { '''facebook/mbart-large-en-ro''': ( '''https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model''' ), '''facebook/mbart-large-cc25''': ( '''https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model''' ), } } A_ = { '''facebook/mbart-large-en-ro''': 1024, '''facebook/mbart-large-cc25''': 1024, } # fmt: off A_ = ['''ar_AR''', '''cs_CZ''', '''de_DE''', '''en_XX''', '''es_XX''', '''et_EE''', '''fi_FI''', '''fr_XX''', '''gu_IN''', '''hi_IN''', '''it_IT''', '''ja_XX''', '''kk_KZ''', '''ko_KR''', '''lt_LT''', '''lv_LV''', '''my_MM''', '''ne_NP''', '''nl_XX''', '''ro_RO''', '''ru_RU''', '''si_LK''', '''tr_TR''', '''vi_VN''', '''zh_CN'''] class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = ['input_ids', 'attention_mask'] snake_case_ = [] snake_case_ = [] def __init__( self : str , snake_case : List[str] , snake_case : Optional[int]="<s>" , snake_case : List[str]="</s>" , snake_case : Any="</s>" , snake_case : int="<s>" , snake_case : Optional[Any]="<unk>" , snake_case : List[str]="<pad>" , snake_case : List[Any]="<mask>" , snake_case : int=None , snake_case : Dict=None , snake_case : int=None , snake_case : Optional[Dict[str, Any]] = None , snake_case : Optional[int]=None , **snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : List[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , cls_token=snake_case , pad_token=snake_case , mask_token=snake_case , tokenizer_file=snake_case , src_lang=snake_case , tgt_lang=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(snake_case ) ) A__ : Any = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token A__ : Union[str, Any] = {"""<s>""": 0, """<pad>""": 1, """</s>""": 2, """<unk>""": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab A__ : Union[str, Any] = 1 A__ : Tuple = len(self.sp_model ) A__ : List[str] = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(snake_case ) } A__ : str = {v: k for k, v in self.lang_code_to_id.items()} A__ : str = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) A__ : Tuple = {v: k for k, v in self.fairseq_tokens_to_ids.items()} A__ : Optional[int] = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) A__ : Union[str, Any] = src_lang if src_lang is not None else """en_XX""" A__ : Dict = self.lang_code_to_id[self._src_lang] A__ : Optional[Any] = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__( self : Optional[Any] ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : Dict = None A__ : int = self.sp_model.serialized_model_proto() return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : Optional[int] = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Union[str, Any] = {} A__ : List[Any] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return self._src_lang @src_lang.setter def _UpperCamelCase ( self : Dict , snake_case : str ): '''simple docstring''' A__ : List[str] = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def _UpperCamelCase ( self : Optional[Any] , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) A__ : Any = [1] * len(self.prefix_tokens ) A__ : int = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(snake_case )) + suffix_ones return prefix_ones + ([0] * len(snake_case )) + ([0] * len(snake_case )) + suffix_ones def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def _UpperCamelCase ( self : Optional[int] , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Union[str, Any] = [self.sep_token_id] A__ : Optional[int] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def _UpperCamelCase ( self : List[Any] , snake_case : Dict , snake_case : str , snake_case : Optional[str] , snake_case : Optional[str] , **snake_case : Optional[Any] ): '''simple docstring''' if src_lang is None or tgt_lang is None: raise ValueError("""Translation requires a `src_lang` and a `tgt_lang` for this model""" ) A__ : Dict = src_lang A__ : Union[str, Any] = self(snake_case , add_special_tokens=snake_case , return_tensors=snake_case , **snake_case ) A__ : int = self.convert_tokens_to_ids(snake_case ) A__ : Optional[int] = tgt_lang_id return inputs def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def _UpperCamelCase ( self : str , snake_case : str ): '''simple docstring''' return self.sp_model.encode(snake_case , out_type=snake_case ) def _UpperCamelCase ( self : Tuple , snake_case : str ): '''simple docstring''' if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] A__ : List[str] = self.sp_model.PieceToId(snake_case ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def _UpperCamelCase ( self : Tuple , snake_case : str ): '''simple docstring''' A__ : str = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : List[str] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : Dict = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Dict = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,) def _UpperCamelCase ( self : Optional[int] , snake_case : List[str] , snake_case : str = "en_XX" , snake_case : Optional[List[str]] = None , snake_case : str = "ro_RO" , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : Optional[int] = src_lang A__ : List[str] = tgt_lang return super().prepare_seqaseq_batch(snake_case , snake_case , **snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return self.set_src_lang_special_tokens(self.src_lang ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return self.set_tgt_lang_special_tokens(self.tgt_lang ) def _UpperCamelCase ( self : int , snake_case : Tuple ): '''simple docstring''' A__ : List[str] = self.lang_code_to_id[src_lang] A__ : Any = [] A__ : List[str] = [self.eos_token_id, self.cur_lang_code] def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : int = self.lang_code_to_id[lang] A__ : Any = [] A__ : Optional[Any] = [self.eos_token_id, self.cur_lang_code]
296
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
1
"""simple docstring""" import argparse import gc import json import os import re import torch from huggingface_hub import hf_hub_download from transformers import AutoModelForCausalLM, AutoTokenizer, PreTrainedTokenizerFast, RwkvConfig from transformers.modeling_utils import WEIGHTS_INDEX_NAME, shard_checkpoint A_ = { '''169M''': 12, '''430M''': 24, '''1B5''': 24, '''3B''': 32, '''7B''': 32, '''14B''': 40, } A_ = { '''169M''': 768, '''430M''': 1024, '''1B5''': 2048, '''3B''': 2560, '''7B''': 4096, '''14B''': 5120, } def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->List[Any]: A__ : Tuple = list(state_dict.keys() ) for name in state_dict_keys: A__ : Optional[Any] = state_dict.pop(UpperCAmelCase__ ) # emb -> embedding if name.startswith("""emb.""" ): A__ : Optional[Any] = name.replace("""emb.""", """embeddings.""" ) # ln_0 -> pre_ln (only present at block 0) if name.startswith("""blocks.0.ln0""" ): A__ : List[Any] = name.replace("""blocks.0.ln0""", """blocks.0.pre_ln""" ) # att -> attention A__ : Union[str, Any] = re.sub(R"""blocks\.(\d+)\.att""", R"""blocks.\1.attention""", UpperCAmelCase__ ) # ffn -> feed_forward A__ : int = re.sub(R"""blocks\.(\d+)\.ffn""", R"""blocks.\1.feed_forward""", UpperCAmelCase__ ) # time_mix_k -> time_mix_key and reshape if name.endswith(""".time_mix_k""" ): A__ : List[str] = name.replace(""".time_mix_k""", """.time_mix_key""" ) # time_mix_v -> time_mix_value and reshape if name.endswith(""".time_mix_v""" ): A__ : int = name.replace(""".time_mix_v""", """.time_mix_value""" ) # time_mix_r -> time_mix_key and reshape if name.endswith(""".time_mix_r""" ): A__ : int = name.replace(""".time_mix_r""", """.time_mix_receptance""" ) if name != "head.weight": A__ : Union[str, Any] = """rwkv.""" + name A__ : Optional[int] = weight return state_dict def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : int, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Any=None, UpperCAmelCase__ : int=None, UpperCAmelCase__ : Dict=False, UpperCAmelCase__ : Union[str, Any]=None ) ->Any: # 1. If possible, build the tokenizer. if tokenizer_file is None: print("""No `--tokenizer_file` provided, we will use the default tokenizer.""" ) A__ : str = 5_0_2_7_7 A__ : Optional[Any] = AutoTokenizer.from_pretrained("""EleutherAI/gpt-neox-20b""" ) else: A__ : Optional[int] = PreTrainedTokenizerFast(tokenizer_file=UpperCAmelCase__ ) A__ : List[Any] = len(UpperCAmelCase__ ) tokenizer.save_pretrained(UpperCAmelCase__ ) # 2. Build the config A__ : Optional[int] = list(NUM_HIDDEN_LAYERS_MAPPING.keys() ) if size is None: # Try to infer size from the checkpoint name for candidate in possible_sizes: if candidate in checkpoint_file: A__ : int = candidate break if size is None: raise ValueError("""Could not infer the size, please provide it with the `--size` argument.""" ) if size not in possible_sizes: raise ValueError(f'`size` should be one of {possible_sizes}, got {size}.' ) A__ : int = RwkvConfig( vocab_size=UpperCAmelCase__, num_hidden_layers=NUM_HIDDEN_LAYERS_MAPPING[size], hidden_size=HIDEN_SIZE_MAPPING[size], ) config.save_pretrained(UpperCAmelCase__ ) # 3. Download model file then convert state_dict A__ : Optional[Any] = hf_hub_download(UpperCAmelCase__, UpperCAmelCase__ ) A__ : Dict = torch.load(UpperCAmelCase__, map_location="""cpu""" ) A__ : List[str] = convert_state_dict(UpperCAmelCase__ ) # 4. Split in shards and save A__ , A__ : str = shard_checkpoint(UpperCAmelCase__ ) for shard_file, shard in shards.items(): torch.save(UpperCAmelCase__, os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) ) if index is not None: A__ : List[Any] = os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) # Save the index as well with open(UpperCAmelCase__, """w""", encoding="""utf-8""" ) as f: A__ : Dict = json.dumps(UpperCAmelCase__, indent=2, sort_keys=UpperCAmelCase__ ) + """\n""" f.write(UpperCAmelCase__ ) # 5. Clean up shards (for some reason the file PyTorch saves take the same space as the whole state_dict print( """Cleaning up shards. This may error with an OOM error, it this is the case don't worry you still have converted the model.""" ) A__ : Optional[Any] = list(shards.keys() ) del state_dict del shards gc.collect() for shard_file in shard_files: A__ : Union[str, Any] = torch.load(os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) ) torch.save({k: v.cpu().clone() for k, v in state_dict.items()}, os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) ) del state_dict gc.collect() if push_to_hub: if model_name is None: raise ValueError("""Please provide a `model_name` to push the model to the Hub.""" ) A__ : Union[str, Any] = AutoModelForCausalLM.from_pretrained(UpperCAmelCase__ ) model.push_to_hub(UpperCAmelCase__, max_shard_size="""2GB""" ) tokenizer.push_to_hub(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--repo_id''', default=None, type=str, required=True, help='''Repo ID from which to pull the checkpoint.''' ) parser.add_argument( '''--checkpoint_file''', default=None, type=str, required=True, help='''Name of the checkpoint file in the repo.''' ) parser.add_argument( '''--output_dir''', default=None, type=str, required=True, help='''Where to save the converted model.''' ) parser.add_argument( '''--tokenizer_file''', default=None, type=str, help='''Path to the tokenizer file to use (if not provided, only the model is converted).''', ) parser.add_argument( '''--size''', default=None, type=str, help='''Size of the model. Will be inferred from the `checkpoint_file` if not passed.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Push to the Hub the converted model.''', ) parser.add_argument( '''--model_name''', default=None, type=str, help='''Name of the pushed model on the Hub, including the username / organization.''', ) A_ = parser.parse_args() convert_rmkv_checkpoint_to_hf_format( args.repo_id, args.checkpoint_file, args.output_dir, size=args.size, tokenizer_file=args.tokenizer_file, push_to_hub=args.push_to_hub, model_name=args.model_name, )
296
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = logging.get_logger(__name__) A_ = Dict[str, Any] A_ = List[Prediction] @add_end_docstrings(UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : str , *snake_case : Tuple , **snake_case : Tuple ): '''simple docstring''' super().__init__(*snake_case , **snake_case ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _UpperCamelCase ( self : List[Any] , **snake_case : Optional[int] ): '''simple docstring''' A__ : Dict = {} if "threshold" in kwargs: A__ : int = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return super().__call__(*snake_case , **snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : List[str] = load_image(snake_case ) A__ : int = torch.IntTensor([[image.height, image.width]] ) A__ : Union[str, Any] = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: A__ : str = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) A__ : List[str] = target_size return inputs def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : str = model_inputs.pop("""target_size""" ) A__ : Dict = self.model(**snake_case ) A__ : Optional[Any] = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: A__ : str = model_inputs["""bbox"""] return model_outputs def _UpperCamelCase ( self : Tuple , snake_case : Optional[int] , snake_case : int=0.9 ): '''simple docstring''' A__ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A__ , A__ : Tuple = target_size[0].tolist() def unnormalize(snake_case : Optional[int] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) A__ , A__ : Optional[int] = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A__ : Optional[Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A__ : List[str] = [unnormalize(snake_case ) for bbox in model_outputs["""bbox"""].squeeze(0 )] A__ : Tuple = ["""score""", """label""", """box"""] A__ : Any = [dict(zip(snake_case , snake_case ) ) for vals in zip(scores.tolist() , snake_case , snake_case ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A__ : Union[str, Any] = self.image_processor.post_process_object_detection(snake_case , snake_case , snake_case ) A__ : str = raw_annotations[0] A__ : str = raw_annotation["""scores"""] A__ : List[Any] = raw_annotation["""labels"""] A__ : int = raw_annotation["""boxes"""] A__ : str = scores.tolist() A__ : Any = [self.model.config.idalabel[label.item()] for label in labels] A__ : int = [self._get_bounding_box(snake_case ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A__ : str = ["""score""", """label""", """box"""] A__ : Dict = [ dict(zip(snake_case , snake_case ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def _UpperCamelCase ( self : Union[str, Any] , snake_case : "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) A__ , A__ , A__ , A__ : Any = box.int().tolist() A__ : Any = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
296
1
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
"""simple docstring""" import argparse import json import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import AutoImageProcessor, SwinConfig, SwinForImageClassification def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->str: A__ : str = SwinConfig() A__ : Dict = swin_name.split("""_""" ) A__ : Union[str, Any] = name_split[1] A__ : List[Any] = int(name_split[4] ) A__ : Tuple = int(name_split[3][-1] ) if model_size == "tiny": A__ : Optional[Any] = 9_6 A__ : Any = (2, 2, 6, 2) A__ : Union[str, Any] = (3, 6, 1_2, 2_4) elif model_size == "small": A__ : Optional[Any] = 9_6 A__ : Union[str, Any] = (2, 2, 1_8, 2) A__ : List[str] = (3, 6, 1_2, 2_4) elif model_size == "base": A__ : Dict = 1_2_8 A__ : str = (2, 2, 1_8, 2) A__ : List[str] = (4, 8, 1_6, 3_2) else: A__ : Any = 1_9_2 A__ : Tuple = (2, 2, 1_8, 2) A__ : Dict = (6, 1_2, 2_4, 4_8) if "in22k" in swin_name: A__ : List[str] = 2_1_8_4_1 else: A__ : List[str] = 1_0_0_0 A__ : Dict = """huggingface/label-files""" A__ : Optional[int] = """imagenet-1k-id2label.json""" A__ : str = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : List[str] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : List[Any] = idalabel A__ : int = {v: k for k, v in idalabel.items()} A__ : Union[str, Any] = img_size A__ : Any = num_classes A__ : int = embed_dim A__ : Tuple = depths A__ : Any = num_heads A__ : int = window_size return config def _lowerCAmelCase ( UpperCAmelCase__ : Dict ) ->Any: if "patch_embed.proj" in name: A__ : str = name.replace("""patch_embed.proj""", """embeddings.patch_embeddings.projection""" ) if "patch_embed.norm" in name: A__ : Optional[Any] = name.replace("""patch_embed.norm""", """embeddings.norm""" ) if "layers" in name: A__ : List[str] = """encoder.""" + name if "attn.proj" in name: A__ : List[Any] = name.replace("""attn.proj""", """attention.output.dense""" ) if "attn" in name: A__ : str = name.replace("""attn""", """attention.self""" ) if "norm1" in name: A__ : Tuple = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : Union[str, Any] = name.replace("""norm2""", """layernorm_after""" ) if "mlp.fc1" in name: A__ : Tuple = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if name == "norm.weight": A__ : Union[str, Any] = """layernorm.weight""" if name == "norm.bias": A__ : int = """layernorm.bias""" if "head" in name: A__ : List[Any] = name.replace("""head""", """classifier""" ) else: A__ : str = """swin.""" + name return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[Any] ) ->int: for key in orig_state_dict.copy().keys(): A__ : Dict = orig_state_dict.pop(UpperCAmelCase__ ) if "mask" in key: continue elif "qkv" in key: A__ : List[Any] = key.split(""".""" ) A__ : int = int(key_split[1] ) A__ : Union[str, Any] = int(key_split[3] ) A__ : Optional[int] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: A__ : Union[str, Any] = val[:dim, :] A__ : Optional[int] = val[ dim : dim * 2, : ] A__ : int = val[-dim:, :] else: A__ : List[Any] = val[ :dim ] A__ : Optional[int] = val[ dim : dim * 2 ] A__ : Dict = val[ -dim: ] else: A__ : str = val return orig_state_dict def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str ) ->str: A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() A__ : List[str] = get_swin_config(UpperCAmelCase__ ) A__ : Dict = SwinForImageClassification(UpperCAmelCase__ ) model.eval() A__ : Any = convert_state_dict(timm_model.state_dict(), UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) A__ : Any = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : List[Any] = AutoImageProcessor.from_pretrained("""microsoft/{}""".format(swin_name.replace("""_""", """-""" ) ) ) A__ : Tuple = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) A__ : str = image_processor(images=UpperCAmelCase__, return_tensors="""pt""" ) A__ : Optional[int] = timm_model(inputs["""pixel_values"""] ) A__ : Tuple = model(**UpperCAmelCase__ ).logits assert torch.allclose(UpperCAmelCase__, UpperCAmelCase__, atol=1e-3 ) print(f'Saving model {swin_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--swin_name''', default='''swin_tiny_patch4_window7_224''', type=str, help='''Name of the Swin timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_swin_checkpoint(args.swin_name, args.pytorch_dump_folder_path)
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''unc-nlp/lxmert-base-uncased''': '''https://huggingface.co/unc-nlp/lxmert-base-uncased/resolve/main/config.json''', } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'lxmert' snake_case_ = {} def __init__( self : int , snake_case : Dict=3_0522 , snake_case : Optional[int]=768 , snake_case : str=12 , snake_case : List[Any]=9500 , snake_case : Optional[Any]=1600 , snake_case : str=400 , snake_case : Dict=3072 , snake_case : Tuple="gelu" , snake_case : str=0.1 , snake_case : int=0.1 , snake_case : Any=512 , snake_case : List[str]=2 , snake_case : List[str]=0.02 , snake_case : Optional[int]=1e-12 , snake_case : List[str]=9 , snake_case : List[Any]=5 , snake_case : str=5 , snake_case : int=2048 , snake_case : int=4 , snake_case : Optional[Any]=6.67 , snake_case : List[str]=True , snake_case : List[Any]=True , snake_case : Tuple=True , snake_case : Any=True , snake_case : Tuple=True , snake_case : Tuple=True , snake_case : Tuple=True , **snake_case : List[Any] , ): '''simple docstring''' A__ : List[Any] = vocab_size A__ : List[Any] = hidden_size A__ : Union[str, Any] = num_attention_heads A__ : int = hidden_act A__ : Optional[int] = intermediate_size A__ : Tuple = hidden_dropout_prob A__ : Union[str, Any] = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : Dict = type_vocab_size A__ : Union[str, Any] = initializer_range A__ : Tuple = layer_norm_eps A__ : str = num_qa_labels A__ : List[str] = num_object_labels A__ : List[str] = num_attr_labels A__ : Any = l_layers A__ : Dict = x_layers A__ : Optional[Any] = r_layers A__ : Optional[int] = visual_feat_dim A__ : Union[str, Any] = visual_pos_dim A__ : Dict = visual_loss_normalizer A__ : str = task_matched A__ : List[Any] = task_mask_lm A__ : int = task_obj_predict A__ : Optional[Any] = task_qa A__ : Union[str, Any] = visual_obj_loss A__ : int = visual_attr_loss A__ : str = visual_feat_loss A__ : Union[str, Any] = {"""vision""": r_layers, """cross_encoder""": x_layers, """language""": l_layers} super().__init__(**snake_case )
296
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
1
"""simple docstring""" import unittest import numpy as np from diffusers import LMSDiscreteScheduler, OnnxStableDiffusionInpaintPipeline from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): # FIXME: add fast tests pass @nightly @require_onnxruntime @require_torch_gpu class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @property def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Any = ort.SessionOptions() A__ : Tuple = False return options def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) A__ : Optional[int] = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) A__ : Tuple = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , safety_checker=snake_case , feature_extractor=snake_case , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case ) A__ : Dict = """A red cat sitting on a park bench""" A__ : Optional[Any] = np.random.RandomState(0 ) A__ : Union[str, Any] = pipe( prompt=snake_case , image=snake_case , mask_image=snake_case , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case , output_type="""np""" , ) A__ : List[Any] = output.images A__ : int = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) A__ : Union[str, Any] = np.array([0.2514, 0.3007, 0.3517, 0.1790, 0.2382, 0.3167, 0.1944, 0.2273, 0.2464] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3 def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo.png""" ) A__ : int = load_image( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main""" """/in_paint/overture-creations-5sI6fQgYIuo_mask.png""" ) A__ : List[str] = LMSDiscreteScheduler.from_pretrained( """runwayml/stable-diffusion-inpainting""" , subfolder="""scheduler""" , revision="""onnx""" ) A__ : List[str] = OnnxStableDiffusionInpaintPipeline.from_pretrained( """runwayml/stable-diffusion-inpainting""" , revision="""onnx""" , scheduler=snake_case , safety_checker=snake_case , feature_extractor=snake_case , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case ) A__ : List[str] = """A red cat sitting on a park bench""" A__ : Union[str, Any] = np.random.RandomState(0 ) A__ : int = pipe( prompt=snake_case , image=snake_case , mask_image=snake_case , guidance_scale=7.5 , num_inference_steps=20 , generator=snake_case , output_type="""np""" , ) A__ : Dict = output.images A__ : Optional[int] = images[0, 255:258, 255:258, -1] assert images.shape == (1, 512, 512, 3) A__ : List[str] = np.array([0.0086, 0.0077, 0.0083, 0.0093, 0.0107, 0.0139, 0.0094, 0.0097, 0.0125] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-3
296
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[Any]=13 , snake_case : Union[str, Any]=7 , snake_case : Optional[Any]=True , snake_case : str=True , snake_case : Dict=False , snake_case : Union[str, Any]=True , snake_case : Optional[Any]=99 , snake_case : str=32 , snake_case : Tuple=5 , snake_case : List[str]=4 , snake_case : Optional[int]=37 , snake_case : str="gelu" , snake_case : Tuple=0.1 , snake_case : Optional[int]=0.1 , snake_case : int=512 , snake_case : List[str]=16 , snake_case : str=2 , snake_case : Optional[int]=0.02 , snake_case : str=3 , snake_case : Dict=4 , snake_case : Optional[Any]=None , ): '''simple docstring''' A__ : int = parent A__ : Union[str, Any] = batch_size A__ : Optional[int] = seq_length A__ : List[Any] = is_training A__ : List[str] = use_input_mask A__ : Optional[Any] = use_token_type_ids A__ : List[Any] = use_labels A__ : Union[str, Any] = vocab_size A__ : List[Any] = hidden_size A__ : Any = num_hidden_layers A__ : Any = num_attention_heads A__ : Optional[int] = intermediate_size A__ : Any = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Dict = attention_probs_dropout_prob A__ : Optional[int] = max_position_embeddings A__ : Tuple = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[str] = initializer_range A__ : Any = num_labels A__ : Any = num_choices A__ : int = scope def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = None if self.use_input_mask: A__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_token_type_ids: A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : int = None A__ : int = None A__ : List[str] = None if self.use_labels: A__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) A__ : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Any , snake_case : Dict , snake_case : Any , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case ) A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Optional[int] , snake_case : List[str] , snake_case : str , snake_case : Optional[Any] , snake_case : List[str] , snake_case : List[Any] , snake_case : Tuple , snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = BioGptForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Any , snake_case : str , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Any , *snake_case : Dict ): '''simple docstring''' A__ : Union[str, Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() # create attention mask A__ : List[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) A__ : Any = self.seq_length // 2 A__ : str = 0 # first forward pass A__ , A__ : List[Any] = model(snake_case , attention_mask=snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids A__ : int = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A__ : List[str] = ids_tensor((1,) , snake_case ).item() + 1 A__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A__ : int = random_other_next_tokens # append to next input_ids and attn_mask A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case )] , dim=1 , ) # get two different outputs A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Optional[int] = model(snake_case , past_key_values=snake_case , attention_mask=snake_case )["""last_hidden_state"""] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() A__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str , snake_case : int , snake_case : Optional[Any] , *snake_case : str ): '''simple docstring''' A__ : Dict = BioGptModel(config=snake_case ).to(snake_case ).eval() A__ : Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) # first forward pass A__ : Dict = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ , A__ : List[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Optional[int] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A__ : Any = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , past_key_values=snake_case )[ """last_hidden_state""" ] # select random slice A__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Tuple , *snake_case : Union[str, Any] , snake_case : Union[str, Any]=False ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM(snake_case ) model.to(snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() A__ : Optional[Any] = model(snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self : int , snake_case : Optional[Any] , *snake_case : Optional[int] ): '''simple docstring''' A__ : int = BioGptModel(snake_case ) A__ : Union[str, Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _UpperCamelCase ( self : Any , snake_case : Dict , snake_case : Tuple , snake_case : int , snake_case : Union[str, Any] , snake_case : Dict , *snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = self.num_labels A__ : int = BioGptForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : str = config_and_inputs A__ : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case_ = (BioGptForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[str] = BioGptModelTester(self ) A__ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : str = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case , gradient_checkpointing=snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) A__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = """left""" # Define PAD Token = EOS Token = 50256 A__ : Optional[int] = tokenizer.eos_token A__ : Dict = model.config.eos_token_id # use different length sentences to test batching A__ : Union[str, Any] = [ """Hello, my dog is a little""", """Today, I""", ] A__ : List[str] = tokenizer(snake_case , return_tensors="""pt""" , padding=snake_case ) A__ : str = inputs["""input_ids"""].to(snake_case ) A__ : Dict = model.generate( input_ids=snake_case , attention_mask=inputs["""attention_mask"""].to(snake_case ) , ) A__ : Optional[int] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Any = model.generate(input_ids=snake_case ) A__ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A__ : str = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Dict = model.generate(input_ids=snake_case , max_length=model.config.max_length - num_paddings ) A__ : Optional[Any] = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case ) A__ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case ) A__ : Optional[int] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(snake_case , [non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = BioGptModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(snake_case ) A__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Union[str, Any] = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Any = 3 A__ : List[Any] = """multi_label_classification""" A__ : Dict = input_dict["""input_ids"""] A__ : Tuple = input_ids.ne(1 ).to(snake_case ) A__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Tuple = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A__ : str = torch.tensor([[2, 4805, 9, 656, 21]] ) A__ : Dict = model(snake_case )[0] A__ : Tuple = 4_2384 A__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : str = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) torch.manual_seed(0 ) A__ : Tuple = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(snake_case ) A__ : Optional[int] = model.generate( **snake_case , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case , ) A__ : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case ) A__ : List[str] = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar A_ = TypeVar('''KEY''') A_ = TypeVar('''VAL''') @dataclass(frozen=UpperCamelCase , slots=UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( Generic[KEY, VAL] ): snake_case_ = 42 snake_case_ = 42 class __SCREAMING_SNAKE_CASE ( _Item ): def __init__( self : Optional[Any] ): '''simple docstring''' super().__init__(snake_case , snake_case ) def __bool__( self : str ): '''simple docstring''' return False A_ = _DeletedItem() class __SCREAMING_SNAKE_CASE ( MutableMapping[KEY, VAL] ): def __init__( self : Any , snake_case : int = 8 , snake_case : float = 0.75 ): '''simple docstring''' A__ : List[Any] = initial_block_size A__ : list[_Item | None] = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 A__ : Dict = capacity_factor A__ : Dict = 0 def _UpperCamelCase ( self : Optional[int] , snake_case : KEY ): '''simple docstring''' return hash(snake_case ) % len(self._buckets ) def _UpperCamelCase ( self : Tuple , snake_case : int ): '''simple docstring''' return (ind + 1) % len(self._buckets ) def _UpperCamelCase ( self : Any , snake_case : int , snake_case : KEY , snake_case : VAL ): '''simple docstring''' A__ : int = self._buckets[ind] if not stored: A__ : Optional[Any] = _Item(snake_case , snake_case ) self._len += 1 return True elif stored.key == key: A__ : int = _Item(snake_case , snake_case ) return True else: return False def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = len(self._buckets ) * self._capacity_factor return len(self ) >= int(snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' if len(self._buckets ) <= self._initial_block_size: return False A__ : List[Any] = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def _UpperCamelCase ( self : Any , snake_case : int ): '''simple docstring''' A__ : int = self._buckets A__ : Dict = [None] * new_size A__ : Union[str, Any] = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' self._resize(len(self._buckets ) * 2 ) def _UpperCamelCase ( self : Any ): '''simple docstring''' self._resize(len(self._buckets ) // 2 ) def _UpperCamelCase ( self : Optional[Any] , snake_case : KEY ): '''simple docstring''' A__ : Optional[Any] = self._get_bucket_index(snake_case ) for _ in range(len(self._buckets ) ): yield ind A__ : Tuple = self._get_next_ind(snake_case ) def _UpperCamelCase ( self : Dict , snake_case : KEY , snake_case : VAL ): '''simple docstring''' for ind in self._iterate_buckets(snake_case ): if self._try_set(snake_case , snake_case , snake_case ): break def __setitem__( self : str , snake_case : KEY , snake_case : VAL ): '''simple docstring''' if self._is_full(): self._size_up() self._add_item(snake_case , snake_case ) def __delitem__( self : Optional[int] , snake_case : KEY ): '''simple docstring''' for ind in self._iterate_buckets(snake_case ): A__ : List[str] = self._buckets[ind] if item is None: raise KeyError(snake_case ) if item is _deleted: continue if item.key == key: A__ : Optional[int] = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__( self : str , snake_case : KEY ): '''simple docstring''' for ind in self._iterate_buckets(snake_case ): A__ : List[str] = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(snake_case ) def __len__( self : Any ): '''simple docstring''' return self._len def __iter__( self : Tuple ): '''simple docstring''' yield from (item.key for item in self._buckets if item) def __repr__( self : Union[str, Any] ): '''simple docstring''' A__ : int = """ ,""".join( F'{item.key}: {item.val}' for item in self._buckets if item ) return F'HashMap({val_string})'
296
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : list[list] ) ->list[list]: A__ : Tuple = current_set.copy() for row_index, row in enumerate(UpperCAmelCase__ ): A__ : Union[str, Any] = row[0] for column_index, column in enumerate(UpperCAmelCase__ ): if magnitude == 0: A__ : int = column continue A__ : Tuple = column / magnitude # Subtract to cancel term A__ : int = current_set[0] A__ : Any = [first_row] A__ : Optional[Any] = current_set[1::] for row in current_set: A__ : Dict = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(UpperCAmelCase__ ) continue for column_index in range(len(UpperCAmelCase__ ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(UpperCAmelCase__ ) # Create next recursion iteration set if len(final_set[0] ) != 3: A__ : int = final_set[0] A__ : Union[str, Any] = [] A__ : Dict = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) A__ : int = simplify(UpperCAmelCase__ ) for i in range(len(UpperCAmelCase__ ) ): resultant[i].insert(0, current_first_column[i] ) resultant.insert(0, UpperCAmelCase__ ) A__ : int = resultant return final_set def _lowerCAmelCase ( UpperCAmelCase__ : list[list] ) ->list: if len(UpperCAmelCase__ ) == 0: raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) A__ : Tuple = len(UpperCAmelCase__ ) + 1 if any(len(UpperCAmelCase__ ) != _length for item in equations ): raise IndexError("""solve_simultaneous() requires n lists of length n+1""" ) for row in equations: if any(not isinstance(UpperCAmelCase__, (int, float) ) for column in row ): raise ValueError("""solve_simultaneous() requires lists of integers""" ) if len(UpperCAmelCase__ ) == 1: return [equations[0][-1] / equations[0][0]] A__ : str = equations.copy() if any(0 in row for row in data_set ): A__ : List[str] = data_set.copy() A__ : Tuple = [] for row_index, row in enumerate(UpperCAmelCase__ ): if 0 not in row: A__ : str = data_set.pop(UpperCAmelCase__ ) break if not full_row: raise ValueError("""solve_simultaneous() requires at least 1 full equation""" ) data_set.insert(0, UpperCAmelCase__ ) A__ : int = data_set.copy() A__ : int = simplify(UpperCAmelCase__ ) A__ : int = simplified[::-1] A__ : list = [] for row in simplified: A__ : int = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue A__ : int = row.copy()[: len(UpperCAmelCase__ ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(UpperCAmelCase__ ) == 0: solutions.append(0 ) continue A__ : Dict = temp_row[1::] A__ : Any = temp_row[::-1] for column_index, column in enumerate(UpperCAmelCase__ ): current_solution -= column * solutions[column_index] solutions.append(UpperCAmelCase__ ) A__ : Optional[int] = [] for item in solutions: final.append(float(round(UpperCAmelCase__, 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() A_ = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
1
"""simple docstring""" from __future__ import annotations def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : List[Any] ) ->List[str]: # noqa: E741 while r - l > 1: A__ : List[str] = (l + r) // 2 if v[m] >= key: A__ : Any = m else: A__ : List[str] = m # noqa: E741 return r def _lowerCAmelCase ( UpperCAmelCase__ : list[int] ) ->int: if len(UpperCAmelCase__ ) == 0: return 0 A__ : Optional[Any] = [0] * len(UpperCAmelCase__ ) A__ : Tuple = 1 A__ : int = v[0] for i in range(1, len(UpperCAmelCase__ ) ): if v[i] < tail[0]: A__ : List[Any] = v[i] elif v[i] > tail[length - 1]: A__ : Dict = v[i] length += 1 else: A__ : Optional[int] = v[i] return length if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A_ = '''src/diffusers''' A_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. A_ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) A_ = spec.loader.load_module() def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any] ) ->Any: return line.startswith(UpperCAmelCase__ ) or len(UpperCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", UpperCAmelCase__ ) is not None def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Union[str, Any]: A__ : Any = object_name.split(""".""" ) A__ : int = 0 # First let's find the module where our object lives. A__ : str = parts[i] while i < len(UpperCAmelCase__ ) and not os.path.isfile(os.path.join(UpperCAmelCase__, f'{module}.py' ) ): i += 1 if i < len(UpperCAmelCase__ ): A__ : Union[str, Any] = os.path.join(UpperCAmelCase__, parts[i] ) if i >= len(UpperCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(UpperCAmelCase__, f'{module}.py' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : List[Any] = f.readlines() # Now let's find the class / func in the code! A__ : Optional[Any] = """""" A__ : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A__ : List[Any] = line_index while line_index < len(UpperCAmelCase__ ) and _should_continue(lines[line_index], UpperCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : List[Any] = lines[start_index:line_index] return "".join(UpperCAmelCase__ ) A_ = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') A_ = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') A_ = re.compile(r'''<FILL\s+[^>]*>''') def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Optional[Any]: A__ : Dict = code.split("""\n""" ) A__ : List[Any] = 0 while idx < len(UpperCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase__ ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->int: A__ : str = len(get_indent(UpperCAmelCase__ ) ) > 0 if has_indent: A__ : Union[str, Any] = f'class Bla:\n{code}' A__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=1_1_9, preview=UpperCAmelCase__ ) A__ : Tuple = black.format_str(UpperCAmelCase__, mode=UpperCAmelCase__ ) A__ , A__ : List[Any] = style_docstrings_in_code(UpperCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : Dict=False ) ->List[Any]: with open(UpperCAmelCase__, """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : int = f.readlines() A__ : Dict = [] A__ : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase__ ): A__ : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A__ , A__ , A__ : Dict = search.groups() A__ : Tuple = find_code_in_diffusers(UpperCAmelCase__ ) A__ : int = get_indent(UpperCAmelCase__ ) A__ : List[str] = line_index + 1 if indent == theoretical_indent else line_index + 2 A__ : Tuple = theoretical_indent A__ : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A__ : Tuple = True while line_index < len(UpperCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase__ ): break A__ : Optional[int] = lines[line_index] A__ : Tuple = _should_continue(UpperCAmelCase__, UpperCAmelCase__ ) and re.search(f'^{indent}# End copy', UpperCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : Dict = lines[start_index:line_index] A__ : Tuple = """""".join(UpperCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies A__ : Optional[int] = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase__ ) is None] A__ : Optional[Any] = """\n""".join(UpperCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase__ ) > 0: A__ : int = replace_pattern.replace("""with""", """""" ).split(""",""" ) A__ : List[Any] = [_re_replace_pattern.search(UpperCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue A__ , A__ , A__ : Union[str, Any] = pattern.groups() A__ : Union[str, Any] = re.sub(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if option.strip() == "all-casing": A__ : List[Any] = re.sub(obja.lower(), obja.lower(), UpperCAmelCase__ ) A__ : Tuple = re.sub(obja.upper(), obja.upper(), UpperCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A__ : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) A__ : List[Any] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A__ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] A__ : Tuple = start_index + 1 if overwrite and len(UpperCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(UpperCAmelCase__, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(UpperCAmelCase__ ) return diffs def _lowerCAmelCase ( UpperCAmelCase__ : bool = False ) ->Any: A__ : Dict = glob.glob(os.path.join(UpperCAmelCase__, """**/*.py""" ), recursive=UpperCAmelCase__ ) A__ : str = [] for filename in all_files: A__ : Any = is_copy_consistent(UpperCAmelCase__, UpperCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(UpperCAmelCase__ ) > 0: A__ : Any = """\n""".join(UpperCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() check_copies(args.fix_and_overwrite)
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available A_ = { '''configuration_bigbird_pegasus''': [ '''BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''BigBirdPegasusConfig''', '''BigBirdPegasusOnnxConfig''', ], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST''', '''BigBirdPegasusForCausalLM''', '''BigBirdPegasusForConditionalGeneration''', '''BigBirdPegasusForQuestionAnswering''', '''BigBirdPegasusForSequenceClassification''', '''BigBirdPegasusModel''', '''BigBirdPegasusPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_CONFIG_ARCHIVE_MAP, BigBirdPegasusConfig, BigBirdPegasusOnnxConfig, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bigbird_pegasus import ( BIGBIRD_PEGASUS_PRETRAINED_MODEL_ARCHIVE_LIST, BigBirdPegasusForCausalLM, BigBirdPegasusForConditionalGeneration, BigBirdPegasusForQuestionAnswering, BigBirdPegasusForSequenceClassification, BigBirdPegasusModel, BigBirdPegasusPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
1
"""simple docstring""" import logging import torch from accelerate import Accelerator from arguments import EvaluationArguments from datasets import load_dataset from torch.utils.data import IterableDataset from torch.utils.data.dataloader import DataLoader from transformers import AutoModelForCausalLM, AutoTokenizer, HfArgumentParser, set_seed class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Any , snake_case : Any , snake_case : Any , snake_case : Any=1024 , snake_case : Optional[int]=1024 , snake_case : Optional[Any]=3.6 ): '''simple docstring''' A__ : Optional[int] = tokenizer A__ : Optional[int] = tokenizer.bos_token_id A__ : Dict = dataset A__ : Optional[int] = seq_length A__ : Union[str, Any] = seq_length * chars_per_token * num_of_sequences def __iter__( self : Any ): '''simple docstring''' A__ : int = iter(self.dataset ) A__ : str = True while more_examples: A__ , A__ : Any = [], 0 while True: if buffer_len >= self.input_characters: break try: buffer.append(next(snake_case )["""content"""] ) buffer_len += len(buffer[-1] ) except StopIteration: A__ : str = False break A__ : Optional[Any] = tokenizer(snake_case , truncation=snake_case )["""input_ids"""] A__ : Optional[Any] = [] for tokenized_input in tokenized_inputs: all_token_ids.extend(tokenized_input + [self.concat_token_id] ) for i in range(0 , len(snake_case ) , self.seq_length ): A__ : int = all_token_ids[i : i + self.seq_length] if len(snake_case ) == self.seq_length: yield torch.tensor(snake_case ) def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->Dict: A__ : List[Any] = {"""streaming""": True} A__ : Optional[int] = load_dataset(args.dataset_name, split="""train""", **UpperCAmelCase__ ) A__ : str = ConstantLengthDataset(UpperCAmelCase__, UpperCAmelCase__, seq_length=args.seq_length ) A__ : Union[str, Any] = DataLoader(UpperCAmelCase__, batch_size=args.batch_size ) return eval_dataloader def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->int: model.eval() A__ : Tuple = [] for step, batch in enumerate(UpperCAmelCase__ ): with torch.no_grad(): A__ : List[str] = model(UpperCAmelCase__, labels=UpperCAmelCase__ ) A__ : Dict = outputs.loss.repeat(args.batch_size ) losses.append(accelerator.gather(UpperCAmelCase__ ) ) if args.max_eval_steps > 0 and step >= args.max_eval_steps: break A__ : Union[str, Any] = torch.mean(torch.cat(UpperCAmelCase__ ) ) try: A__ : Union[str, Any] = torch.exp(UpperCAmelCase__ ) except OverflowError: A__ : Union[str, Any] = float("""inf""" ) return loss.item(), perplexity.item() # Setup Accelerator A_ = Accelerator() # Parse configuration A_ = HfArgumentParser(EvaluationArguments) A_ = parser.parse_args() set_seed(args.seed) # Logging A_ = logging.getLogger(__name__) logging.basicConfig( format='''%(asctime)s - %(levelname)s - %(name)s - %(message)s''', datefmt='''%m/%d/%Y %H:%M:%S''', level=logging.INFO ) # Load model and tokenizer A_ = AutoModelForCausalLM.from_pretrained(args.model_ckpt) A_ = AutoTokenizer.from_pretrained(args.model_ckpt) # Load dataset and dataloader A_ = create_dataloader(args) # Prepare everything with our `accelerator`. A_ , A_ = accelerator.prepare(model, eval_dataloader) # Evaluate and save the last checkpoint logger.info('''Evaluating and saving model after training''') A_ , A_ = evaluate(args) logger.info(F'loss/eval: {eval_loss}, perplexity: {perplexity}')
296
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels A_ = object() # For specifying empty leaf dict `{}` A_ = object() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any] ) ->Dict: A__ : Union[str, Any] = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) + 1 ): A__ : Optional[Any] = [x.match(UpperCAmelCase__ ) for x, y in zip(UpperCAmelCase__, ks[i:] )] if matches and all(UpperCAmelCase__ ): return True return False def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Dict: def replace(UpperCAmelCase__ : int, UpperCAmelCase__ : List[str] ): for rule, replacement in rules: if _match(UpperCAmelCase__, UpperCAmelCase__ ): return replacement return val return replace def _lowerCAmelCase ( ) ->Tuple: return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""", UpperCAmelCase__ )), (("transformer", "wte", "embedding"), P("""mp""", UpperCAmelCase__ )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(UpperCAmelCase__, """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(UpperCAmelCase__, """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : Union[str, Any] = _get_partition_rules() A__ : int = _replacement_rules(UpperCAmelCase__ ) A__ : Tuple = {k: _unmatched for k in flatten_dict(UpperCAmelCase__ )} A__ : Optional[int] = {k: replace(UpperCAmelCase__, UpperCAmelCase__ ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(UpperCAmelCase__ ) )
296
1
"""simple docstring""" from torch import nn def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int] ) ->Any: if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'Unsupported activation function: {act_fn}' )
296
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] , snake_case : Tuple , snake_case : List[str]=2 , snake_case : List[str]=8 , snake_case : List[Any]=True , snake_case : Optional[Any]=True , snake_case : List[Any]=True , snake_case : Dict=True , snake_case : Tuple=99 , snake_case : Dict=16 , snake_case : Dict=5 , snake_case : int=2 , snake_case : Any=36 , snake_case : str="gelu" , snake_case : Dict=0.0 , snake_case : List[Any]=0.0 , snake_case : int=512 , snake_case : List[Any]=16 , snake_case : Tuple=2 , snake_case : Any=0.02 , snake_case : Optional[Any]=3 , snake_case : List[Any]=4 , snake_case : str=None , ): '''simple docstring''' A__ : Union[str, Any] = parent A__ : Optional[Any] = batch_size A__ : Dict = seq_length A__ : str = is_training A__ : Tuple = use_input_mask A__ : Dict = use_token_type_ids A__ : Dict = use_labels A__ : int = vocab_size A__ : List[str] = hidden_size A__ : Union[str, Any] = num_hidden_layers A__ : int = num_attention_heads A__ : List[str] = intermediate_size A__ : int = hidden_act A__ : str = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : Any = max_position_embeddings A__ : Optional[int] = type_vocab_size A__ : int = type_sequence_label_size A__ : Optional[Any] = initializer_range A__ : int = num_labels A__ : Optional[int] = num_choices A__ : Optional[int] = scope def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Any = None if self.use_input_mask: A__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Optional[int] = None if self.use_token_type_ids: A__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Dict = None A__ : List[str] = None A__ : Union[str, Any] = None if self.use_labels: A__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Any = ids_tensor([self.batch_size] , self.num_choices ) A__ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.get_config() A__ : List[str] = 300 return config def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Tuple = self.prepare_config_and_inputs() A__ : List[str] = True A__ : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _UpperCamelCase ( self : Any , snake_case : Any , snake_case : Tuple , snake_case : Any , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Dict ): '''simple docstring''' A__ : List[str] = MraModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A__ : List[str] = model(snake_case , token_type_ids=snake_case ) A__ : Union[str, Any] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : List[Any] , snake_case : Any , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Dict , snake_case : str , snake_case : Dict , snake_case : str , ): '''simple docstring''' A__ : Dict = True A__ : Optional[Any] = MraModel(snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , ) A__ : Optional[int] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : str , snake_case : Union[str, Any] , snake_case : Dict , snake_case : List[str] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Dict , snake_case : Dict , snake_case : Dict , snake_case : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : Dict = MraForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : Tuple , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Optional[int] , snake_case : List[str] , snake_case : Union[str, Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Optional[Any] = MraForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict , snake_case : str , snake_case : List[Any] , snake_case : Any , snake_case : Dict , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Union[str, Any] = MraForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : Tuple , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Dict , snake_case : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.num_choices A__ : str = MraForMultipleChoice(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Dict = config_and_inputs A__ : Optional[int] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = () def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Optional[Any] = MraModelTester(self ) A__ : List[str] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : List[str] = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : str = MraModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip(reason="""MRA does not output attentions""" ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Any = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : List[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , snake_case ) A__ : int = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Dict = 5_0265 A__ : List[str] = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : List[Any] = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) A__ : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Union[str, Any] = 5_0265 A__ : Optional[Any] = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : Optional[int] = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) )
296
1
"""simple docstring""" # Lint as: python3 import itertools import os import re A_ = re.compile(r'''([A-Z]+)([A-Z][a-z])''') A_ = re.compile(r'''([a-z\d])([A-Z])''') A_ = re.compile(r'''(?<!_)_(?!_)''') A_ = re.compile(r'''(_{2,})''') A_ = r'''^\w+(\.\w+)*$''' A_ = r'''<>:/\|?*''' def _lowerCAmelCase ( UpperCAmelCase__ : Any ) ->int: A__ : Any = _uppercase_uppercase_re.sub(R"""\1_\2""", UpperCAmelCase__ ) A__ : Dict = _lowercase_uppercase_re.sub(R"""\1_\2""", UpperCAmelCase__ ) return name.lower() def _lowerCAmelCase ( UpperCAmelCase__ : Any ) ->int: A__ : Tuple = _single_underscore_re.split(UpperCAmelCase__ ) A__ : List[Any] = [_multiple_underscores_re.split(UpperCAmelCase__ ) for n in name] return "".join(n.capitalize() for n in itertools.chain.from_iterable(UpperCAmelCase__ ) if n != """""" ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->Any: if os.path.basename(UpperCAmelCase__ ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) return camelcase_to_snakecase(UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Optional[int] ) ->Dict: if os.path.basename(UpperCAmelCase__ ) != name: raise ValueError(f'Should be a dataset name, not a path: {name}' ) if not re.match(_split_re, UpperCAmelCase__ ): raise ValueError(f'Split name should match \'{_split_re}\'\' but got \'{split}\'.' ) return f'{filename_prefix_for_name(UpperCAmelCase__ )}-{split}' def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : List[Any]=None ) ->str: A__ : Union[str, Any] = filename_prefix_for_split(UpperCAmelCase__, UpperCAmelCase__ ) if filetype_suffix: prefix += f'.{filetype_suffix}' A__ : List[Any] = os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) return f'{filepath}*' def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : int=None ) ->Dict: A__ : Union[str, Any] = filename_prefix_for_split(UpperCAmelCase__, UpperCAmelCase__ ) A__ : Dict = os.path.join(UpperCAmelCase__, UpperCAmelCase__ ) if shard_lengths: A__ : Any = len(UpperCAmelCase__ ) A__ : Optional[Any] = [f'{prefix}-{shard_id:05d}-of-{num_shards:05d}' for shard_id in range(UpperCAmelCase__ )] if filetype_suffix: A__ : int = [filename + f'.{filetype_suffix}' for filename in filenames] return filenames else: A__ : str = prefix if filetype_suffix: filename += f'.{filetype_suffix}' return [filename]
296
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets A_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' A_ = '''\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. ''' A_ = ''' Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {"raw_values", "uniform_average"} or array-like of shape (n_outputs,), default="uniform_average" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. "raw_values" : Returns a full set of errors in case of multioutput input. "uniform_average" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric("mse") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {\'mse\': 0.6123724356957945} If you\'re using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric("mse", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput=\'raw_values\') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mse\': array([0.41666667, 1. ])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : Dict ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html""" ] , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("""float""" ) ), "references": datasets.Sequence(datasets.Value("""float""" ) ), } else: return { "predictions": datasets.Value("""float""" ), "references": datasets.Value("""float""" ), } def _UpperCamelCase ( self : List[str] , snake_case : Dict , snake_case : List[Any] , snake_case : List[str]=None , snake_case : List[Any]="uniform_average" , snake_case : int=True ): '''simple docstring''' A__ : Optional[int] = mean_squared_error( snake_case , snake_case , sample_weight=snake_case , multioutput=snake_case , squared=snake_case ) return {"mse": mse}
296
1
"""simple docstring""" import random import torch from huggingface_hub import HfApi from diffusers import UNetaDModel A_ = HfApi() A_ = {} # fmt: off A_ = torch.tensor([ -0.7515, -1.6883, 0.2420, 0.0300, 0.6347, 1.3433, -1.1743, -3.7467, 1.2342, -2.2485, 0.4636, 0.8076, -0.7991, 0.3969, 0.8498, 0.9189, -1.8887, -3.3522, 0.7639, 0.2040, 0.6271, -2.7148, -1.6316, 3.0839, 0.3186, 0.2721, -0.9759, -1.2461, 2.6257, 1.3557 ]) A_ = torch.tensor([ -2.3639, -2.5344, 0.0054, -0.6674, 1.5990, 1.0158, 0.3124, -2.1436, 1.8795, -2.5429, -0.1566, -0.3973, 1.2490, 2.6447, 1.2283, -0.5208, -2.8154, -3.5119, 2.3838, 1.2033, 1.7201, -2.1256, -1.4576, 2.7948, 2.4204, -0.9752, -1.2546, 0.8027, 3.2758, 3.1365 ]) A_ = torch.tensor([ -0.6531, -0.6891, -0.3172, -0.5375, -0.9140, -0.5367, -0.1175, -0.7869, -0.3808, -0.4513, -0.2098, -0.0083, 0.3183, 0.5140, 0.2247, -0.1304, -0.1302, -0.2802, -0.2084, -0.2025, -0.4967, -0.4873, -0.0861, 0.6925, 0.0250, 0.1290, -0.1543, 0.6316, 1.0460, 1.4943 ]) A_ = torch.tensor([ 0.0911, 0.1107, 0.0182, 0.0435, -0.0805, -0.0608, 0.0381, 0.2172, -0.0280, 0.1327, -0.0299, -0.0255, -0.0050, -0.1170, -0.1046, 0.0309, 0.1367, 0.1728, -0.0533, -0.0748, -0.0534, 0.1624, 0.0384, -0.1805, -0.0707, 0.0642, 0.0220, -0.0134, -0.1333, -0.1505 ]) A_ = torch.tensor([ 0.1321, 0.1337, 0.0440, 0.0622, -0.0591, -0.0370, 0.0503, 0.2133, -0.0177, 0.1415, -0.0116, -0.0112, 0.0044, -0.0980, -0.0789, 0.0395, 0.1502, 0.1785, -0.0488, -0.0514, -0.0404, 0.1539, 0.0454, -0.1559, -0.0665, 0.0659, 0.0383, -0.0005, -0.1266, -0.1386 ]) A_ = torch.tensor([ 0.1154, 0.1218, 0.0307, 0.0526, -0.0711, -0.0541, 0.0366, 0.2078, -0.0267, 0.1317, -0.0226, -0.0193, -0.0014, -0.1055, -0.0902, 0.0330, 0.1391, 0.1709, -0.0562, -0.0693, -0.0560, 0.1482, 0.0381, -0.1683, -0.0681, 0.0661, 0.0331, -0.0046, -0.1268, -0.1431 ]) A_ = torch.tensor([ 0.1192, 0.1240, 0.0414, 0.0606, -0.0557, -0.0412, 0.0430, 0.2042, -0.0200, 0.1385, -0.0115, -0.0132, 0.0017, -0.0965, -0.0802, 0.0398, 0.1433, 0.1747, -0.0458, -0.0533, -0.0407, 0.1545, 0.0419, -0.1574, -0.0645, 0.0626, 0.0341, -0.0010, -0.1199, -0.1390 ]) A_ = torch.tensor([ 0.1075, 0.1074, 0.0205, 0.0431, -0.0774, -0.0607, 0.0298, 0.2042, -0.0320, 0.1267, -0.0281, -0.0250, -0.0064, -0.1091, -0.0946, 0.0290, 0.1328, 0.1650, -0.0580, -0.0738, -0.0586, 0.1440, 0.0337, -0.1746, -0.0712, 0.0605, 0.0250, -0.0099, -0.1316, -0.1473 ]) A_ = torch.tensor([ -1.4572, -2.0481, -0.0414, -0.6005, 1.4136, 0.5848, 0.4028, -2.7330, 1.2212, -2.1228, 0.2155, 0.4039, 0.7662, 2.0535, 0.7477, -0.3243, -2.1758, -2.7648, 1.6947, 0.7026, 1.2338, -1.6078, -0.8682, 2.2810, 1.8574, -0.5718, -0.5586, -0.0186, 2.3415, 2.1251]) A_ = torch.tensor([ -1.3690, -1.9720, -0.4090, -0.6966, 1.4660, 0.9938, -0.1385, -2.7324, 0.7736, -1.8917, 0.2923, 0.4293, 0.1693, 1.4112, 1.1887, -0.3181, -2.2160, -2.6381, 1.3170, 0.8163, 0.9240, -1.6544, -0.6099, 2.5259, 1.6430, -0.9090, -0.9392, -0.0126, 2.4268, 2.3266 ]) A_ = torch.tensor([ -1.3525, -1.9628, -0.3956, -0.6860, 1.4664, 1.0014, -0.1259, -2.7212, 0.7772, -1.8811, 0.2996, 0.4388, 0.1704, 1.4029, 1.1701, -0.3027, -2.2053, -2.6287, 1.3350, 0.8131, 0.9274, -1.6292, -0.6098, 2.5131, 1.6505, -0.8958, -0.9298, -0.0151, 2.4257, 2.3355 ]) A_ = torch.tensor([ -2.0585, -2.7897, -0.2850, -0.8940, 1.9052, 0.5702, 0.6345, -3.8959, 1.5932, -3.2319, 0.1974, 0.0287, 1.7566, 2.6543, 0.8387, -0.5351, -3.2736, -4.3375, 2.9029, 1.6390, 1.4640, -2.1701, -1.9013, 2.9341, 3.4981, -0.6255, -1.1644, -0.1591, 3.7097, 3.2066 ]) A_ = torch.tensor([ -2.3139, -2.5594, -0.0197, -0.6785, 1.7001, 1.1606, 0.3075, -2.1740, 1.8071, -2.5630, -0.0926, -0.3811, 1.2116, 2.6246, 1.2731, -0.5398, -2.8153, -3.6140, 2.3893, 1.3262, 1.6258, -2.1856, -1.3267, 2.8395, 2.3779, -1.0623, -1.2468, 0.8959, 3.3367, 3.2243 ]) A_ = torch.tensor([ -2.0628, -2.7667, -0.2089, -0.8263, 2.0539, 0.5992, 0.6495, -3.8336, 1.6025, -3.2817, 0.1721, -0.0633, 1.7516, 2.7039, 0.8100, -0.5908, -3.2113, -4.4343, 2.9257, 1.3632, 1.5562, -2.1489, -1.9894, 3.0560, 3.3396, -0.7328, -1.0417, 0.0383, 3.7093, 3.2343 ]) A_ = torch.tensor([ -1.4574, -2.0569, -0.0473, -0.6117, 1.4018, 0.5769, 0.4129, -2.7344, 1.2241, -2.1397, 0.2000, 0.3937, 0.7616, 2.0453, 0.7324, -0.3391, -2.1746, -2.7744, 1.6963, 0.6921, 1.2187, -1.6172, -0.8877, 2.2439, 1.8471, -0.5839, -0.5605, -0.0464, 2.3250, 2.1219 ]) # fmt: on A_ = api.list_models(filter='''diffusers''') for mod in models: if "google" in mod.author or mod.modelId == "CompVis/ldm-celebahq-256": A_ = '''/home/patrick/google_checkpoints/''' + mod.modelId.split('''/''')[-1] print(F'Started running {mod.modelId}!!!') if mod.modelId.startswith('''CompVis'''): A_ = UNetaDModel.from_pretrained(local_checkpoint, subfolder='''unet''') else: A_ = UNetaDModel.from_pretrained(local_checkpoint) torch.manual_seed(0) random.seed(0) A_ = torch.randn(1, model.config.in_channels, model.config.sample_size, model.config.sample_size) A_ = torch.tensor([10] * noise.shape[0]) with torch.no_grad(): A_ = model(noise, time_step).sample assert torch.allclose( logits[0, 0, 0, :30], results['''_'''.join('''_'''.join(mod.modelId.split('''/''')).split('''-'''))], atol=1e-3 ) print(F'{mod.modelId} has passed successfully!!!')
296
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Optional[int] , snake_case : List[str]=None , **snake_case : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , snake_case , ) super().__init__(args=snake_case , **snake_case )
296
1
"""simple docstring""" from argparse import ArgumentParser from .env import EnvironmentCommand def _lowerCAmelCase ( ) ->str: A__ : str = ArgumentParser("""Diffusers CLI tool""", usage="""diffusers-cli <command> [<args>]""" ) A__ : List[str] = parser.add_subparsers(help="""diffusers-cli command helpers""" ) # Register commands EnvironmentCommand.register_subcommand(UpperCAmelCase__ ) # Let's go A__ : Union[str, Any] = parser.parse_args() if not hasattr(UpperCAmelCase__, """func""" ): parser.print_help() exit(1 ) # Run A__ : int = args.func(UpperCAmelCase__ ) service.run() if __name__ == "__main__": main()
296
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch A_ = random.Random() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Tuple=1.0, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : str=None ) ->Union[str, Any]: if rng is None: A__ : Optional[int] = global_rng A__ : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[str]=7 , snake_case : str=400 , snake_case : Optional[Any]=2000 , snake_case : Union[str, Any]=10 , snake_case : str=160 , snake_case : List[str]=8 , snake_case : List[Any]=0.0 , snake_case : Optional[Any]=4000 , snake_case : Any=False , snake_case : int=True , ): '''simple docstring''' A__ : Any = parent A__ : str = batch_size A__ : List[str] = min_seq_length A__ : Dict = max_seq_length A__ : str = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) A__ : Dict = padding_value A__ : Optional[Any] = sampling_rate A__ : Any = return_attention_mask A__ : Optional[int] = do_normalize A__ : Tuple = feature_size A__ : Optional[Any] = chunk_length A__ : Union[str, Any] = hop_length def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict=False , snake_case : Optional[Any]=False ): '''simple docstring''' def _flatten(snake_case : Dict ): return list(itertools.chain(*snake_case ) ) if equal_length: A__ : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size A__ : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: A__ : List[str] = [np.asarray(snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = WhisperFeatureExtractor if is_speech_available() else None def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : str = WhisperFeatureExtractionTester(self ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] = feat_extract_first.save_pretrained(snake_case )[0] check_json_file_has_correct_format(snake_case ) A__ : Union[str, Any] = self.feature_extraction_class.from_pretrained(snake_case ) A__ : str = feat_extract_first.to_dict() A__ : Union[str, Any] = feat_extract_second.to_dict() A__ : List[Any] = feat_extract_first.mel_filters A__ : Optional[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : Any = os.path.join(snake_case , """feat_extract.json""" ) feat_extract_first.to_json_file(snake_case ) A__ : int = self.feature_extraction_class.from_json_file(snake_case ) A__ : Dict = feat_extract_first.to_dict() A__ : str = feat_extract_second.to_dict() A__ : str = feat_extract_first.mel_filters A__ : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 A__ : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] # Test feature size A__ : Dict = feature_extractor(snake_case , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input A__ : str = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test batched A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. A__ : Tuple = [floats_list((1, x) )[0] for x in (800, 800, 800)] A__ : str = np.asarray(snake_case ) A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test truncation required A__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] A__ : Union[str, Any] = [x[: feature_extractor.n_samples] for x in speech_inputs] A__ : str = [np.asarray(snake_case ) for speech_input in speech_inputs_truncated] A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : str = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' import torch A__ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : List[str] = np.random.rand(100 , 32 ).astype(np.floataa ) A__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: A__ : Optional[Any] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) A__ : Optional[int] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' A__ : int = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech A__ : Union[str, Any] = ds.sort("""id""" ).select(range(snake_case ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on A__ : Optional[Any] = self._load_datasamples(1 ) A__ : Union[str, Any] = WhisperFeatureExtractor() A__ : List[str] = feature_extractor(snake_case , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , snake_case , atol=1e-4 ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : Union[str, Any] = self._load_datasamples(1 )[0] A__ : Any = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue A__ : str = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=snake_case )[0] self.assertTrue(np.all(np.mean(snake_case ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case ) - 1 ) < 1e-3 ) )
296
1
"""simple docstring""" from typing import Dict, List, Optional, Union import numpy as np from ...image_processing_utils import BaseImageProcessor, BatchFeature, get_size_dict from ...image_transforms import ( center_crop, convert_to_rgb, get_resize_output_image_size, normalize, rescale, resize, to_channel_dimension_format, ) from ...image_utils import ( OPENAI_CLIP_MEAN, OPENAI_CLIP_STD, ChannelDimension, ImageInput, PILImageResampling, make_list_of_images, to_numpy_array, valid_images, ) from ...utils import TensorType, is_vision_available, logging A_ = logging.get_logger(__name__) if is_vision_available(): import PIL class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = ['pixel_values'] def __init__( self : str , snake_case : bool = True , snake_case : Dict[str, int] = None , snake_case : PILImageResampling = PILImageResampling.BICUBIC , snake_case : bool = True , snake_case : Dict[str, int] = None , snake_case : bool = True , snake_case : Union[int, float] = 1 / 255 , snake_case : bool = True , snake_case : Optional[Union[float, List[float]]] = None , snake_case : Optional[Union[float, List[float]]] = None , snake_case : bool = True , **snake_case : List[Any] , ): '''simple docstring''' super().__init__(**snake_case ) A__ : Dict = size if size is not None else {"""shortest_edge""": 224} A__ : Any = get_size_dict(snake_case , default_to_square=snake_case ) A__ : Any = crop_size if crop_size is not None else {"""height""": 224, """width""": 224} A__ : List[str] = get_size_dict(snake_case , default_to_square=snake_case , param_name="""crop_size""" ) A__ : int = do_resize A__ : Dict = size A__ : Tuple = resample A__ : Dict = do_center_crop A__ : Optional[Any] = crop_size A__ : Any = do_rescale A__ : Dict = rescale_factor A__ : Any = do_normalize A__ : Any = image_mean if image_mean is not None else OPENAI_CLIP_MEAN A__ : Optional[int] = image_std if image_std is not None else OPENAI_CLIP_STD A__ : Union[str, Any] = do_convert_rgb def _UpperCamelCase ( self : Union[str, Any] , snake_case : np.ndarray , snake_case : Dict[str, int] , snake_case : PILImageResampling = PILImageResampling.BICUBIC , snake_case : Optional[Union[str, ChannelDimension]] = None , **snake_case : Optional[int] , ): '''simple docstring''' A__ : Dict = get_size_dict(snake_case , default_to_square=snake_case ) if "shortest_edge" not in size: raise ValueError(F'The `size` parameter must contain the key `shortest_edge`. Got {size.keys()}' ) A__ : int = get_resize_output_image_size(snake_case , size=size["""shortest_edge"""] , default_to_square=snake_case ) return resize(snake_case , size=snake_case , resample=snake_case , data_format=snake_case , **snake_case ) def _UpperCamelCase ( self : Dict , snake_case : np.ndarray , snake_case : Dict[str, int] , snake_case : Optional[Union[str, ChannelDimension]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : List[Any] = get_size_dict(snake_case ) if "height" not in size or "width" not in size: raise ValueError(F'The `size` parameter must contain the keys (height, width). Got {size.keys()}' ) return center_crop(snake_case , size=(size["""height"""], size["""width"""]) , data_format=snake_case , **snake_case ) def _UpperCamelCase ( self : Tuple , snake_case : np.ndarray , snake_case : Union[int, float] , snake_case : Optional[Union[str, ChannelDimension]] = None , **snake_case : Tuple , ): '''simple docstring''' return rescale(snake_case , scale=snake_case , data_format=snake_case , **snake_case ) def _UpperCamelCase ( self : Optional[Any] , snake_case : np.ndarray , snake_case : Union[float, List[float]] , snake_case : Union[float, List[float]] , snake_case : Optional[Union[str, ChannelDimension]] = None , **snake_case : List[str] , ): '''simple docstring''' return normalize(snake_case , mean=snake_case , std=snake_case , data_format=snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : ImageInput , snake_case : bool = None , snake_case : Dict[str, int] = None , snake_case : PILImageResampling = None , snake_case : bool = None , snake_case : int = None , snake_case : bool = None , snake_case : float = None , snake_case : bool = None , snake_case : Optional[Union[float, List[float]]] = None , snake_case : Optional[Union[float, List[float]]] = None , snake_case : bool = None , snake_case : Optional[Union[str, TensorType]] = None , snake_case : Optional[ChannelDimension] = ChannelDimension.FIRST , **snake_case : Optional[int] , ): '''simple docstring''' A__ : str = do_resize if do_resize is not None else self.do_resize A__ : Tuple = size if size is not None else self.size A__ : Tuple = get_size_dict(snake_case , param_name="""size""" , default_to_square=snake_case ) A__ : Tuple = resample if resample is not None else self.resample A__ : str = do_center_crop if do_center_crop is not None else self.do_center_crop A__ : List[Any] = crop_size if crop_size is not None else self.crop_size A__ : Union[str, Any] = get_size_dict(snake_case , param_name="""crop_size""" , default_to_square=snake_case ) A__ : int = do_rescale if do_rescale is not None else self.do_rescale A__ : List[Any] = rescale_factor if rescale_factor is not None else self.rescale_factor A__ : Optional[int] = do_normalize if do_normalize is not None else self.do_normalize A__ : List[Any] = image_mean if image_mean is not None else self.image_mean A__ : Optional[int] = image_std if image_std is not None else self.image_std A__ : Optional[int] = do_convert_rgb if do_convert_rgb is not None else self.do_convert_rgb A__ : int = make_list_of_images(snake_case ) if not valid_images(snake_case ): raise ValueError( """Invalid image type. Must be of type PIL.Image.Image, numpy.ndarray, """ """torch.Tensor, tf.Tensor or jax.ndarray.""" ) if do_resize and size is None: raise ValueError("""Size must be specified if do_resize is True.""" ) if do_center_crop and crop_size is None: raise ValueError("""Crop size must be specified if do_center_crop is True.""" ) if do_rescale and rescale_factor is None: raise ValueError("""Rescale factor must be specified if do_rescale is True.""" ) if do_normalize and (image_mean is None or image_std is None): raise ValueError("""Image mean and std must be specified if do_normalize is True.""" ) # PIL RGBA images are converted to RGB if do_convert_rgb: A__ : Any = [convert_to_rgb(snake_case ) for image in images] # All transformations expect numpy arrays. A__ : Union[str, Any] = [to_numpy_array(snake_case ) for image in images] if do_resize: A__ : Optional[int] = [self.resize(image=snake_case , size=snake_case , resample=snake_case ) for image in images] if do_center_crop: A__ : List[Any] = [self.center_crop(image=snake_case , size=snake_case ) for image in images] if do_rescale: A__ : int = [self.rescale(image=snake_case , scale=snake_case ) for image in images] if do_normalize: A__ : int = [self.normalize(image=snake_case , mean=snake_case , std=snake_case ) for image in images] A__ : List[str] = [to_channel_dimension_format(snake_case , snake_case ) for image in images] A__ : Union[str, Any] = {"""pixel_values""": images} return BatchFeature(data=snake_case , tensor_type=snake_case )
296
"""simple docstring""" import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] ): '''simple docstring''' A__ : Optional[int] = (0, 0) A__ : Dict = None A__ : int = 0 A__ : str = 0 A__ : Optional[Any] = 0 def __eq__( self : str , snake_case : Optional[int] ): '''simple docstring''' return self.position == cell.position def _UpperCamelCase ( self : List[str] ): '''simple docstring''' print(self.position ) class __SCREAMING_SNAKE_CASE : def __init__( self : int , snake_case : Any=(5, 5) ): '''simple docstring''' A__ : Optional[int] = np.zeros(snake_case ) A__ : List[Any] = world_size[0] A__ : Dict = world_size[1] def _UpperCamelCase ( self : Any ): '''simple docstring''' print(self.w ) def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : int = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] A__ : int = cell.position[0] A__ : str = cell.position[1] A__ : Any = [] for n in neughbour_cord: A__ : List[Any] = current_x + n[0] A__ : Tuple = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: A__ : List[Any] = Cell() A__ : str = (x, y) A__ : Optional[Any] = cell neighbours.append(snake_case ) return neighbours def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Dict ) ->Dict: A__ : Union[str, Any] = [] A__ : Optional[int] = [] _open.append(UpperCAmelCase__ ) while _open: A__ : List[Any] = np.argmin([n.f for n in _open] ) A__ : Union[str, Any] = _open[min_f] _closed.append(_open.pop(UpperCAmelCase__ ) ) if current == goal: break for n in world.get_neigbours(UpperCAmelCase__ ): for c in _closed: if c == n: continue A__ : Dict = current.g + 1 A__ , A__ : int = n.position A__ , A__ : Optional[int] = goal.position A__ : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 A__ : Optional[int] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(UpperCAmelCase__ ) A__ : List[str] = [] while current.parent is not None: path.append(current.position ) A__ : Union[str, Any] = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": A_ = Gridworld() # Start position and goal A_ = Cell() A_ = (0, 0) A_ = Cell() A_ = (4, 4) print(F'path from {start.position} to {goal.position}') A_ = astar(world, start, goal) # Just for visual reasons. for i in s: A_ = 1 print(world.w)
296
1
"""simple docstring""" from ..utils import DummyObject, requires_backends class __SCREAMING_SNAKE_CASE ( metaclass=UpperCamelCase ): snake_case_ = ['note_seq'] def __init__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : List[Any] ): '''simple docstring''' requires_backends(self , ["""note_seq"""] ) @classmethod def _UpperCamelCase ( cls : str , *snake_case : str , **snake_case : Dict ): '''simple docstring''' requires_backends(cls , ["""note_seq"""] ) @classmethod def _UpperCamelCase ( cls : Dict , *snake_case : List[str] , **snake_case : Any ): '''simple docstring''' requires_backends(cls , ["""note_seq"""] )
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple=False ) ->str: A__ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ : Optional[int] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]=False ) ->str: for i in range(config.num_hidden_layers ): if base_model: A__ : Any = """""" else: A__ : Tuple = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A__ : str = in_proj_bias[: config.hidden_size] A__ : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A__ : Any = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : int = dct.pop(UpperCAmelCase__ ) A__ : Tuple = val def _lowerCAmelCase ( ) ->List[Any]: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Any ) ->Tuple: A__ : List[Any] = DeiTConfig() # all deit models have fine-tuned heads A__ : Tuple = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ : str = 1_0_0_0 A__ : List[str] = """huggingface/label-files""" A__ : Dict = """imagenet-1k-id2label.json""" A__ : List[str] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Dict = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[int] = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} A__ : List[str] = int(deit_name[-6:-4] ) A__ : str = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): A__ : List[str] = 1_9_2 A__ : int = 7_6_8 A__ : List[Any] = 1_2 A__ : Dict = 3 elif deit_name[9:].startswith("""small""" ): A__ : List[Any] = 3_8_4 A__ : List[str] = 1_5_3_6 A__ : Any = 1_2 A__ : Union[str, Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): A__ : int = 1_0_2_4 A__ : str = 4_0_9_6 A__ : Any = 2_4 A__ : int = 1_6 # load original model from timm A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ : Tuple = timm_model.state_dict() A__ : str = create_rename_keys(UpperCAmelCase__, UpperCAmelCase__ ) for src, dest in rename_keys: rename_key(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : str = DeiTForImageClassificationWithTeacher(UpperCAmelCase__ ).eval() model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor A__ : int = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ : Any = DeiTImageProcessor(size=UpperCAmelCase__, crop_size=config.image_size ) A__ : Union[str, Any] = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : Optional[Any] = encoding["""pixel_values"""] A__ : Union[str, Any] = model(UpperCAmelCase__ ) A__ : Union[str, Any] = timm_model(UpperCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCAmelCase__, outputs.logits, atol=1e-3 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
296
1
"""simple docstring""" from __future__ import annotations def _lowerCAmelCase ( UpperCAmelCase__ : str, UpperCAmelCase__ : str ) ->bool: A__ : List[str] = get_failure_array(UpperCAmelCase__ ) # 2) Step through text searching for pattern A__ , A__ : Optional[int] = 0, 0 # index into text, pattern while i < len(UpperCAmelCase__ ): if pattern[j] == text[i]: if j == (len(UpperCAmelCase__ ) - 1): return True j += 1 # if this is a prefix in our pattern # just go back far enough to continue elif j > 0: A__ : List[Any] = failure[j - 1] continue i += 1 return False def _lowerCAmelCase ( UpperCAmelCase__ : str ) ->list[int]: A__ : Any = [0] A__ : Optional[Any] = 0 A__ : Union[str, Any] = 1 while j < len(UpperCAmelCase__ ): if pattern[i] == pattern[j]: i += 1 elif i > 0: A__ : Tuple = failure[i - 1] continue j += 1 failure.append(UpperCAmelCase__ ) return failure if __name__ == "__main__": # Test 1) A_ = '''abc1abc12''' A_ = '''alskfjaldsabc1abc1abc12k23adsfabcabc''' A_ = '''alskfjaldsk23adsfabcabc''' assert kmp(pattern, texta) and not kmp(pattern, texta) # Test 2) A_ = '''ABABX''' A_ = '''ABABZABABYABABX''' assert kmp(pattern, text) # Test 3) A_ = '''AAAB''' A_ = '''ABAAAAAB''' assert kmp(pattern, text) # Test 4) A_ = '''abcdabcy''' A_ = '''abcxabcdabxabcdabcdabcy''' assert kmp(pattern, text) # Test 5) A_ = '''aabaabaaa''' assert get_failure_array(pattern) == [0, 1, 0, 1, 2, 3, 4, 5, 2]
296
"""simple docstring""" from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] A__ : Optional[int] = (low + high) // 2 A__ , A__ , A__ : List[Any] = max_subarray(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_subarray(UpperCAmelCase__, mid + 1, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_cross_sum(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int, int, float]: A__ , A__ : Dict = float("""-inf""" ), -1 A__ , A__ : Optional[Any] = float("""-inf""" ), -1 A__ : int | float = 0 for i in range(UpperCAmelCase__, low - 1, -1 ): summ += arr[i] if summ > left_sum: A__ : Optional[int] = summ A__ : Union[str, Any] = i A__ : Optional[Any] = 0 for i in range(mid + 1, high + 1 ): summ += arr[i] if summ > right_sum: A__ : int = summ A__ : Union[str, Any] = i return max_left, max_right, (left_sum + right_sum) def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->float: A__ : Union[str, Any] = [randint(1, UpperCAmelCase__ ) for _ in range(UpperCAmelCase__ )] A__ : Any = time.time() max_subarray(UpperCAmelCase__, 0, input_size - 1 ) A__ : List[Any] = time.time() return end - start def _lowerCAmelCase ( ) ->None: A__ : List[Any] = [1_0, 1_0_0, 1_0_0_0, 1_0_0_0_0, 5_0_0_0_0, 1_0_0_0_0_0, 2_0_0_0_0_0, 3_0_0_0_0_0, 4_0_0_0_0_0, 5_0_0_0_0_0] A__ : Any = [time_max_subarray(UpperCAmelCase__ ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(UpperCAmelCase__, UpperCAmelCase__ ): print(UpperCAmelCase__, """\t\t""", UpperCAmelCase__ ) plt.plot(UpperCAmelCase__, UpperCAmelCase__ ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
296
1
"""simple docstring""" A_ = 6_5521 def _lowerCAmelCase ( UpperCAmelCase__ : str ) ->int: A__ : List[Any] = 1 A__ : str = 0 for plain_chr in plain_text: A__ : int = (a + ord(UpperCAmelCase__ )) % MOD_ADLER A__ : int = (b + a) % MOD_ADLER return (b << 1_6) | a
296
"""simple docstring""" from __future__ import annotations class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , snake_case : int ): '''simple docstring''' A__ : List[Any] = order # a_{0} ... a_{k} A__ : List[Any] = [1.0] + [0.0] * order # b_{0} ... b_{k} A__ : str = [1.0] + [0.0] * order # x[n-1] ... x[n-k] A__ : Union[str, Any] = [0.0] * self.order # y[n-1] ... y[n-k] A__ : List[str] = [0.0] * self.order def _UpperCamelCase ( self : Optional[int] , snake_case : list[float] , snake_case : list[float] ): '''simple docstring''' if len(snake_case ) < self.order: A__ : Any = [1.0, *a_coeffs] if len(snake_case ) != self.order + 1: A__ : str = ( F'Expected a_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) if len(snake_case ) != self.order + 1: A__ : Union[str, Any] = ( F'Expected b_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) A__ : Dict = a_coeffs A__ : Any = b_coeffs def _UpperCamelCase ( self : List[str] , snake_case : float ): '''simple docstring''' A__ : str = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) A__ : Dict = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] A__ : Tuple = self.input_history[:-1] A__ : int = self.output_history[:-1] A__ : Dict = sample A__ : Tuple = result return result
296
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''xlm-roberta-base''': '''https://huggingface.co/xlm-roberta-base/resolve/main/config.json''', '''xlm-roberta-large''': '''https://huggingface.co/xlm-roberta-large/resolve/main/config.json''', '''xlm-roberta-large-finetuned-conll02-dutch''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-dutch/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll02-spanish''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll02-spanish/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-english''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-english/resolve/main/config.json''' ), '''xlm-roberta-large-finetuned-conll03-german''': ( '''https://huggingface.co/xlm-roberta-large-finetuned-conll03-german/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'xlm-roberta' def __init__( self : Tuple , snake_case : Optional[Any]=3_0522 , snake_case : Optional[Any]=768 , snake_case : Any=12 , snake_case : Optional[int]=12 , snake_case : Any=3072 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : Dict=0.1 , snake_case : Tuple=512 , snake_case : Union[str, Any]=2 , snake_case : Any=0.02 , snake_case : Union[str, Any]=1e-12 , snake_case : Any=1 , snake_case : Optional[int]=0 , snake_case : Optional[int]=2 , snake_case : int="absolute" , snake_case : Dict=True , snake_case : str=None , **snake_case : Dict , ): '''simple docstring''' super().__init__(pad_token_id=snake_case , bos_token_id=snake_case , eos_token_id=snake_case , **snake_case ) A__ : Optional[int] = vocab_size A__ : Optional[Any] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : List[str] = num_attention_heads A__ : Union[str, Any] = hidden_act A__ : Any = intermediate_size A__ : Dict = hidden_dropout_prob A__ : Any = attention_probs_dropout_prob A__ : Tuple = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Optional[int] = initializer_range A__ : Any = layer_norm_eps A__ : List[str] = position_embedding_type A__ : str = use_cache A__ : int = classifier_dropout class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): @property def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' if self.task == "multiple-choice": A__ : int = {0: """batch""", 1: """choice""", 2: """sequence"""} else: A__ : List[str] = {0: """batch""", 1: """sequence"""} return OrderedDict( [ ("""input_ids""", dynamic_axis), ("""attention_mask""", dynamic_axis), ] )
296
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , snake_case : Optional[Any] , snake_case : Tuple=13 , snake_case : Dict=7 , snake_case : Optional[int]=True , snake_case : Union[str, Any]=True , snake_case : Dict=True , snake_case : Any=True , snake_case : List[str]=99 , snake_case : str=64 , snake_case : Optional[int]=5 , snake_case : str=4 , snake_case : List[Any]=37 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : Dict=16 , snake_case : List[Any]=2 , snake_case : Optional[int]=0.02 , snake_case : Any=3 , snake_case : Union[str, Any]=4 , snake_case : Dict=None , ): '''simple docstring''' A__ : Tuple = parent A__ : Union[str, Any] = batch_size A__ : List[str] = seq_length A__ : Optional[int] = is_training A__ : Dict = use_input_mask A__ : Any = use_token_type_ids A__ : Optional[Any] = use_labels A__ : List[str] = vocab_size A__ : Optional[int] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : Any = num_attention_heads A__ : List[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Optional[int] = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[Any] = initializer_range A__ : Optional[int] = num_labels A__ : Dict = num_choices A__ : Dict = scope A__ : List[Any] = vocab_size - 1 def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[Any] = None if self.use_input_mask: A__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_labels: A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Tuple = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ , A__ , A__ : str = self.prepare_config_and_inputs() A__ : Union[str, Any] = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : List[str] , snake_case : int ): '''simple docstring''' A__ : Any = GPTNeoXModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case ) A__ : Optional[int] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : str , snake_case : Any , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = True A__ : str = GPTNeoXModel(snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Dict , snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : int = self.num_labels A__ : int = GPTNeoXForQuestionAnswering(snake_case ) model.to(snake_case ) model.eval() A__ : Optional[Any] = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : str , snake_case : Tuple , snake_case : int , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : List[Any] = self.num_labels A__ : Tuple = GPTNeoXForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Any , snake_case : Union[str, Any] , snake_case : int , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : Any = GPTNeoXForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Optional[int] = True A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass A__ : Tuple = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ : str = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Any = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Tuple = model(snake_case , attention_mask=snake_case , output_hidden_states=snake_case ) A__ : List[Any] = output_from_no_past["""hidden_states"""][0] A__ : List[str] = model( snake_case , attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )["""hidden_states"""][0] # select random slice A__ : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Dict = config_and_inputs A__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) snake_case_ = (GPTNeoXForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = GPTNeoXModelTester(self ) A__ : Any = ConfigTester(self , config_class=snake_case , hidden_size=64 , num_attention_heads=8 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ , A__ , A__ , A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() A__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ , A__ , A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] = ids_tensor([1, 10] , config.vocab_size ) A__ : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Union[str, Any] = GPTNeoXModel(snake_case ) original_model.to(snake_case ) original_model.eval() A__ : Optional[int] = original_model(snake_case ).last_hidden_state A__ : List[str] = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Optional[int] = {"""type""": scaling_type, """factor""": 10.0} A__ : Optional[int] = GPTNeoXModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() A__ : List[str] = scaled_model(snake_case ).last_hidden_state A__ : Tuple = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: A__ : Optional[Any] = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(snake_case ) A__ : Optional[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" A__ : Tuple = model.generate(**snake_case , do_sample=snake_case , max_new_tokens=20 ) A__ : Tuple = tokenizer.batch_decode(snake_case )[0] self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" from .glue import GlueDataset, GlueDataTrainingArguments from .language_modeling import ( LineByLineTextDataset, LineByLineWithRefDataset, LineByLineWithSOPTextDataset, TextDataset, TextDatasetForNextSentencePrediction, ) from .squad import SquadDataset, SquadDataTrainingArguments
296
"""simple docstring""" from collections import defaultdict from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_5_0_0_0_0_0 ) ->int: A__ : defaultdict = defaultdict(UpperCAmelCase__ ) A__ : Any = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, UpperCAmelCase__, 2 ): if gcd(UpperCAmelCase__, UpperCAmelCase__ ) > 1: continue A__ : str = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'{solution() = }')
296
1
"""simple docstring""" import json import os import unittest from transformers import OpenAIGPTTokenizer, OpenAIGPTTokenizerFast from transformers.models.openai.tokenization_openai import VOCAB_FILES_NAMES from transformers.testing_utils import require_ftfy, require_spacy, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = OpenAIGPTTokenizer snake_case_ = OpenAIGPTTokenizerFast snake_case_ = True snake_case_ = False def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' super().setUp() # Adapted from Sennrich et al. 2015 and https://github.com/rsennrich/subword-nmt A__ : Dict = [ """l""", """o""", """w""", """e""", """r""", """s""", """t""", """i""", """d""", """n""", """w</w>""", """r</w>""", """t</w>""", """lo""", """low""", """er</w>""", """low</w>""", """lowest</w>""", """newer</w>""", """wider</w>""", """<unk>""", ] A__ : Tuple = dict(zip(snake_case , range(len(snake_case ) ) ) ) A__ : List[Any] = ["""#version: 0.2""", """l o""", """lo w""", """e r</w>""", """"""] A__ : List[Any] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""vocab_file"""] ) A__ : Tuple = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["""merges_file"""] ) with open(self.vocab_file , """w""" ) as fp: fp.write(json.dumps(snake_case ) ) with open(self.merges_file , """w""" ) as fp: fp.write("""\n""".join(snake_case ) ) def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' return "lower newer", "lower newer" def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = OpenAIGPTTokenizer(self.vocab_file , self.merges_file ) A__ : Union[str, Any] = """lower""" A__ : int = ["""low""", """er</w>"""] A__ : Optional[int] = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , snake_case ) A__ : Any = tokens + ["""<unk>"""] A__ : Tuple = [14, 15, 20] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def _UpperCamelCase ( self : Any , snake_case : int=15 ): '''simple docstring''' for tokenizer, pretrained_name, kwargs in self.tokenizers_list: with self.subTest(F'{tokenizer.__class__.__name__} ({pretrained_name})' ): A__ : List[str] = self.rust_tokenizer_class.from_pretrained(snake_case , **snake_case ) # Simple input A__ : Optional[Any] = """This is a simple input""" A__ : str = ["""This is a simple input 1""", """This is a simple input 2"""] A__ : Dict = ("""This is a simple input""", """This is a pair""") A__ : List[Any] = [ ("""This is a simple input 1""", """This is a simple input 2"""), ("""This is a simple pair 1""", """This is a simple pair 2"""), ] # Simple input tests self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding="""max_length""" ) # Simple input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding="""max_length""" ) # Simple input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding="""max_length""" , ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode , snake_case , max_length=snake_case , padding="""max_length""" ) # Pair input self.assertRaises(snake_case , tokenizer_r.encode_plus , snake_case , max_length=snake_case , padding="""max_length""" ) # Pair input self.assertRaises( snake_case , tokenizer_r.batch_encode_plus , snake_case , max_length=snake_case , padding="""max_length""" , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass @require_ftfy @require_spacy @require_tokenizers class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): pass
296
"""simple docstring""" import os from distutils.util import strtobool def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any] ) ->List[str]: for e in env_keys: A__ : List[Any] = int(os.environ.get(UpperCAmelCase__, -1 ) ) if val >= 0: return val return default def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : str=False ) ->List[str]: A__ : List[Any] = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return strtobool(UpperCAmelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]="no" ) ->int: A__ : str = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return value
296
1
"""simple docstring""" import os import sys A_ = os.path.join(os.path.dirname(__file__), '''src''') sys.path.append(SRC_DIR) from transformers import ( AutoConfig, AutoModel, AutoModelForCausalLM, AutoModelForMaskedLM, AutoModelForQuestionAnswering, AutoModelForSequenceClassification, AutoTokenizer, add_start_docstrings, ) A_ = [ '''torch''', '''numpy''', '''tokenizers''', '''filelock''', '''requests''', '''tqdm''', '''regex''', '''sentencepiece''', '''sacremoses''', '''importlib_metadata''', '''huggingface_hub''', ] @add_start_docstrings(AutoConfig.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : Optional[Any], **UpperCAmelCase__ : Optional[int] ) ->Dict: return AutoConfig.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoTokenizer.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : Optional[Any], **UpperCAmelCase__ : List[str] ) ->Tuple: return AutoTokenizer.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoModel.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : Any, **UpperCAmelCase__ : int ) ->List[Any]: return AutoModel.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoModelForCausalLM.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : List[Any], **UpperCAmelCase__ : Union[str, Any] ) ->Tuple: return AutoModelForCausalLM.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoModelForMaskedLM.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : Dict, **UpperCAmelCase__ : Optional[Any] ) ->Any: return AutoModelForMaskedLM.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoModelForSequenceClassification.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : str, **UpperCAmelCase__ : Any ) ->int: return AutoModelForSequenceClassification.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ ) @add_start_docstrings(AutoModelForQuestionAnswering.__doc__ ) def _lowerCAmelCase ( *UpperCAmelCase__ : Optional[int], **UpperCAmelCase__ : int ) ->int: return AutoModelForQuestionAnswering.from_pretrained(*UpperCAmelCase__, **UpperCAmelCase__ )
296
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
1
"""simple docstring""" import pytest import requests from datasets.utils.file_utils import http_head from .utils import OfflineSimulationMode, RequestWouldHangIndefinitelyError, offline @pytest.mark.integration def _lowerCAmelCase ( ) ->Dict: with offline(OfflineSimulationMode.CONNECTION_TIMES_OUT ): with pytest.raises(UpperCAmelCase__ ): requests.request("""GET""", """https://huggingface.co""" ) with pytest.raises(requests.exceptions.ConnectTimeout ): requests.request("""GET""", """https://huggingface.co""", timeout=1.0 ) @pytest.mark.integration def _lowerCAmelCase ( ) ->Any: with offline(OfflineSimulationMode.CONNECTION_FAILS ): with pytest.raises(requests.exceptions.ConnectionError ): requests.request("""GET""", """https://huggingface.co""" ) def _lowerCAmelCase ( ) ->List[str]: with offline(OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1 ): with pytest.raises(UpperCAmelCase__ ): http_head("""https://huggingface.co""" )
296
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = logging.get_logger(__name__) A_ = Dict[str, Any] A_ = List[Prediction] @add_end_docstrings(UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : str , *snake_case : Tuple , **snake_case : Tuple ): '''simple docstring''' super().__init__(*snake_case , **snake_case ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _UpperCamelCase ( self : List[Any] , **snake_case : Optional[int] ): '''simple docstring''' A__ : Dict = {} if "threshold" in kwargs: A__ : int = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return super().__call__(*snake_case , **snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : List[str] = load_image(snake_case ) A__ : int = torch.IntTensor([[image.height, image.width]] ) A__ : Union[str, Any] = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: A__ : str = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) A__ : List[str] = target_size return inputs def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : str = model_inputs.pop("""target_size""" ) A__ : Dict = self.model(**snake_case ) A__ : Optional[Any] = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: A__ : str = model_inputs["""bbox"""] return model_outputs def _UpperCamelCase ( self : Tuple , snake_case : Optional[int] , snake_case : int=0.9 ): '''simple docstring''' A__ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A__ , A__ : Tuple = target_size[0].tolist() def unnormalize(snake_case : Optional[int] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) A__ , A__ : Optional[int] = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A__ : Optional[Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A__ : List[str] = [unnormalize(snake_case ) for bbox in model_outputs["""bbox"""].squeeze(0 )] A__ : Tuple = ["""score""", """label""", """box"""] A__ : Any = [dict(zip(snake_case , snake_case ) ) for vals in zip(scores.tolist() , snake_case , snake_case ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A__ : Union[str, Any] = self.image_processor.post_process_object_detection(snake_case , snake_case , snake_case ) A__ : str = raw_annotations[0] A__ : str = raw_annotation["""scores"""] A__ : List[Any] = raw_annotation["""labels"""] A__ : int = raw_annotation["""boxes"""] A__ : str = scores.tolist() A__ : Any = [self.model.config.idalabel[label.item()] for label in labels] A__ : int = [self._get_bounding_box(snake_case ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A__ : str = ["""score""", """label""", """box"""] A__ : Dict = [ dict(zip(snake_case , snake_case ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def _UpperCamelCase ( self : Union[str, Any] , snake_case : "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) A__ , A__ , A__ , A__ : Any = box.int().tolist() A__ : Any = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
296
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''microsoft/cvt-13''': '''https://huggingface.co/microsoft/cvt-13/resolve/main/config.json''', # See all Cvt models at https://huggingface.co/models?filter=cvt } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'cvt' def __init__( self : str , snake_case : Optional[Any]=3 , snake_case : Tuple=[7, 3, 3] , snake_case : Optional[int]=[4, 2, 2] , snake_case : Dict=[2, 1, 1] , snake_case : Any=[64, 192, 384] , snake_case : Any=[1, 3, 6] , snake_case : Dict=[1, 2, 10] , snake_case : int=[4.0, 4.0, 4.0] , snake_case : int=[0.0, 0.0, 0.0] , snake_case : List[Any]=[0.0, 0.0, 0.0] , snake_case : Union[str, Any]=[0.0, 0.0, 0.1] , snake_case : Dict=[True, True, True] , snake_case : List[Any]=[False, False, True] , snake_case : str=["dw_bn", "dw_bn", "dw_bn"] , snake_case : Union[str, Any]=[3, 3, 3] , snake_case : Union[str, Any]=[1, 1, 1] , snake_case : Union[str, Any]=[2, 2, 2] , snake_case : Optional[int]=[1, 1, 1] , snake_case : Union[str, Any]=[1, 1, 1] , snake_case : List[str]=0.02 , snake_case : int=1e-12 , **snake_case : str , ): '''simple docstring''' super().__init__(**snake_case ) A__ : int = num_channels A__ : Dict = patch_sizes A__ : Any = patch_stride A__ : Union[str, Any] = patch_padding A__ : Any = embed_dim A__ : str = num_heads A__ : Optional[int] = depth A__ : int = mlp_ratio A__ : Dict = attention_drop_rate A__ : Optional[Any] = drop_rate A__ : Any = drop_path_rate A__ : int = qkv_bias A__ : Dict = cls_token A__ : Any = qkv_projection_method A__ : Tuple = kernel_qkv A__ : List[Any] = padding_kv A__ : Any = stride_kv A__ : Optional[int] = padding_q A__ : List[str] = stride_q A__ : Any = initializer_range A__ : Any = layer_norm_eps
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , snake_case : Optional[Any] , snake_case : Tuple=13 , snake_case : Dict=7 , snake_case : Optional[int]=True , snake_case : Union[str, Any]=True , snake_case : Dict=True , snake_case : Any=True , snake_case : List[str]=99 , snake_case : str=64 , snake_case : Optional[int]=5 , snake_case : str=4 , snake_case : List[Any]=37 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : Dict=16 , snake_case : List[Any]=2 , snake_case : Optional[int]=0.02 , snake_case : Any=3 , snake_case : Union[str, Any]=4 , snake_case : Dict=None , ): '''simple docstring''' A__ : Tuple = parent A__ : Union[str, Any] = batch_size A__ : List[str] = seq_length A__ : Optional[int] = is_training A__ : Dict = use_input_mask A__ : Any = use_token_type_ids A__ : Optional[Any] = use_labels A__ : List[str] = vocab_size A__ : Optional[int] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : Any = num_attention_heads A__ : List[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Optional[int] = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[Any] = initializer_range A__ : Optional[int] = num_labels A__ : Dict = num_choices A__ : Dict = scope A__ : List[Any] = vocab_size - 1 def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[Any] = None if self.use_input_mask: A__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_labels: A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Tuple = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ , A__ , A__ : str = self.prepare_config_and_inputs() A__ : Union[str, Any] = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : List[str] , snake_case : int ): '''simple docstring''' A__ : Any = GPTNeoXModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case ) A__ : Optional[int] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : str , snake_case : Any , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = True A__ : str = GPTNeoXModel(snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Dict , snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : int = self.num_labels A__ : int = GPTNeoXForQuestionAnswering(snake_case ) model.to(snake_case ) model.eval() A__ : Optional[Any] = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : str , snake_case : Tuple , snake_case : int , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : List[Any] = self.num_labels A__ : Tuple = GPTNeoXForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Any , snake_case : Union[str, Any] , snake_case : int , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : Any = GPTNeoXForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Optional[int] = True A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass A__ : Tuple = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ : str = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Any = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Tuple = model(snake_case , attention_mask=snake_case , output_hidden_states=snake_case ) A__ : List[Any] = output_from_no_past["""hidden_states"""][0] A__ : List[str] = model( snake_case , attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )["""hidden_states"""][0] # select random slice A__ : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Dict = config_and_inputs A__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) snake_case_ = (GPTNeoXForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = GPTNeoXModelTester(self ) A__ : Any = ConfigTester(self , config_class=snake_case , hidden_size=64 , num_attention_heads=8 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ , A__ , A__ , A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() A__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ , A__ , A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] = ids_tensor([1, 10] , config.vocab_size ) A__ : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Union[str, Any] = GPTNeoXModel(snake_case ) original_model.to(snake_case ) original_model.eval() A__ : Optional[int] = original_model(snake_case ).last_hidden_state A__ : List[str] = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Optional[int] = {"""type""": scaling_type, """factor""": 10.0} A__ : Optional[int] = GPTNeoXModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() A__ : List[str] = scaled_model(snake_case ).last_hidden_state A__ : Tuple = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: A__ : Optional[Any] = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(snake_case ) A__ : Optional[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" A__ : Tuple = model.generate(**snake_case , do_sample=snake_case , max_new_tokens=20 ) A__ : Tuple = tokenizer.batch_decode(snake_case )[0] self.assertEqual(snake_case , snake_case )
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
"""simple docstring""" import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset A_ = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : str , snake_case : str ): '''simple docstring''' super().__init__() A__ : Optional[Any] = torchvision.models.resnetaaa(pretrained=snake_case ) A__ : int = list(model.children() )[:-2] A__ : Optional[Any] = nn.Sequential(*snake_case ) A__ : Tuple = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def _UpperCamelCase ( self : List[Any] , snake_case : str ): '''simple docstring''' A__ : List[str] = self.pool(self.model(snake_case ) ) A__ : str = torch.flatten(snake_case , start_dim=2 ) A__ : Union[str, Any] = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Union[str, Any] , snake_case : str , snake_case : Optional[Any] , snake_case : List[Any] , snake_case : str , snake_case : int ): '''simple docstring''' A__ : Optional[Any] = [json.loads(snake_case ) for l in open(snake_case )] A__ : List[Any] = os.path.dirname(snake_case ) A__ : Any = tokenizer A__ : int = labels A__ : Optional[Any] = len(snake_case ) A__ : Dict = max_seq_length A__ : Union[str, Any] = transforms def __len__( self : Optional[Any] ): '''simple docstring''' return len(self.data ) def __getitem__( self : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : List[str] = torch.LongTensor(self.tokenizer.encode(self.data[index]["""text"""] , add_special_tokens=snake_case ) ) A__ , A__ , A__ : Union[str, Any] = sentence[0], sentence[1:-1], sentence[-1] A__ : Tuple = sentence[: self.max_seq_length] A__ : str = torch.zeros(self.n_classes ) A__ : List[str] = 1 A__ : Dict = Image.open(os.path.join(self.data_dir , self.data[index]["""img"""] ) ).convert("""RGB""" ) A__ : Optional[Any] = self.transforms(snake_case ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : int = Counter() for row in self.data: label_freqs.update(row["""label"""] ) return label_freqs def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Optional[int]: A__ : Union[str, Any] = [len(row["""sentence"""] ) for row in batch] A__ , A__ : Union[str, Any] = len(UpperCAmelCase__ ), max(UpperCAmelCase__ ) A__ : Union[str, Any] = torch.zeros(UpperCAmelCase__, UpperCAmelCase__, dtype=torch.long ) A__ : Union[str, Any] = torch.zeros(UpperCAmelCase__, UpperCAmelCase__, dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(UpperCAmelCase__, UpperCAmelCase__ ) ): A__ : Union[str, Any] = input_row["""sentence"""] A__ : Union[str, Any] = 1 A__ : Dict = torch.stack([row["""image"""] for row in batch] ) A__ : Any = torch.stack([row["""label"""] for row in batch] ) A__ : List[str] = torch.stack([row["""image_start_token"""] for row in batch] ) A__ : List[Any] = torch.stack([row["""image_end_token"""] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def _lowerCAmelCase ( ) ->Union[str, Any]: return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def _lowerCAmelCase ( ) ->Optional[int]: return transforms.Compose( [ transforms.Resize(2_5_6 ), transforms.CenterCrop(2_2_4 ), transforms.ToTensor(), transforms.Normalize( mean=[0.4677_7044, 0.4453_1429, 0.4066_1017], std=[0.1222_1994, 0.1214_5835, 0.1438_0469], ), ] )
296
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
1
"""simple docstring""" import unittest from transformers import RoFormerTokenizer, RoFormerTokenizerFast from transformers.testing_utils import require_rjieba, require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_rjieba @require_tokenizers class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = RoFormerTokenizer snake_case_ = RoFormerTokenizerFast snake_case_ = True snake_case_ = True def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' super().setUp() def _UpperCamelCase ( self : Optional[Any] , **snake_case : Optional[int] ): '''simple docstring''' return self.tokenizer_class.from_pretrained("""junnyu/roformer_chinese_base""" , **snake_case ) def _UpperCamelCase ( self : str , **snake_case : int ): '''simple docstring''' return self.rust_tokenizer_class.from_pretrained("""junnyu/roformer_chinese_base""" , **snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Dict = """永和服装饰品有限公司,今天天气非常好""" A__ : List[str] = """永和 服装 饰品 有限公司 , 今 天 天 气 非常 好""" return input_text, output_text def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Dict = self.get_tokenizer() A__ , A__ : Any = self.get_chinese_input_output_texts() A__ : Dict = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , output_text.split() ) A__ : Union[str, Any] = tokens + [tokenizer.unk_token] A__ : List[str] = [2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = self.get_rust_tokenizer() A__ , A__ : Tuple = self.get_chinese_input_output_texts() A__ : Tuple = tokenizer.tokenize(snake_case ) self.assertListEqual(snake_case , output_text.split() ) A__ : Tuple = tokens + [tokenizer.unk_token] A__ : int = [2_2943, 2_1332, 3_4431, 4_5904, 117, 306, 1231, 1231, 2653, 3_3994, 1266, 100] self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case ) , snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' pass def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass def _UpperCamelCase ( self : List[str] ): '''simple docstring''' pass
296
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[Any]=13 , snake_case : Union[str, Any]=7 , snake_case : Optional[Any]=True , snake_case : str=True , snake_case : Dict=False , snake_case : Union[str, Any]=True , snake_case : Optional[Any]=99 , snake_case : str=32 , snake_case : Tuple=5 , snake_case : List[str]=4 , snake_case : Optional[int]=37 , snake_case : str="gelu" , snake_case : Tuple=0.1 , snake_case : Optional[int]=0.1 , snake_case : int=512 , snake_case : List[str]=16 , snake_case : str=2 , snake_case : Optional[int]=0.02 , snake_case : str=3 , snake_case : Dict=4 , snake_case : Optional[Any]=None , ): '''simple docstring''' A__ : int = parent A__ : Union[str, Any] = batch_size A__ : Optional[int] = seq_length A__ : List[Any] = is_training A__ : List[str] = use_input_mask A__ : Optional[Any] = use_token_type_ids A__ : List[Any] = use_labels A__ : Union[str, Any] = vocab_size A__ : List[Any] = hidden_size A__ : Any = num_hidden_layers A__ : Any = num_attention_heads A__ : Optional[int] = intermediate_size A__ : Any = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Dict = attention_probs_dropout_prob A__ : Optional[int] = max_position_embeddings A__ : Tuple = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[str] = initializer_range A__ : Any = num_labels A__ : Any = num_choices A__ : int = scope def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = None if self.use_input_mask: A__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_token_type_ids: A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : int = None A__ : int = None A__ : List[str] = None if self.use_labels: A__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) A__ : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Any , snake_case : Dict , snake_case : Any , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case ) A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Optional[int] , snake_case : List[str] , snake_case : str , snake_case : Optional[Any] , snake_case : List[str] , snake_case : List[Any] , snake_case : Tuple , snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = BioGptForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Any , snake_case : str , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Any , *snake_case : Dict ): '''simple docstring''' A__ : Union[str, Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() # create attention mask A__ : List[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) A__ : Any = self.seq_length // 2 A__ : str = 0 # first forward pass A__ , A__ : List[Any] = model(snake_case , attention_mask=snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids A__ : int = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A__ : List[str] = ids_tensor((1,) , snake_case ).item() + 1 A__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A__ : int = random_other_next_tokens # append to next input_ids and attn_mask A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case )] , dim=1 , ) # get two different outputs A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Optional[int] = model(snake_case , past_key_values=snake_case , attention_mask=snake_case )["""last_hidden_state"""] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() A__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str , snake_case : int , snake_case : Optional[Any] , *snake_case : str ): '''simple docstring''' A__ : Dict = BioGptModel(config=snake_case ).to(snake_case ).eval() A__ : Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) # first forward pass A__ : Dict = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ , A__ : List[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Optional[int] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A__ : Any = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , past_key_values=snake_case )[ """last_hidden_state""" ] # select random slice A__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Tuple , *snake_case : Union[str, Any] , snake_case : Union[str, Any]=False ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM(snake_case ) model.to(snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() A__ : Optional[Any] = model(snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self : int , snake_case : Optional[Any] , *snake_case : Optional[int] ): '''simple docstring''' A__ : int = BioGptModel(snake_case ) A__ : Union[str, Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _UpperCamelCase ( self : Any , snake_case : Dict , snake_case : Tuple , snake_case : int , snake_case : Union[str, Any] , snake_case : Dict , *snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = self.num_labels A__ : int = BioGptForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : str = config_and_inputs A__ : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case_ = (BioGptForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[str] = BioGptModelTester(self ) A__ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : str = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case , gradient_checkpointing=snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) A__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = """left""" # Define PAD Token = EOS Token = 50256 A__ : Optional[int] = tokenizer.eos_token A__ : Dict = model.config.eos_token_id # use different length sentences to test batching A__ : Union[str, Any] = [ """Hello, my dog is a little""", """Today, I""", ] A__ : List[str] = tokenizer(snake_case , return_tensors="""pt""" , padding=snake_case ) A__ : str = inputs["""input_ids"""].to(snake_case ) A__ : Dict = model.generate( input_ids=snake_case , attention_mask=inputs["""attention_mask"""].to(snake_case ) , ) A__ : Optional[int] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Any = model.generate(input_ids=snake_case ) A__ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A__ : str = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Dict = model.generate(input_ids=snake_case , max_length=model.config.max_length - num_paddings ) A__ : Optional[Any] = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case ) A__ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case ) A__ : Optional[int] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(snake_case , [non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = BioGptModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(snake_case ) A__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Union[str, Any] = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Any = 3 A__ : List[Any] = """multi_label_classification""" A__ : Dict = input_dict["""input_ids"""] A__ : Tuple = input_ids.ne(1 ).to(snake_case ) A__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Tuple = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A__ : str = torch.tensor([[2, 4805, 9, 656, 21]] ) A__ : Dict = model(snake_case )[0] A__ : Tuple = 4_2384 A__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : str = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) torch.manual_seed(0 ) A__ : Tuple = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(snake_case ) A__ : Optional[int] = model.generate( **snake_case , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case , ) A__ : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case ) A__ : List[str] = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" import unittest from transformers import DonutProcessor A_ = '''naver-clova-ix/donut-base''' class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = DonutProcessor.from_pretrained(snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : int = { """name""": """John Doe""", """age""": """99""", """city""": """Atlanta""", """state""": """GA""", """zip""": """30301""", """phone""": """123-4567""", """nicknames""": [{"""nickname""": """Johnny"""}, {"""nickname""": """JD"""}], } A__ : Optional[Any] = ( """<s_name>John Doe</s_name><s_age>99</s_age><s_city>Atlanta</s_city>""" """<s_state>GA</s_state><s_zip>30301</s_zip><s_phone>123-4567</s_phone>""" """<s_nicknames><s_nickname>Johnny</s_nickname>""" """<sep/><s_nickname>JD</s_nickname></s_nicknames>""" ) A__ : int = self.processor.tokenajson(snake_case ) self.assertDictEqual(snake_case , snake_case )
296
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
1
"""simple docstring""" import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast A_ = datasets.utils.logging.get_logger(__name__) @dataclass class __SCREAMING_SNAKE_CASE ( datasets.BuilderConfig ): snake_case_ = 10000 snake_case_ = None snake_case_ = None class __SCREAMING_SNAKE_CASE ( datasets.ArrowBasedBuilder ): snake_case_ = ParquetConfig def _UpperCamelCase ( self : int ): '''simple docstring''' return datasets.DatasetInfo(features=self.config.features ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}' ) A__ : Tuple = dl_manager.download_and_extract(self.config.data_files ) if isinstance(snake_case , (str, list, tuple) ): A__ : Dict = data_files if isinstance(snake_case , snake_case ): A__ : List[Any] = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A__ : Union[str, Any] = [dl_manager.iter_files(snake_case ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"""files""": files} )] A__ : Optional[int] = [] for split_name, files in data_files.items(): if isinstance(snake_case , snake_case ): A__ : Dict = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive A__ : Dict = [dl_manager.iter_files(snake_case ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(snake_case ): with open(snake_case , """rb""" ) as f: A__ : str = datasets.Features.from_arrow_schema(pq.read_schema(snake_case ) ) break splits.append(datasets.SplitGenerator(name=snake_case , gen_kwargs={"""files""": files} ) ) return splits def _UpperCamelCase ( self : str , snake_case : pa.Table ): '''simple docstring''' if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example A__ : int = table_cast(snake_case , self.info.features.arrow_schema ) return pa_table def _UpperCamelCase ( self : List[str] , snake_case : List[Any] ): '''simple docstring''' A__ : Dict = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( F'Tried to load parquet data with columns \'{self.config.columns}\' with mismatching features \'{self.info.features}\'' ) for file_idx, file in enumerate(itertools.chain.from_iterable(snake_case ) ): with open(snake_case , """rb""" ) as f: A__ : int = pq.ParquetFile(snake_case ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): A__ : List[Any] = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield F'{file_idx}_{batch_idx}', self._cast_table(snake_case ) except ValueError as e: logger.error(F'Failed to read file \'{file}\' with error {type(snake_case )}: {e}' ) raise
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_torch_available, is_vision_available, ) A_ = {'''configuration_vit''': ['''VIT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''ViTConfig''', '''ViTOnnxConfig''']} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''ViTFeatureExtractor'''] A_ = ['''ViTImageProcessor'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''VIT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''ViTForImageClassification''', '''ViTForMaskedImageModeling''', '''ViTModel''', '''ViTPreTrainedModel''', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''TFViTForImageClassification''', '''TFViTModel''', '''TFViTPreTrainedModel''', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''FlaxViTForImageClassification''', '''FlaxViTModel''', '''FlaxViTPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_vit import VIT_PRETRAINED_CONFIG_ARCHIVE_MAP, ViTConfig, ViTOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_vit import ViTFeatureExtractor from .image_processing_vit import ViTImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_vit import ( VIT_PRETRAINED_MODEL_ARCHIVE_LIST, ViTForImageClassification, ViTForMaskedImageModeling, ViTModel, ViTPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_vit import TFViTForImageClassification, TFViTModel, TFViTPreTrainedModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_vit import FlaxViTForImageClassification, FlaxViTModel, FlaxViTPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A_ = '''src/diffusers''' A_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. A_ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) A_ = spec.loader.load_module() def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any] ) ->Any: return line.startswith(UpperCAmelCase__ ) or len(UpperCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", UpperCAmelCase__ ) is not None def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Union[str, Any]: A__ : Any = object_name.split(""".""" ) A__ : int = 0 # First let's find the module where our object lives. A__ : str = parts[i] while i < len(UpperCAmelCase__ ) and not os.path.isfile(os.path.join(UpperCAmelCase__, f'{module}.py' ) ): i += 1 if i < len(UpperCAmelCase__ ): A__ : Union[str, Any] = os.path.join(UpperCAmelCase__, parts[i] ) if i >= len(UpperCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(UpperCAmelCase__, f'{module}.py' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : List[Any] = f.readlines() # Now let's find the class / func in the code! A__ : Optional[Any] = """""" A__ : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A__ : List[Any] = line_index while line_index < len(UpperCAmelCase__ ) and _should_continue(lines[line_index], UpperCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : List[Any] = lines[start_index:line_index] return "".join(UpperCAmelCase__ ) A_ = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') A_ = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') A_ = re.compile(r'''<FILL\s+[^>]*>''') def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Optional[Any]: A__ : Dict = code.split("""\n""" ) A__ : List[Any] = 0 while idx < len(UpperCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase__ ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->int: A__ : str = len(get_indent(UpperCAmelCase__ ) ) > 0 if has_indent: A__ : Union[str, Any] = f'class Bla:\n{code}' A__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=1_1_9, preview=UpperCAmelCase__ ) A__ : Tuple = black.format_str(UpperCAmelCase__, mode=UpperCAmelCase__ ) A__ , A__ : List[Any] = style_docstrings_in_code(UpperCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : Dict=False ) ->List[Any]: with open(UpperCAmelCase__, """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : int = f.readlines() A__ : Dict = [] A__ : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase__ ): A__ : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A__ , A__ , A__ : Dict = search.groups() A__ : Tuple = find_code_in_diffusers(UpperCAmelCase__ ) A__ : int = get_indent(UpperCAmelCase__ ) A__ : List[str] = line_index + 1 if indent == theoretical_indent else line_index + 2 A__ : Tuple = theoretical_indent A__ : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A__ : Tuple = True while line_index < len(UpperCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase__ ): break A__ : Optional[int] = lines[line_index] A__ : Tuple = _should_continue(UpperCAmelCase__, UpperCAmelCase__ ) and re.search(f'^{indent}# End copy', UpperCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : Dict = lines[start_index:line_index] A__ : Tuple = """""".join(UpperCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies A__ : Optional[int] = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase__ ) is None] A__ : Optional[Any] = """\n""".join(UpperCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase__ ) > 0: A__ : int = replace_pattern.replace("""with""", """""" ).split(""",""" ) A__ : List[Any] = [_re_replace_pattern.search(UpperCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue A__ , A__ , A__ : Union[str, Any] = pattern.groups() A__ : Union[str, Any] = re.sub(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if option.strip() == "all-casing": A__ : List[Any] = re.sub(obja.lower(), obja.lower(), UpperCAmelCase__ ) A__ : Tuple = re.sub(obja.upper(), obja.upper(), UpperCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A__ : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) A__ : List[Any] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A__ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] A__ : Tuple = start_index + 1 if overwrite and len(UpperCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(UpperCAmelCase__, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(UpperCAmelCase__ ) return diffs def _lowerCAmelCase ( UpperCAmelCase__ : bool = False ) ->Any: A__ : Dict = glob.glob(os.path.join(UpperCAmelCase__, """**/*.py""" ), recursive=UpperCAmelCase__ ) A__ : str = [] for filename in all_files: A__ : Any = is_copy_consistent(UpperCAmelCase__, UpperCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(UpperCAmelCase__ ) > 0: A__ : Any = """\n""".join(UpperCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() check_copies(args.fix_and_overwrite)
296
1
"""simple docstring""" import argparse import random import joblib import numpy as np import torch from igf.igf import ( SecondaryLearner, collect_objective_set, compute_perplexity, generate_datasets, load_gpta, recopy_gpta, set_seed, train_secondary_learner, ) from torch.utils.data import DataLoader, RandomSampler from transformers import GPTaLMHeadModel def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int]=3_2, UpperCAmelCase__ : str=1_0, UpperCAmelCase__ : str=1_0_0, UpperCAmelCase__ : Optional[int]=1_0_2_6, UpperCAmelCase__ : List[Any]=True, UpperCAmelCase__ : Optional[int]="data/tokenized_stories_train_wikitext103.jbl", UpperCAmelCase__ : str="igf_context_pairs.jbl", ) ->Tuple: set_seed(3 ) # generate train_data and objective_set A__ , A__ : Any = generate_datasets( UpperCAmelCase__, UpperCAmelCase__, number=UpperCAmelCase__, min_len=1_0_2_6, trim=UpperCAmelCase__ ) # keeps model same across runs set_seed(4 ) # model, lm_optimizer, lm_scheduler = recopy_gpt2(model, device, max_steps) # store original model weights # can we train on GPU? A__ : str = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) # load pretrained model A__ : Any = load_gpta("""gpt2""" ).to(UpperCAmelCase__ ) print("""computing perplexity on objective set""" ) A__ : str = compute_perplexity(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ).item() print("""perplexity on objective set:""", UpperCAmelCase__ ) # collect igf pairs and save to file demo.jbl collect_objective_set(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # clean up, delete model and data we don't need anymore del model, train_data, objective_set torch.cuda.empty_cache() def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : int=1_5, UpperCAmelCase__ : List[Any]=1_2_8, UpperCAmelCase__ : Tuple=1_0_0, UpperCAmelCase__ : Union[str, Any]="igf_model.pt", ) ->Optional[int]: set_seed(4_2 ) # Load pre-trained model A__ : int = GPTaLMHeadModel.from_pretrained("""gpt2""" ) # Initialize secondary learner to use embedding weights of model A__ : Union[str, Any] = SecondaryLearner(UpperCAmelCase__ ) # Train secondary learner A__ : Dict = train_secondary_learner( UpperCAmelCase__, UpperCAmelCase__, max_epochs=UpperCAmelCase__, batch_size=UpperCAmelCase__, eval_freq=1_0_0, igf_model_path=UpperCAmelCase__, ) del model, secondary_learner_train_data torch.cuda.empty_cache() return secondary_learner def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Dict, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any]=3_2, UpperCAmelCase__ : str=1_0_0_0, UpperCAmelCase__ : Optional[Any]=1_6, UpperCAmelCase__ : List[Any]=1.0, UpperCAmelCase__ : str=recopy_gpta, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : Optional[int]=1_0, UpperCAmelCase__ : Dict="gpt2_finetuned.pt", ) ->Tuple: A__ : int = torch.device("""cuda:0""" if torch.cuda.is_available() else """cpu""" ) A__ : Optional[Any] = RandomSampler(UpperCAmelCase__ ) A__ : Tuple = DataLoader(UpperCAmelCase__, sampler=UpperCAmelCase__ ) A__ : List[str] = max_steps // (len(UpperCAmelCase__ )) + 1 A__ : List[Any] = 0 A__ : Optional[Any] = torch.zeros((1, context_len), dtype=torch.long, device=UpperCAmelCase__ ) A__ , A__ , A__ : Any = recopy_model(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) model.train() if secondary_learner is not None: secondary_learner.to(UpperCAmelCase__ ) secondary_learner.eval() A__ : Union[str, Any] = [] A__ : Optional[int] = 0 A__ : Union[str, Any] = [] A__ : Union[str, Any] = [] # Compute the performance of the transformer model at the beginning A__ : Any = compute_perplexity(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) test_perps.append(UpperCAmelCase__ ) print("""Test perplexity, step""", UpperCAmelCase__, """:""", UpperCAmelCase__ ) for epoch in range(int(UpperCAmelCase__ ) ): for step, example in enumerate(UpperCAmelCase__ ): torch.cuda.empty_cache() A__ : Optional[int] = random.randint(0, example.size(2 ) - context_len - 1 ) A__ : Optional[int] = example[0, 0, start : start + context_len] lm_optimizer.zero_grad() A__ : int = model(UpperCAmelCase__, labels=UpperCAmelCase__ ) A__ : Dict = True if secondary_learner is not None: A__ : Union[str, Any] = secondary_learner.forward( torch.tensor(UpperCAmelCase__, dtype=torch.long, device=UpperCAmelCase__ ).unsqueeze(0 ) )[0].item() observed_qs.append(float(UpperCAmelCase__ ) ) # Here we implement the simple non-constant threshold for the predicted IG(X) value # We will decay the selectivity of our secondary learner filter from # 1 standard deviation above average to 1 below average after 10 batches. if global_step == 1_0: A__ : Optional[int] = -1 if predicted_q < threshold: A__ : Optional[int] = False # If we passed the filter, add the context to the batch! if do_backprop: contexts.append(np.array(context.cpu() ) ) A__ : List[Any] = outputs[0] lm_loss.backward() examples += 1 del outputs # Once the batch is filled with enough contexts, backprop on the batch. if examples == batch_size: torch.cuda.empty_cache() A__ : int = 0 # Do LM backprop torch.nn.utils.clip_grad_norm_(model.parameters(), 3.0 ) lm_optimizer.step() lm_scheduler.step() # Update learning rate schedule global_step += 1 # Compute the performance of the transformer model at this batch if global_step % eval_interval == 0: A__ : Optional[int] = compute_perplexity(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) test_perps.append(UpperCAmelCase__ ) print("""Test perplexity, step""", UpperCAmelCase__, """:""", UpperCAmelCase__ ) # Break out of the loop after 60 batches if max_steps > 0 and global_step > 6_0: break if max_steps > 0 and global_step > 6_0: break # save finetuned transformer model torch.save(model.state_dict(), UpperCAmelCase__ ) torch.cuda.empty_cache() # Do some cleaning up so we can reinitialize for the next run of this function del lm_optimizer del lm_scheduler return model def _lowerCAmelCase ( ) ->Optional[Any]: A__ : Dict = argparse.ArgumentParser(description="""Fine-tune a transformer model with IGF on a language modeling task""" ) # Required parameters parser.add_argument( """--data_dir""", default=UpperCAmelCase__, type=UpperCAmelCase__, required=UpperCAmelCase__, help="""The input data dir. Should contain data files for WikiText.""", ) parser.add_argument( """--model_name_or_path""", default=UpperCAmelCase__, type=UpperCAmelCase__, required=UpperCAmelCase__, help="""Path to pretrained model or model identifier from huggingface.co/models""", ) parser.add_argument( """--data_file""", type=UpperCAmelCase__, default=UpperCAmelCase__, help=( """A jbl file containing tokenized data which can be split as objective dataset, """ """train_dataset and test_dataset.""" ), ) parser.add_argument( """--igf_data_file""", type=UpperCAmelCase__, default=UpperCAmelCase__, help="""A jbl file containing the context and information gain pairs to train secondary learner.""", ) parser.add_argument( """--output_dir""", default=UpperCAmelCase__, type=UpperCAmelCase__, required=UpperCAmelCase__, help="""The output directory where the final fine-tuned model is stored.""", ) parser.add_argument( """--tokenizer_name""", default=UpperCAmelCase__, type=UpperCAmelCase__, help="""Pretrained tokenizer name or path if not the same as model_name""", ) parser.add_argument("""--seed""", type=UpperCAmelCase__, default=UpperCAmelCase__, help="""A seed for reproducible training.""" ) parser.add_argument( """--context_len""", default=3_2, type=UpperCAmelCase__, help=( """The maximum total input sequence length after tokenization. Sequences longer """ """than this will be truncated, sequences shorter will be padded.""" ), ) parser.add_argument( """--size_objective_set""", default=1_0_0, type=UpperCAmelCase__, help="""number of articles that are long enough to be used as our objective set""", ) parser.add_argument( """--eval_freq""", default=1_0_0, type=UpperCAmelCase__, help="""secondary model evaluation is triggered at eval_freq""" ) parser.add_argument("""--max_steps""", default=1_0_0_0, type=UpperCAmelCase__, help="""To calculate training epochs""" ) parser.add_argument( """--secondary_learner_batch_size""", default=1_2_8, type=UpperCAmelCase__, help="""batch size of training data for secondary learner""", ) parser.add_argument( """--batch_size""", default=1_6, type=UpperCAmelCase__, help="""batch size of training data of language model(gpt2) """ ) parser.add_argument( """--eval_interval""", default=1_0, type=UpperCAmelCase__, help=( """decay the selectivity of our secondary learner filter from""" """1 standard deviation above average to 1 below average after 10 batches""" ), ) parser.add_argument( """--number""", default=1_0_0, type=UpperCAmelCase__, help="""The number of examples split to be used as objective_set/test_data""" ) parser.add_argument( """--min_len""", default=1_0_2_6, type=UpperCAmelCase__, help="""The minimum length of the article to be used as objective set""" ) parser.add_argument( """--secondary_learner_max_epochs""", default=1_5, type=UpperCAmelCase__, help="""number of epochs to train secondary learner""" ) parser.add_argument("""--trim""", default=UpperCAmelCase__, type=UpperCAmelCase__, help="""truncate the example if it exceeds context length""" ) parser.add_argument( """--threshold""", default=1.0, type=UpperCAmelCase__, help=( """The threshold value used by secondary learner to filter the train_data and allow only""" """ informative data as input to the model""" ), ) parser.add_argument("""--finetuned_model_name""", default="""gpt2_finetuned.pt""", type=UpperCAmelCase__, help="""finetuned_model_name""" ) parser.add_argument( """--recopy_model""", default=UpperCAmelCase__, type=UpperCAmelCase__, help="""Reset the model to the original pretrained GPT-2 weights after each iteration""", ) # function calls # Collecting *n* pairs of context and information gain(X, IG(X)) for training the secondary learner generate_n_pairs( context_len=3_2, max_steps=1_0, size_objective_set=1_0_0, min_len=1_0_2_6, trim=UpperCAmelCase__, data_file="""data/tokenized_stories_train_wikitext103.jbl""", igf_data_file="""igf_context_pairs.jbl""", ) # Load train data for secondary learner A__ : Tuple = joblib.load("""data/IGF_values.jbl""" ) # Train secondary learner A__ : Tuple = training_secondary_learner( UpperCAmelCase__, secondary_learner_max_epochs=1_5, secondary_learner_batch_size=1_2_8, eval_freq=1_0_0, igf_model_path="""igf_model.pt""", ) # load pretrained gpt2 model A__ : str = GPTaLMHeadModel.from_pretrained("""gpt2""" ) set_seed(4_2 ) # Generate train and test data to train and evaluate gpt2 model A__ , A__ : Union[str, Any] = generate_datasets( context_len=3_2, file="""data/tokenized_stories_train_wikitext103.jbl""", number=1_0_0, min_len=1_0_2_6, trim=UpperCAmelCase__ ) # fine-tuning of the gpt2 model using igf (Information Gain Filtration) finetune( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, context_len=3_2, max_steps=1_0_0_0, batch_size=1_6, threshold=1.0, recopy_model=UpperCAmelCase__, secondary_learner=UpperCAmelCase__, eval_interval=1_0, finetuned_model_name="""gpt2_finetuned.pt""", ) if __name__ == "__main__": main()
296
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_0 ) ->str: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ) or n < 0: raise ValueError("""Invalid input""" ) A__ : Union[str, Any] = 1_0**n A__ : Optional[Any] = 2_8_4_3_3 * (pow(2, 7_8_3_0_4_5_7, UpperCAmelCase__ )) + 1 return str(number % modulus ) if __name__ == "__main__": from doctest import testmod testmod() print(F'{solution(10) = }')
296
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels A_ = object() # For specifying empty leaf dict `{}` A_ = object() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any] ) ->Dict: A__ : Union[str, Any] = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) + 1 ): A__ : Optional[Any] = [x.match(UpperCAmelCase__ ) for x, y in zip(UpperCAmelCase__, ks[i:] )] if matches and all(UpperCAmelCase__ ): return True return False def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Dict: def replace(UpperCAmelCase__ : int, UpperCAmelCase__ : List[str] ): for rule, replacement in rules: if _match(UpperCAmelCase__, UpperCAmelCase__ ): return replacement return val return replace def _lowerCAmelCase ( ) ->Tuple: return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""", UpperCAmelCase__ )), (("transformer", "wte", "embedding"), P("""mp""", UpperCAmelCase__ )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(UpperCAmelCase__, """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(UpperCAmelCase__, """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : Union[str, Any] = _get_partition_rules() A__ : int = _replacement_rules(UpperCAmelCase__ ) A__ : Tuple = {k: _unmatched for k in flatten_dict(UpperCAmelCase__ )} A__ : Optional[int] = {k: replace(UpperCAmelCase__, UpperCAmelCase__ ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(UpperCAmelCase__ ) )
296
1
"""simple docstring""" import math from datetime import datetime, timedelta def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->datetime: A__ : List[str] = year % 1_9 A__ : Any = year % 4 A__ : str = year % 7 A__ : Any = math.floor(year / 1_0_0 ) A__ : List[str] = math.floor((1_3 + 8 * leap_day_inhibits) / 2_5 ) A__ : Tuple = leap_day_inhibits / 4 A__ : Dict = ( 1_5 - lunar_orbit_correction + leap_day_inhibits - leap_day_reinstall_number ) % 3_0 A__ : int = (4 + leap_day_inhibits - leap_day_reinstall_number) % 7 # days to be added to March 21 A__ : Union[str, Any] = (1_9 * metonic_cycle + secular_moon_shift) % 3_0 # PHM -> Paschal Full Moon A__ : List[str] = ( 2 * julian_leap_year + 4 * non_leap_year + 6 * days_to_add + century_starting_point ) % 7 if days_to_add == 2_9 and days_from_phm_to_sunday == 6: return datetime(UpperCAmelCase__, 4, 1_9 ) elif days_to_add == 2_8 and days_from_phm_to_sunday == 6: return datetime(UpperCAmelCase__, 4, 1_8 ) else: return datetime(UpperCAmelCase__, 3, 2_2 ) + timedelta( days=int(days_to_add + days_from_phm_to_sunday ) ) if __name__ == "__main__": for year in (1994, 2000, 2010, 2021, 2023): A_ = '''will be''' if year > datetime.now().year else '''was''' print(F'Easter in {year} {tense} {gauss_easter(year)}')
296
"""simple docstring""" import unittest from transformers import MraConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask if is_torch_available(): import torch from transformers import ( MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, MraModel, ) from transformers.models.mra.modeling_mra import MRA_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] , snake_case : Tuple , snake_case : List[str]=2 , snake_case : List[str]=8 , snake_case : List[Any]=True , snake_case : Optional[Any]=True , snake_case : List[Any]=True , snake_case : Dict=True , snake_case : Tuple=99 , snake_case : Dict=16 , snake_case : Dict=5 , snake_case : int=2 , snake_case : Any=36 , snake_case : str="gelu" , snake_case : Dict=0.0 , snake_case : List[Any]=0.0 , snake_case : int=512 , snake_case : List[Any]=16 , snake_case : Tuple=2 , snake_case : Any=0.02 , snake_case : Optional[Any]=3 , snake_case : List[Any]=4 , snake_case : str=None , ): '''simple docstring''' A__ : Union[str, Any] = parent A__ : Optional[Any] = batch_size A__ : Dict = seq_length A__ : str = is_training A__ : Tuple = use_input_mask A__ : Dict = use_token_type_ids A__ : Dict = use_labels A__ : int = vocab_size A__ : List[str] = hidden_size A__ : Union[str, Any] = num_hidden_layers A__ : int = num_attention_heads A__ : List[str] = intermediate_size A__ : int = hidden_act A__ : str = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : Any = max_position_embeddings A__ : Optional[int] = type_vocab_size A__ : int = type_sequence_label_size A__ : Optional[Any] = initializer_range A__ : int = num_labels A__ : Optional[int] = num_choices A__ : Optional[int] = scope def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Dict = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Any = None if self.use_input_mask: A__ : Any = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Optional[int] = None if self.use_token_type_ids: A__ : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : Dict = None A__ : List[str] = None A__ : Union[str, Any] = None if self.use_labels: A__ : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Any = ids_tensor([self.batch_size] , self.num_choices ) A__ : Optional[int] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return MraConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.get_config() A__ : List[str] = 300 return config def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Tuple = self.prepare_config_and_inputs() A__ : List[str] = True A__ : List[str] = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) A__ : int = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels, encoder_hidden_states, encoder_attention_mask, ) def _UpperCamelCase ( self : Any , snake_case : Any , snake_case : Tuple , snake_case : Any , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Optional[int] , snake_case : Dict ): '''simple docstring''' A__ : List[str] = MraModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) A__ : List[str] = model(snake_case , token_type_ids=snake_case ) A__ : Union[str, Any] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : List[Any] , snake_case : Any , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Dict , snake_case : str , snake_case : Dict , snake_case : str , ): '''simple docstring''' A__ : Dict = True A__ : Optional[Any] = MraModel(snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , encoder_attention_mask=snake_case , ) A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , encoder_hidden_states=snake_case , ) A__ : Optional[int] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : str , snake_case : Union[str, Any] , snake_case : Dict , snake_case : List[str] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Dict , snake_case : Dict , snake_case : Dict , snake_case : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : Dict = MraForQuestionAnswering(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , start_positions=snake_case , end_positions=snake_case , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : Tuple , snake_case : List[Any] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Optional[int] , snake_case : List[str] , snake_case : Union[str, Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Optional[Any] = MraForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict , snake_case : str , snake_case : List[Any] , snake_case : Any , snake_case : Dict , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : str = self.num_labels A__ : Union[str, Any] = MraForTokenClassification(config=snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : Tuple , snake_case : Optional[Any] , snake_case : Optional[int] , snake_case : int , snake_case : Optional[Any] , snake_case : List[str] , snake_case : Dict , snake_case : Optional[Any] ): '''simple docstring''' A__ : List[str] = self.num_choices A__ : str = MraForMultipleChoice(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = input_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Dict = token_type_ids.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : Tuple = input_mask.unsqueeze(1 ).expand(-1 , self.num_choices , -1 ).contiguous() A__ : str = model( snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : Dict = config_and_inputs A__ : Optional[int] = {"""input_ids""": input_ids, """token_type_ids""": token_type_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( MraModel, MraForMaskedLM, MraForMultipleChoice, MraForQuestionAnswering, MraForSequenceClassification, MraForTokenClassification, ) if is_torch_available() else () ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = () def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Optional[Any] = MraModelTester(self ) A__ : List[str] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : List[str] = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : Any ): '''simple docstring''' for model_name in MRA_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : str = MraModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) @unittest.skip(reason="""MRA does not output attentions""" ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = MraModel.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Any = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : List[Any] = torch.Size((1, 256, 768) ) self.assertEqual(output.shape , snake_case ) A__ : int = torch.tensor( [[[-0.0140, 0.0830, -0.0381], [0.1546, 0.1402, 0.0220], [0.1162, 0.0851, 0.0165]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Union[str, Any] = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-512-4""" ) A__ : Tuple = torch.arange(256 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Dict = 5_0265 A__ : List[str] = torch.Size((1, 256, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : List[Any] = torch.tensor( [[[9.2595, -3.6038, 11.8819], [9.3869, -3.2693, 11.0956], [11.8524, -3.4938, 13.1210]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Any = MraForMaskedLM.from_pretrained("""uw-madison/mra-base-4096-8-d3""" ) A__ : List[Any] = torch.arange(4096 ).unsqueeze(0 ) with torch.no_grad(): A__ : List[Any] = model(snake_case )[0] A__ : Union[str, Any] = 5_0265 A__ : Optional[Any] = torch.Size((1, 4096, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : Optional[int] = torch.tensor( [[[5.4789, -2.3564, 7.5064], [7.9067, -1.3369, 9.9668], [9.0712, -1.8106, 7.0380]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) )
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->bool: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): A__ : List[str] = f'Input value of [number={number}] must be an integer' raise TypeError(UpperCAmelCase__ ) if number < 0: return False A__ : List[str] = number * number while number > 0: if number % 1_0 != number_square % 1_0: return False number //= 1_0 number_square //= 1_0 return True if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets A_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' A_ = '''\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. ''' A_ = ''' Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {"raw_values", "uniform_average"} or array-like of shape (n_outputs,), default="uniform_average" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. "raw_values" : Returns a full set of errors in case of multioutput input. "uniform_average" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric("mse") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {\'mse\': 0.6123724356957945} If you\'re using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric("mse", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput=\'raw_values\') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mse\': array([0.41666667, 1. ])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : Dict ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html""" ] , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("""float""" ) ), "references": datasets.Sequence(datasets.Value("""float""" ) ), } else: return { "predictions": datasets.Value("""float""" ), "references": datasets.Value("""float""" ), } def _UpperCamelCase ( self : List[str] , snake_case : Dict , snake_case : List[Any] , snake_case : List[str]=None , snake_case : List[Any]="uniform_average" , snake_case : int=True ): '''simple docstring''' A__ : Optional[int] = mean_squared_error( snake_case , snake_case , sample_weight=snake_case , multioutput=snake_case , squared=snake_case ) return {"mse": mse}
296
1
"""simple docstring""" import argparse import fairseq import torch from transformers import UniSpeechSatConfig, UniSpeechSatForCTC, UniSpeechSatForPreTraining, logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) A_ = { '''post_extract_proj''': '''feature_projection.projection''', '''encoder.pos_conv.0''': '''encoder.pos_conv_embed.conv''', '''self_attn.k_proj''': '''encoder.layers.*.attention.k_proj''', '''self_attn.v_proj''': '''encoder.layers.*.attention.v_proj''', '''self_attn.q_proj''': '''encoder.layers.*.attention.q_proj''', '''self_attn.out_proj''': '''encoder.layers.*.attention.out_proj''', '''self_attn_layer_norm''': '''encoder.layers.*.layer_norm''', '''fc1''': '''encoder.layers.*.feed_forward.intermediate_dense''', '''fc2''': '''encoder.layers.*.feed_forward.output_dense''', '''final_layer_norm''': '''encoder.layers.*.final_layer_norm''', '''encoder.layer_norm''': '''encoder.layer_norm''', '''encoder.layer_norm_for_extract''': '''layer_norm_for_extract''', '''w2v_model.layer_norm''': '''feature_projection.layer_norm''', '''quantizer.weight_proj''': '''quantizer.weight_proj''', '''quantizer.vars''': '''quantizer.codevectors''', '''project_q''': '''project_q''', '''final_proj''': '''project_hid''', '''w2v_encoder.proj''': '''lm_head''', '''label_embs_concat''': '''label_embeddings_concat''', '''mask_emb''': '''masked_spec_embed''', '''spk_proj''': '''speaker_proj''', } A_ = [ '''lm_head''', '''quantizer.weight_proj''', '''quantizer.codevectors''', '''project_q''', '''project_hid''', '''label_embeddings_concat''', '''speaker_proj''', '''layer_norm_for_extract''', ] def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : str, UpperCAmelCase__ : Dict, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[str] ) ->str: for attribute in key.split(""".""" ): A__ : Dict = getattr(UpperCAmelCase__, UpperCAmelCase__ ) if weight_type is not None: A__ : Union[str, Any] = getattr(UpperCAmelCase__, UpperCAmelCase__ ).shape else: A__ : Union[str, Any] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": A__ : List[str] = value elif weight_type == "weight_g": A__ : Any = value elif weight_type == "weight_v": A__ : Union[str, Any] = value elif weight_type == "bias": A__ : Any = value else: A__ : List[Any] = value logger.info(f'{key + "." + weight_type if weight_type is not None else ""} was initialized from {full_name}.' ) def _lowerCAmelCase ( UpperCAmelCase__ : str, UpperCAmelCase__ : Optional[int] ) ->Union[str, Any]: A__ : List[Any] = [] A__ : int = fairseq_model.state_dict() A__ : List[Any] = hf_model.unispeech_sat.feature_extractor for name, value in fairseq_dict.items(): A__ : Dict = False if "conv_layers" in name: load_conv_layer( UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, hf_model.config.feat_extract_norm == """group""", ) A__ : str = True else: for key, mapped_key in MAPPING.items(): A__ : int = """unispeech_sat.""" + mapped_key if mapped_key not in TOP_LEVEL_KEYS else mapped_key if key in name or key.split("""w2v_model.""" )[-1] == name.split(""".""" )[0]: if "layer_norm_for_extract" in name and (".".join(name.split(""".""" )[:-1] ) != key): # special case since naming is very similar continue A__ : Dict = True if "*" in mapped_key: A__ : Any = name.split(UpperCAmelCase__ )[0].split(""".""" )[-2] A__ : str = mapped_key.replace("""*""", UpperCAmelCase__ ) if "weight_g" in name: A__ : Any = """weight_g""" elif "weight_v" in name: A__ : Dict = """weight_v""" elif "bias" in name: A__ : Optional[int] = """bias""" elif "weight" in name: # TODO: don't match quantizer.weight_proj A__ : List[str] = """weight""" else: A__ : int = None set_recursively(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) continue if not is_used: unused_weights.append(UpperCAmelCase__ ) logger.warning(f'Unused weights: {unused_weights}' ) def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : List[str], UpperCAmelCase__ : str, UpperCAmelCase__ : str, UpperCAmelCase__ : Optional[int] ) ->Union[str, Any]: A__ : Tuple = full_name.split("""conv_layers.""" )[-1] A__ : Dict = name.split(""".""" ) A__ : List[str] = int(items[0] ) A__ : Optional[Any] = int(items[1] ) if type_id == 0: if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.bias.data.shape} was found.' ) A__ : Optional[int] = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].conv.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor.conv_layers[layer_id].conv.weight.data.shape} was found.' ) A__ : str = value logger.info(f'Feat extract conv layer {layer_id} was initialized from {full_name}.' ) elif (type_id == 2 and not use_group_norm) or (type_id == 2 and layer_id == 0 and use_group_norm): if "bias" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.bias.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.bias.data.shape} was found.' ) A__ : List[str] = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) elif "weight" in name: if value.shape != feature_extractor.conv_layers[layer_id].layer_norm.weight.data.shape: raise ValueError( f'{full_name} has size {value.shape}, but' f' {feature_extractor[layer_id].layer_norm.weight.data.shape} was found.' ) A__ : List[Any] = value logger.info(f'Feat extract layer norm weight of layer {layer_id} was initialized from {full_name}.' ) else: unused_weights.append(UpperCAmelCase__ ) @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Union[str, Any]=None, UpperCAmelCase__ : Dict=None, UpperCAmelCase__ : str=True ) ->int: if config_path is not None: A__ : int = UniSpeechSatConfig.from_pretrained(UpperCAmelCase__ ) else: A__ : List[str] = UniSpeechSatConfig() A__ : List[Any] = """""" if is_finetuned: A__ : List[Any] = UniSpeechSatForCTC(UpperCAmelCase__ ) else: A__ : List[str] = UniSpeechSatForPreTraining(UpperCAmelCase__ ) A__ , A__ , A__ : Any = fairseq.checkpoint_utils.load_model_ensemble_and_task( [checkpoint_path], arg_overrides={"""data""": """/""".join(dict_path.split("""/""" )[:-1] )} ) A__ : Dict = model[0].eval() recursively_load_weights(UpperCAmelCase__, UpperCAmelCase__ ) hf_wavavec.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model.''') parser.add_argument('''--checkpoint_path''', default=None, type=str, help='''Path to fairseq checkpoint''') parser.add_argument('''--dict_path''', default=None, type=str, help='''Path to dict of fine-tuned model''') parser.add_argument('''--config_path''', default=None, type=str, help='''Path to hf config.json of model to convert''') parser.add_argument( '''--not_finetuned''', action='''store_true''', help='''Whether the model to convert is a fine-tuned model or not''' ) A_ = parser.parse_args() convert_unispeech_sat_checkpoint( args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.dict_path, not args.not_finetuned )
296
"""simple docstring""" import warnings from ..trainer import Trainer from ..utils import logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : Optional[int] , snake_case : List[str]=None , **snake_case : Any ): '''simple docstring''' warnings.warn( """`SageMakerTrainer` is deprecated and will be removed in v5 of Transformers. You can use `Trainer` """ """instead.""" , snake_case , ) super().__init__(args=snake_case , **snake_case )
296
1
"""simple docstring""" import re from flax.core.frozen_dict import freeze from flax.traverse_util import flatten_dict, unflatten_dict from jax.experimental import PartitionSpec as P # Sentinels A_ = object() # For specifying empty leaf dict `{}` A_ = object() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any] ) ->Dict: A__ : Union[str, Any] = tuple((re.compile(x + """$""" ) for x in qs) ) for i in range(len(UpperCAmelCase__ ) - len(UpperCAmelCase__ ) + 1 ): A__ : Optional[Any] = [x.match(UpperCAmelCase__ ) for x, y in zip(UpperCAmelCase__, ks[i:] )] if matches and all(UpperCAmelCase__ ): return True return False def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Dict: def replace(UpperCAmelCase__ : int, UpperCAmelCase__ : List[str] ): for rule, replacement in rules: if _match(UpperCAmelCase__, UpperCAmelCase__ ): return replacement return val return replace def _lowerCAmelCase ( ) ->Tuple: return [ # embeddings (("transformer", "wpe", "embedding"), P("""mp""", UpperCAmelCase__ )), (("transformer", "wte", "embedding"), P("""mp""", UpperCAmelCase__ )), # atention (("attention", "(q_proj|k_proj|v_proj)", "kernel"), P(UpperCAmelCase__, """mp""" )), (("attention", "out_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("attention", "out_proj", "bias"), None), # mlp (("mlp", "c_fc", "kernel"), P(UpperCAmelCase__, """mp""" )), (("mlp", "c_fc", "bias"), P("""mp""" )), (("mlp", "c_proj", "kernel"), P("""mp""", UpperCAmelCase__ )), (("mlp", "c_proj", "bias"), None), # layer norms ((r"ln_\d+", "bias"), None), ((r"\d+", r"ln_\d+", "scale"), None), (("ln_f", "bias"), None), (("ln_f", "scale"), None), ] def _lowerCAmelCase ( UpperCAmelCase__ : Tuple ) ->Any: A__ : Union[str, Any] = _get_partition_rules() A__ : int = _replacement_rules(UpperCAmelCase__ ) A__ : Tuple = {k: _unmatched for k in flatten_dict(UpperCAmelCase__ )} A__ : Optional[int] = {k: replace(UpperCAmelCase__, UpperCAmelCase__ ) for k, v in initd.items()} assert _unmatched not in result.values(), "Incomplete partition spec." return freeze(unflatten_dict(UpperCAmelCase__ ) )
296
"""simple docstring""" import itertools import os import random import tempfile import unittest import numpy as np from datasets import load_dataset from transformers import is_speech_available from transformers.testing_utils import check_json_file_has_correct_format, require_torch, require_torchaudio from transformers.utils.import_utils import is_torch_available from ...test_sequence_feature_extraction_common import SequenceFeatureExtractionTestMixin if is_speech_available(): from transformers import WhisperFeatureExtractor if is_torch_available(): import torch A_ = random.Random() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Tuple=1.0, UpperCAmelCase__ : Optional[int]=None, UpperCAmelCase__ : str=None ) ->Union[str, Any]: if rng is None: A__ : Optional[int] = global_rng A__ : Optional[Any] = [] for batch_idx in range(shape[0] ): values.append([] ) for _ in range(shape[1] ): values[-1].append(rng.random() * scale ) return values @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[str]=7 , snake_case : str=400 , snake_case : Optional[Any]=2000 , snake_case : Union[str, Any]=10 , snake_case : str=160 , snake_case : List[str]=8 , snake_case : List[Any]=0.0 , snake_case : Optional[Any]=4000 , snake_case : Any=False , snake_case : int=True , ): '''simple docstring''' A__ : Any = parent A__ : str = batch_size A__ : List[str] = min_seq_length A__ : Dict = max_seq_length A__ : str = (self.max_seq_length - self.min_seq_length) // (self.batch_size - 1) A__ : Dict = padding_value A__ : Optional[Any] = sampling_rate A__ : Any = return_attention_mask A__ : Optional[int] = do_normalize A__ : Tuple = feature_size A__ : Optional[Any] = chunk_length A__ : Union[str, Any] = hop_length def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return { "feature_size": self.feature_size, "hop_length": self.hop_length, "chunk_length": self.chunk_length, "padding_value": self.padding_value, "sampling_rate": self.sampling_rate, "return_attention_mask": self.return_attention_mask, "do_normalize": self.do_normalize, } def _UpperCamelCase ( self : Union[str, Any] , snake_case : Dict=False , snake_case : Optional[Any]=False ): '''simple docstring''' def _flatten(snake_case : Dict ): return list(itertools.chain(*snake_case ) ) if equal_length: A__ : Dict = [floats_list((self.max_seq_length, self.feature_size) ) for _ in range(self.batch_size )] else: # make sure that inputs increase in size A__ : Optional[int] = [ floats_list((x, self.feature_size) ) for x in range(self.min_seq_length , self.max_seq_length , self.seq_length_diff ) ] if numpify: A__ : List[str] = [np.asarray(snake_case ) for x in speech_inputs] return speech_inputs @require_torch @require_torchaudio class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = WhisperFeatureExtractor if is_speech_available() else None def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : str = WhisperFeatureExtractionTester(self ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : List[Any] = feat_extract_first.save_pretrained(snake_case )[0] check_json_file_has_correct_format(snake_case ) A__ : Union[str, Any] = self.feature_extraction_class.from_pretrained(snake_case ) A__ : str = feat_extract_first.to_dict() A__ : Union[str, Any] = feat_extract_second.to_dict() A__ : List[Any] = feat_extract_first.mel_filters A__ : Optional[Any] = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_dict ) with tempfile.TemporaryDirectory() as tmpdirname: A__ : Any = os.path.join(snake_case , """feat_extract.json""" ) feat_extract_first.to_json_file(snake_case ) A__ : int = self.feature_extraction_class.from_json_file(snake_case ) A__ : Dict = feat_extract_first.to_dict() A__ : str = feat_extract_second.to_dict() A__ : str = feat_extract_first.mel_filters A__ : Dict = feat_extract_second.mel_filters self.assertTrue(np.allclose(snake_case , snake_case ) ) self.assertEqual(snake_case , snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : Any = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) # create three inputs of length 800, 1000, and 1200 A__ : str = [floats_list((1, x) )[0] for x in range(800 , 1400 , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] # Test feature size A__ : Dict = feature_extractor(snake_case , padding="""max_length""" , return_tensors="""np""" ).input_features self.assertTrue(input_features.ndim == 3 ) self.assertTrue(input_features.shape[-1] == feature_extractor.nb_max_frames ) self.assertTrue(input_features.shape[-2] == feature_extractor.feature_size ) # Test not batched input A__ : str = feature_extractor(speech_inputs[0] , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(np_speech_inputs[0] , return_tensors="""np""" ).input_features self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test batched A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test 2-D numpy arrays are batched. A__ : Tuple = [floats_list((1, x) )[0] for x in (800, 800, 800)] A__ : str = np.asarray(snake_case ) A__ : List[str] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) # Test truncation required A__ : Optional[Any] = [floats_list((1, x) )[0] for x in range(200 , (feature_extractor.n_samples + 500) , 200 )] A__ : Union[str, Any] = [np.asarray(snake_case ) for speech_input in speech_inputs] A__ : Union[str, Any] = [x[: feature_extractor.n_samples] for x in speech_inputs] A__ : str = [np.asarray(snake_case ) for speech_input in speech_inputs_truncated] A__ : Optional[int] = feature_extractor(snake_case , return_tensors="""np""" ).input_features A__ : str = feature_extractor(snake_case , return_tensors="""np""" ).input_features for enc_seq_a, enc_seq_a in zip(snake_case , snake_case ): self.assertTrue(np.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' import torch A__ : Tuple = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : List[str] = np.random.rand(100 , 32 ).astype(np.floataa ) A__ : Tuple = np_speech_inputs.tolist() for inputs in [py_speech_inputs, np_speech_inputs]: A__ : Optional[Any] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""np""" ) self.assertTrue(np_processed.input_features.dtype == np.floataa ) A__ : Optional[int] = feature_extractor.pad([{"""input_features""": inputs}] , return_tensors="""pt""" ) self.assertTrue(pt_processed.input_features.dtype == torch.floataa ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[int] ): '''simple docstring''' A__ : int = load_dataset("""hf-internal-testing/librispeech_asr_dummy""" , """clean""" , split="""validation""" ) # automatic decoding with librispeech A__ : Union[str, Any] = ds.sort("""id""" ).select(range(snake_case ) )[:num_samples]["""audio"""] return [x["array"] for x in speech_samples] def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = torch.tensor( [ 0.1193, -0.0946, -0.1098, -0.0196, 0.0225, -0.0690, -0.1736, 0.0951, 0.0971, -0.0817, -0.0702, 0.0162, 0.0260, 0.0017, -0.0192, -0.1678, 0.0709, -0.1867, -0.0655, -0.0274, -0.0234, -0.1884, -0.0516, -0.0554, -0.0274, -0.1425, -0.1423, 0.0837, 0.0377, -0.0854 ] ) # fmt: on A__ : Optional[Any] = self._load_datasamples(1 ) A__ : Union[str, Any] = WhisperFeatureExtractor() A__ : List[str] = feature_extractor(snake_case , return_tensors="""pt""" ).input_features self.assertEqual(input_features.shape , (1, 80, 3000) ) self.assertTrue(torch.allclose(input_features[0, 0, :30] , snake_case , atol=1e-4 ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Union[str, Any] = self.feature_extraction_class(**self.feat_extract_tester.prepare_feat_extract_dict() ) A__ : Union[str, Any] = self._load_datasamples(1 )[0] A__ : Any = ((audio - audio.min()) / (audio.max() - audio.min())) * 6_5535 # Rescale to [0, 65535] to show issue A__ : str = feat_extract.zero_mean_unit_var_norm([audio] , attention_mask=snake_case )[0] self.assertTrue(np.all(np.mean(snake_case ) < 1e-3 ) ) self.assertTrue(np.all(np.abs(np.var(snake_case ) - 1 ) < 1e-3 ) )
296
1
"""simple docstring""" import contextlib import copy import random from typing import Any, Dict, Iterable, Optional, Union import numpy as np import torch from .utils import deprecate, is_transformers_available if is_transformers_available(): import transformers def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Optional[int]: random.seed(UpperCAmelCase__ ) np.random.seed(UpperCAmelCase__ ) torch.manual_seed(UpperCAmelCase__ ) torch.cuda.manual_seed_all(UpperCAmelCase__ ) # ^^ safe to call this function even if cuda is not available class __SCREAMING_SNAKE_CASE : def __init__( self : Any , snake_case : Iterable[torch.nn.Parameter] , snake_case : float = 0.9999 , snake_case : float = 0.0 , snake_case : int = 0 , snake_case : bool = False , snake_case : Union[float, int] = 1.0 , snake_case : Union[float, int] = 2 / 3 , snake_case : Optional[Any] = None , snake_case : Dict[str, Any] = None , **snake_case : Tuple , ): '''simple docstring''' if isinstance(snake_case , torch.nn.Module ): A__ : Any = ( """Passing a `torch.nn.Module` to `ExponentialMovingAverage` is deprecated. """ """Please pass the parameters of the module instead.""" ) deprecate( """passing a `torch.nn.Module` to `ExponentialMovingAverage`""" , """1.0.0""" , snake_case , standard_warn=snake_case , ) A__ : int = parameters.parameters() # set use_ema_warmup to True if a torch.nn.Module is passed for backwards compatibility A__ : Any = True if kwargs.get("""max_value""" , snake_case ) is not None: A__ : Union[str, Any] = """The `max_value` argument is deprecated. Please use `decay` instead.""" deprecate("""max_value""" , """1.0.0""" , snake_case , standard_warn=snake_case ) A__ : Tuple = kwargs["""max_value"""] if kwargs.get("""min_value""" , snake_case ) is not None: A__ : List[str] = """The `min_value` argument is deprecated. Please use `min_decay` instead.""" deprecate("""min_value""" , """1.0.0""" , snake_case , standard_warn=snake_case ) A__ : List[str] = kwargs["""min_value"""] A__ : Any = list(snake_case ) A__ : Optional[Any] = [p.clone().detach() for p in parameters] if kwargs.get("""device""" , snake_case ) is not None: A__ : str = """The `device` argument is deprecated. Please use `to` instead.""" deprecate("""device""" , """1.0.0""" , snake_case , standard_warn=snake_case ) self.to(device=kwargs["""device"""] ) A__ : List[str] = None A__ : Union[str, Any] = decay A__ : Tuple = min_decay A__ : Tuple = update_after_step A__ : Optional[Any] = use_ema_warmup A__ : List[Any] = inv_gamma A__ : Optional[int] = power A__ : Optional[Any] = 0 A__ : int = None # set in `step()` A__ : int = model_cls A__ : Any = model_config @classmethod def _UpperCamelCase ( cls : str , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : str = model_cls.load_config(snake_case , return_unused_kwargs=snake_case ) A__ : Union[str, Any] = model_cls.from_pretrained(snake_case ) A__ : List[Any] = cls(model.parameters() , model_cls=snake_case , model_config=model.config ) ema_model.load_state_dict(snake_case ) return ema_model def _UpperCamelCase ( self : int , snake_case : Optional[Any] ): '''simple docstring''' if self.model_cls is None: raise ValueError("""`save_pretrained` can only be used if `model_cls` was defined at __init__.""" ) if self.model_config is None: raise ValueError("""`save_pretrained` can only be used if `model_config` was defined at __init__.""" ) A__ : List[Any] = self.model_cls.from_config(self.model_config ) A__ : List[str] = self.state_dict() state_dict.pop("""shadow_params""" , snake_case ) model.register_to_config(**snake_case ) self.copy_to(model.parameters() ) model.save_pretrained(snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : Any = max(0 , optimization_step - self.update_after_step - 1 ) if step <= 0: return 0.0 if self.use_ema_warmup: A__ : Optional[Any] = 1 - (1 + step / self.inv_gamma) ** -self.power else: A__ : Optional[int] = (1 + step) / (10 + step) A__ : Dict = min(snake_case , self.decay ) # make sure decay is not smaller than min_decay A__ : Dict = max(snake_case , self.min_decay ) return cur_decay_value @torch.no_grad() def _UpperCamelCase ( self : Optional[int] , snake_case : Iterable[torch.nn.Parameter] ): '''simple docstring''' if isinstance(snake_case , torch.nn.Module ): A__ : Any = ( """Passing a `torch.nn.Module` to `ExponentialMovingAverage.step` is deprecated. """ """Please pass the parameters of the module instead.""" ) deprecate( """passing a `torch.nn.Module` to `ExponentialMovingAverage.step`""" , """1.0.0""" , snake_case , standard_warn=snake_case , ) A__ : Dict = parameters.parameters() A__ : str = list(snake_case ) self.optimization_step += 1 # Compute the decay factor for the exponential moving average. A__ : str = self.get_decay(self.optimization_step ) A__ : Any = decay A__ : str = 1 - decay A__ : List[str] = contextlib.nullcontext if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): import deepspeed for s_param, param in zip(self.shadow_params , snake_case ): if is_transformers_available() and transformers.deepspeed.is_deepspeed_zeroa_enabled(): A__ : int = deepspeed.zero.GatheredParameters(snake_case , modifier_rank=snake_case ) with context_manager(): if param.requires_grad: s_param.sub_(one_minus_decay * (s_param - param) ) else: s_param.copy_(snake_case ) def _UpperCamelCase ( self : Tuple , snake_case : Iterable[torch.nn.Parameter] ): '''simple docstring''' A__ : Tuple = list(snake_case ) for s_param, param in zip(self.shadow_params , snake_case ): param.data.copy_(s_param.to(param.device ).data ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Any=None , snake_case : List[Any]=None ): '''simple docstring''' A__ : str = [ p.to(device=snake_case , dtype=snake_case ) if p.is_floating_point() else p.to(device=snake_case ) for p in self.shadow_params ] def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return { "decay": self.decay, "min_decay": self.min_decay, "optimization_step": self.optimization_step, "update_after_step": self.update_after_step, "use_ema_warmup": self.use_ema_warmup, "inv_gamma": self.inv_gamma, "power": self.power, "shadow_params": self.shadow_params, } def _UpperCamelCase ( self : Union[str, Any] , snake_case : Iterable[torch.nn.Parameter] ): '''simple docstring''' A__ : int = [param.detach().cpu().clone() for param in parameters] def _UpperCamelCase ( self : int , snake_case : Iterable[torch.nn.Parameter] ): '''simple docstring''' if self.temp_stored_params is None: raise RuntimeError("""This ExponentialMovingAverage has no `store()`ed weights """ """to `restore()`""" ) for c_param, param in zip(self.temp_stored_params , snake_case ): param.data.copy_(c_param.data ) # Better memory-wise. A__ : Any = None def _UpperCamelCase ( self : Union[str, Any] , snake_case : dict ): '''simple docstring''' A__ : int = copy.deepcopy(snake_case ) A__ : str = state_dict.get("""decay""" , self.decay ) if self.decay < 0.0 or self.decay > 1.0: raise ValueError("""Decay must be between 0 and 1""" ) A__ : Tuple = state_dict.get("""min_decay""" , self.min_decay ) if not isinstance(self.min_decay , snake_case ): raise ValueError("""Invalid min_decay""" ) A__ : List[Any] = state_dict.get("""optimization_step""" , self.optimization_step ) if not isinstance(self.optimization_step , snake_case ): raise ValueError("""Invalid optimization_step""" ) A__ : Optional[Any] = state_dict.get("""update_after_step""" , self.update_after_step ) if not isinstance(self.update_after_step , snake_case ): raise ValueError("""Invalid update_after_step""" ) A__ : Optional[int] = state_dict.get("""use_ema_warmup""" , self.use_ema_warmup ) if not isinstance(self.use_ema_warmup , snake_case ): raise ValueError("""Invalid use_ema_warmup""" ) A__ : List[Any] = state_dict.get("""inv_gamma""" , self.inv_gamma ) if not isinstance(self.inv_gamma , (float, int) ): raise ValueError("""Invalid inv_gamma""" ) A__ : Optional[int] = state_dict.get("""power""" , self.power ) if not isinstance(self.power , (float, int) ): raise ValueError("""Invalid power""" ) A__ : str = state_dict.get("""shadow_params""" , snake_case ) if shadow_params is not None: A__ : Optional[Any] = shadow_params if not isinstance(self.shadow_params , snake_case ): raise ValueError("""shadow_params must be a list""" ) if not all(isinstance(snake_case , torch.Tensor ) for p in self.shadow_params ): raise ValueError("""shadow_params must all be Tensors""" )
296
"""simple docstring""" import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : List[str] ): '''simple docstring''' A__ : Optional[int] = (0, 0) A__ : Dict = None A__ : int = 0 A__ : str = 0 A__ : Optional[Any] = 0 def __eq__( self : str , snake_case : Optional[int] ): '''simple docstring''' return self.position == cell.position def _UpperCamelCase ( self : List[str] ): '''simple docstring''' print(self.position ) class __SCREAMING_SNAKE_CASE : def __init__( self : int , snake_case : Any=(5, 5) ): '''simple docstring''' A__ : Optional[int] = np.zeros(snake_case ) A__ : List[Any] = world_size[0] A__ : Dict = world_size[1] def _UpperCamelCase ( self : Any ): '''simple docstring''' print(self.w ) def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : int = [ (-1, -1), (-1, 0), (-1, 1), (0, -1), (0, 1), (1, -1), (1, 0), (1, 1), ] A__ : int = cell.position[0] A__ : str = cell.position[1] A__ : Any = [] for n in neughbour_cord: A__ : List[Any] = current_x + n[0] A__ : Tuple = current_y + n[1] if 0 <= x < self.world_x_limit and 0 <= y < self.world_y_limit: A__ : List[Any] = Cell() A__ : str = (x, y) A__ : Optional[Any] = cell neighbours.append(snake_case ) return neighbours def _lowerCAmelCase ( UpperCAmelCase__ : List[str], UpperCAmelCase__ : Optional[Any], UpperCAmelCase__ : Dict ) ->Dict: A__ : Union[str, Any] = [] A__ : Optional[int] = [] _open.append(UpperCAmelCase__ ) while _open: A__ : List[Any] = np.argmin([n.f for n in _open] ) A__ : Union[str, Any] = _open[min_f] _closed.append(_open.pop(UpperCAmelCase__ ) ) if current == goal: break for n in world.get_neigbours(UpperCAmelCase__ ): for c in _closed: if c == n: continue A__ : Dict = current.g + 1 A__ , A__ : int = n.position A__ , A__ : Optional[int] = goal.position A__ : Union[str, Any] = (ya - ya) ** 2 + (xa - xa) ** 2 A__ : Optional[int] = n.h + n.g for c in _open: if c == n and c.f < n.f: continue _open.append(UpperCAmelCase__ ) A__ : List[str] = [] while current.parent is not None: path.append(current.position ) A__ : Union[str, Any] = current.parent path.append(current.position ) return path[::-1] if __name__ == "__main__": A_ = Gridworld() # Start position and goal A_ = Cell() A_ = (0, 0) A_ = Cell() A_ = (4, 4) print(F'path from {start.position} to {goal.position}') A_ = astar(world, start, goal) # Just for visual reasons. for i in s: A_ = 1 print(world.w)
296
1
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = TextToVideoSDPipeline snake_case_ = TEXT_TO_IMAGE_PARAMS snake_case_ = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. snake_case_ = frozenset( [ 'num_inference_steps', 'generator', 'latents', 'return_dict', 'callback', 'callback_steps', ] ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' torch.manual_seed(0 ) A__ : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("""CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """CrossAttnDownBlock3D""", """DownBlock3D""") , up_block_types=("""UpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""", """CrossAttnUpBlock3D""") , cross_attention_dim=32 , attention_head_dim=4 , ) A__ : Optional[int] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="""scaled_linear""" , clip_sample=snake_case , set_alpha_to_one=snake_case , ) torch.manual_seed(0 ) A__ : Optional[int] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["""DownEncoderBlock2D""", """DownEncoderBlock2D"""] , up_block_types=["""UpDecoderBlock2D""", """UpDecoderBlock2D"""] , latent_channels=4 , sample_size=128 , ) torch.manual_seed(0 ) A__ : str = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1e-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1000 , hidden_act="""gelu""" , projection_dim=512 , ) A__ : str = CLIPTextModel(snake_case ) A__ : List[str] = CLIPTokenizer.from_pretrained("""hf-internal-testing/tiny-random-clip""" ) A__ : List[Any] = { """unet""": unet, """scheduler""": scheduler, """vae""": vae, """text_encoder""": text_encoder, """tokenizer""": tokenizer, } return components def _UpperCamelCase ( self : List[Any] , snake_case : Optional[Any] , snake_case : Tuple=0 ): '''simple docstring''' if str(snake_case ).startswith("""mps""" ): A__ : Tuple = torch.manual_seed(snake_case ) else: A__ : str = torch.Generator(device=snake_case ).manual_seed(snake_case ) A__ : Optional[int] = { """prompt""": """A painting of a squirrel eating a burger""", """generator""": generator, """num_inference_steps""": 2, """guidance_scale""": 6.0, """output_type""": """pt""", } return inputs def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = """cpu""" # ensure determinism for the device-dependent torch.Generator A__ : Any = self.get_dummy_components() A__ : Optional[int] = TextToVideoSDPipeline(**snake_case ) A__ : Optional[Any] = sd_pipe.to(snake_case ) sd_pipe.set_progress_bar_config(disable=snake_case ) A__ : Union[str, Any] = self.get_dummy_inputs(snake_case ) A__ : str = """np""" A__ : Dict = sd_pipe(**snake_case ).frames A__ : Optional[Any] = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) A__ : Optional[int] = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1e-2 def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case , expected_max_diff=3e-3 ) @unittest.skipIf( torch_device != """cuda""" or not is_xformers_available() , reason="""XFormers attention is only available with CUDA and `xformers` installed""" , ) def _UpperCamelCase ( self : int ): '''simple docstring''' self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case , expected_max_diff=1e-2 ) @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def _UpperCamelCase ( self : Any ): '''simple docstring''' pass @unittest.skip(reason="""Batching needs to be properly figured out first for this pipeline.""" ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass @unittest.skip(reason="""`num_images_per_prompt` argument is not supported for this pipeline.""" ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' pass def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' return super().test_progress_bar() @slow @skip_mps class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy""" ) A__ : List[Any] = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : Dict = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) A__ : Dict = pipe.to("""cuda""" ) A__ : List[Any] = """Spiderman is surfing""" A__ : Any = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : Any = pipe(snake_case , generator=snake_case , num_inference_steps=25 , output_type="""pt""" ).frames A__ : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2 def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = load_numpy( """https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy""" ) A__ : str = TextToVideoSDPipeline.from_pretrained("""damo-vilab/text-to-video-ms-1.7b""" ) A__ : Dict = pipe.to("""cuda""" ) A__ : List[str] = """Spiderman is surfing""" A__ : Optional[Any] = torch.Generator(device="""cpu""" ).manual_seed(0 ) A__ : List[Any] = pipe(snake_case , generator=snake_case , num_inference_steps=2 , output_type="""pt""" ).frames A__ : List[str] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5e-2
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import DeiTConfig, DeiTForImageClassificationWithTeacher, DeiTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : Tuple=False ) ->str: A__ : Optional[int] = [] for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'deit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'deit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'deit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'deit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'deit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'deit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'deit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'deit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'deit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'deit.encoder.layer.{i}.output.dense.bias') ) # projection layer + position embeddings rename_keys.extend( [ ("""cls_token""", """deit.embeddings.cls_token"""), ("""dist_token""", """deit.embeddings.distillation_token"""), ("""patch_embed.proj.weight""", """deit.embeddings.patch_embeddings.projection.weight"""), ("""patch_embed.proj.bias""", """deit.embeddings.patch_embeddings.projection.bias"""), ("""pos_embed""", """deit.embeddings.position_embeddings"""), ] ) if base_model: # layernorm + pooler rename_keys.extend( [ ("""norm.weight""", """layernorm.weight"""), ("""norm.bias""", """layernorm.bias"""), ("""pre_logits.fc.weight""", """pooler.dense.weight"""), ("""pre_logits.fc.bias""", """pooler.dense.bias"""), ] ) # if just the base model, we should remove "deit" from all keys that start with "deit" A__ : Optional[int] = [(pair[0], pair[1][4:]) if pair[1].startswith("""deit""" ) else pair for pair in rename_keys] else: # layernorm + classification heads rename_keys.extend( [ ("""norm.weight""", """deit.layernorm.weight"""), ("""norm.bias""", """deit.layernorm.bias"""), ("""head.weight""", """cls_classifier.weight"""), ("""head.bias""", """cls_classifier.bias"""), ("""head_dist.weight""", """distillation_classifier.weight"""), ("""head_dist.bias""", """distillation_classifier.bias"""), ] ) return rename_keys def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]=False ) ->str: for i in range(config.num_hidden_layers ): if base_model: A__ : Any = """""" else: A__ : Tuple = """deit.""" # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[Any] = in_proj_weight[ : config.hidden_size, : ] A__ : str = in_proj_bias[: config.hidden_size] A__ : Any = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Dict = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : Optional[Any] = in_proj_weight[ -config.hidden_size :, : ] A__ : Any = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Union[str, Any] ) ->Any: A__ : int = dct.pop(UpperCAmelCase__ ) A__ : Tuple = val def _lowerCAmelCase ( ) ->List[Any]: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Dict, UpperCAmelCase__ : Any ) ->Tuple: A__ : List[Any] = DeiTConfig() # all deit models have fine-tuned heads A__ : Tuple = False # dataset (fine-tuned on ImageNet 2012), patch_size and image_size A__ : str = 1_0_0_0 A__ : List[str] = """huggingface/label-files""" A__ : Dict = """imagenet-1k-id2label.json""" A__ : List[str] = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Dict = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Optional[int] = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} A__ : List[str] = int(deit_name[-6:-4] ) A__ : str = int(deit_name[-3:] ) # size of the architecture if deit_name[9:].startswith("""tiny""" ): A__ : List[str] = 1_9_2 A__ : int = 7_6_8 A__ : List[Any] = 1_2 A__ : Dict = 3 elif deit_name[9:].startswith("""small""" ): A__ : List[Any] = 3_8_4 A__ : List[str] = 1_5_3_6 A__ : Any = 1_2 A__ : Union[str, Any] = 6 if deit_name[9:].startswith("""base""" ): pass elif deit_name[4:].startswith("""large""" ): A__ : int = 1_0_2_4 A__ : str = 4_0_9_6 A__ : Any = 2_4 A__ : int = 1_6 # load original model from timm A__ : Dict = timm.create_model(UpperCAmelCase__, pretrained=UpperCAmelCase__ ) timm_model.eval() # load state_dict of original model, remove and rename some keys A__ : Tuple = timm_model.state_dict() A__ : str = create_rename_keys(UpperCAmelCase__, UpperCAmelCase__ ) for src, dest in rename_keys: rename_key(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : str = DeiTForImageClassificationWithTeacher(UpperCAmelCase__ ).eval() model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by DeiTImageProcessor A__ : int = int( (2_5_6 / 2_2_4) * config.image_size ) # to maintain same ratio w.r.t. 224 images, see https://github.com/facebookresearch/deit/blob/ab5715372db8c6cad5740714b2216d55aeae052e/datasets.py#L103 A__ : Any = DeiTImageProcessor(size=UpperCAmelCase__, crop_size=config.image_size ) A__ : Union[str, Any] = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : Optional[Any] = encoding["""pixel_values"""] A__ : Union[str, Any] = model(UpperCAmelCase__ ) A__ : Union[str, Any] = timm_model(UpperCAmelCase__ ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(UpperCAmelCase__, outputs.logits, atol=1e-3 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {deit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--deit_name''', default='''vit_deit_base_distilled_patch16_224''', type=str, help='''Name of the DeiT timm model you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, help='''Path to the output PyTorch model directory.''' ) A_ = parser.parse_args() convert_deit_checkpoint(args.deit_name, args.pytorch_dump_folder_path)
296
1
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
"""simple docstring""" from __future__ import annotations import time from collections.abc import Sequence from random import randint from matplotlib import pyplot as plt def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int | None, int | None, float]: if not arr: return None, None, 0 if low == high: return low, high, arr[low] A__ : Optional[int] = (low + high) // 2 A__ , A__ , A__ : List[Any] = max_subarray(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_subarray(UpperCAmelCase__, mid + 1, UpperCAmelCase__ ) A__ , A__ , A__ : Union[str, Any] = max_cross_sum(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if left_sum >= right_sum and left_sum >= cross_sum: return left_low, left_high, left_sum elif right_sum >= left_sum and right_sum >= cross_sum: return right_low, right_high, right_sum return cross_left, cross_right, cross_sum def _lowerCAmelCase ( UpperCAmelCase__ : Sequence[float], UpperCAmelCase__ : int, UpperCAmelCase__ : int, UpperCAmelCase__ : int ) ->tuple[int, int, float]: A__ , A__ : Dict = float("""-inf""" ), -1 A__ , A__ : Optional[Any] = float("""-inf""" ), -1 A__ : int | float = 0 for i in range(UpperCAmelCase__, low - 1, -1 ): summ += arr[i] if summ > left_sum: A__ : Optional[int] = summ A__ : Union[str, Any] = i A__ : Optional[Any] = 0 for i in range(mid + 1, high + 1 ): summ += arr[i] if summ > right_sum: A__ : int = summ A__ : Union[str, Any] = i return max_left, max_right, (left_sum + right_sum) def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->float: A__ : Union[str, Any] = [randint(1, UpperCAmelCase__ ) for _ in range(UpperCAmelCase__ )] A__ : Any = time.time() max_subarray(UpperCAmelCase__, 0, input_size - 1 ) A__ : List[Any] = time.time() return end - start def _lowerCAmelCase ( ) ->None: A__ : List[Any] = [1_0, 1_0_0, 1_0_0_0, 1_0_0_0_0, 5_0_0_0_0, 1_0_0_0_0_0, 2_0_0_0_0_0, 3_0_0_0_0_0, 4_0_0_0_0_0, 5_0_0_0_0_0] A__ : Any = [time_max_subarray(UpperCAmelCase__ ) for input_size in input_sizes] print("""No of Inputs\t\tTime Taken""" ) for input_size, runtime in zip(UpperCAmelCase__, UpperCAmelCase__ ): print(UpperCAmelCase__, """\t\t""", UpperCAmelCase__ ) plt.plot(UpperCAmelCase__, UpperCAmelCase__ ) plt.xlabel("""Number of Inputs""" ) plt.ylabel("""Time taken in seconds""" ) plt.show() if __name__ == "__main__": from doctest import testmod testmod()
296
1
"""simple docstring""" from dataclasses import dataclass from typing import List, Optional, Union import numpy as np import torch from ...utils import BaseOutput, OptionalDependencyNotAvailable, is_torch_available, is_transformers_available @dataclass class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 42 try: if not (is_transformers_available() and is_torch_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import * # noqa F403 else: from .pipeline_text_to_video_synth import TextToVideoSDPipeline from .pipeline_text_to_video_synth_imgaimg import VideoToVideoSDPipeline # noqa: F401 from .pipeline_text_to_video_zero import TextToVideoZeroPipeline
296
"""simple docstring""" from __future__ import annotations class __SCREAMING_SNAKE_CASE : def __init__( self : Dict , snake_case : int ): '''simple docstring''' A__ : List[Any] = order # a_{0} ... a_{k} A__ : List[Any] = [1.0] + [0.0] * order # b_{0} ... b_{k} A__ : str = [1.0] + [0.0] * order # x[n-1] ... x[n-k] A__ : Union[str, Any] = [0.0] * self.order # y[n-1] ... y[n-k] A__ : List[str] = [0.0] * self.order def _UpperCamelCase ( self : Optional[int] , snake_case : list[float] , snake_case : list[float] ): '''simple docstring''' if len(snake_case ) < self.order: A__ : Any = [1.0, *a_coeffs] if len(snake_case ) != self.order + 1: A__ : str = ( F'Expected a_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) if len(snake_case ) != self.order + 1: A__ : Union[str, Any] = ( F'Expected b_coeffs to have {self.order + 1} elements ' F'for {self.order}-order filter, got {len(snake_case )}' ) raise ValueError(snake_case ) A__ : Dict = a_coeffs A__ : Any = b_coeffs def _UpperCamelCase ( self : List[str] , snake_case : float ): '''simple docstring''' A__ : str = 0.0 # Start at index 1 and do index 0 at the end. for i in range(1 , self.order + 1 ): result += ( self.b_coeffs[i] * self.input_history[i - 1] - self.a_coeffs[i] * self.output_history[i - 1] ) A__ : Dict = (result + self.b_coeffs[0] * sample) / self.a_coeffs[0] A__ : Tuple = self.input_history[:-1] A__ : int = self.output_history[:-1] A__ : Dict = sample A__ : Tuple = result return result
296
1
"""simple docstring""" import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging A_ = logging.get_logger(__name__) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = ['input_features'] def __init__( self : int , snake_case : str=80 , snake_case : List[str]=1_6000 , snake_case : Union[str, Any]=160 , snake_case : int=30 , snake_case : Optional[Any]=400 , snake_case : Tuple=0.0 , snake_case : str=False , **snake_case : Union[str, Any] , ): '''simple docstring''' super().__init__( feature_size=snake_case , sampling_rate=snake_case , padding_value=snake_case , return_attention_mask=snake_case , **snake_case , ) A__ : Any = n_fft A__ : Dict = hop_length A__ : Optional[Any] = chunk_length A__ : Dict = chunk_length * sampling_rate A__ : int = self.n_samples // hop_length A__ : Optional[Any] = sampling_rate A__ : Optional[Any] = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=snake_case , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=snake_case , norm="""slaney""" , mel_scale="""slaney""" , ) def _UpperCamelCase ( self : Any , snake_case : np.array ): '''simple docstring''' A__ : Union[str, Any] = spectrogram( snake_case , window_function(self.n_fft , """hann""" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="""log10""" , ) A__ : Union[str, Any] = log_spec[:, :-1] A__ : Any = np.maximum(snake_case , log_spec.max() - 8.0 ) A__ : str = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def _UpperCamelCase ( snake_case : List[np.ndarray] , snake_case : List[np.ndarray] , snake_case : float = 0.0 ): '''simple docstring''' if attention_mask is not None: A__ : Optional[int] = np.array(snake_case , np.intaa ) A__ : str = [] for vector, length in zip(snake_case , attention_mask.sum(-1 ) ): A__ : Dict = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1e-7 ) if length < normed_slice.shape[0]: A__ : Union[str, Any] = padding_value normed_input_values.append(snake_case ) else: A__ : Any = [(x - x.mean()) / np.sqrt(x.var() + 1e-7 ) for x in input_values] return normed_input_values def __call__( self : Any , snake_case : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , snake_case : bool = True , snake_case : Optional[int] = None , snake_case : Optional[Union[str, TensorType]] = None , snake_case : Optional[bool] = None , snake_case : Optional[str] = "max_length" , snake_case : Optional[int] = None , snake_case : Optional[int] = None , snake_case : Optional[bool] = None , **snake_case : Optional[int] , ): '''simple docstring''' if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( """It is strongly recommended to pass the `sampling_rate` argument to this function. """ """Failing to do so can result in silent errors that might be hard to debug.""" ) A__ : Tuple = isinstance(snake_case , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) A__ : Union[str, Any] = is_batched_numpy or ( isinstance(snake_case , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: A__ : Tuple = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(snake_case , np.ndarray ): A__ : str = np.asarray(snake_case , dtype=np.floataa ) elif isinstance(snake_case , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): A__ : Any = raw_speech.astype(np.floataa ) # always return batch if not is_batched: A__ : Union[str, Any] = [np.asarray([raw_speech] ).T] A__ : Optional[Any] = BatchFeature({"""input_features""": raw_speech} ) # convert into correct format for padding A__ : Optional[Any] = self.pad( snake_case , padding=snake_case , max_length=max_length if max_length else self.n_samples , truncation=snake_case , pad_to_multiple_of=snake_case , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: A__ : List[Any] = self.zero_mean_unit_var_norm( padded_inputs["""input_features"""] , attention_mask=padded_inputs["""attention_mask"""] , padding_value=self.padding_value , ) A__ : Optional[int] = np.stack(padded_inputs["""input_features"""] , axis=0 ) # make sure list is in array format A__ : Any = padded_inputs.get("""input_features""" ).transpose(2 , 0 , 1 ) A__ : Union[str, Any] = [self._np_extract_fbank_features(snake_case ) for waveform in input_features[0]] if isinstance(input_features[0] , snake_case ): A__ : int = [np.asarray(snake_case , dtype=np.floataa ) for feature in input_features] else: A__ : int = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) A__ : Any = padded_inputs["""attention_mask"""][:, :: self.hop_length] if return_tensors is not None: A__ : List[Any] = padded_inputs.convert_to_tensors(snake_case ) return padded_inputs def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Optional[int] = copy.deepcopy(self.__dict__ ) A__ : List[str] = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
296
"""simple docstring""" import unittest from parameterized import parameterized from transformers import AutoTokenizer, GPTNeoXConfig, is_torch_available, set_seed from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, GPTNeoXModel, ) class __SCREAMING_SNAKE_CASE : def __init__( self : Optional[int] , snake_case : Optional[Any] , snake_case : Tuple=13 , snake_case : Dict=7 , snake_case : Optional[int]=True , snake_case : Union[str, Any]=True , snake_case : Dict=True , snake_case : Any=True , snake_case : List[str]=99 , snake_case : str=64 , snake_case : Optional[int]=5 , snake_case : str=4 , snake_case : List[Any]=37 , snake_case : Optional[Any]="gelu" , snake_case : List[str]=0.1 , snake_case : str=0.1 , snake_case : Optional[int]=512 , snake_case : Dict=16 , snake_case : List[Any]=2 , snake_case : Optional[int]=0.02 , snake_case : Any=3 , snake_case : Union[str, Any]=4 , snake_case : Dict=None , ): '''simple docstring''' A__ : Tuple = parent A__ : Union[str, Any] = batch_size A__ : List[str] = seq_length A__ : Optional[int] = is_training A__ : Dict = use_input_mask A__ : Any = use_token_type_ids A__ : Optional[Any] = use_labels A__ : List[str] = vocab_size A__ : Optional[int] = hidden_size A__ : Optional[Any] = num_hidden_layers A__ : Any = num_attention_heads A__ : List[Any] = intermediate_size A__ : Optional[Any] = hidden_act A__ : Optional[int] = hidden_dropout_prob A__ : Tuple = attention_probs_dropout_prob A__ : str = max_position_embeddings A__ : List[str] = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[Any] = initializer_range A__ : Optional[int] = num_labels A__ : Dict = num_choices A__ : Dict = scope A__ : List[Any] = vocab_size - 1 def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Any = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : List[Any] = None if self.use_input_mask: A__ : List[Any] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_labels: A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Tuple = self.get_config() return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return GPTNeoXConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , pad_token_id=self.pad_token_id , ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ , A__ , A__ , A__ : str = self.prepare_config_and_inputs() A__ : Union[str, Any] = True return config, input_ids, input_mask, token_labels def _UpperCamelCase ( self : Union[str, Any] , snake_case : Optional[int] , snake_case : List[str] , snake_case : int ): '''simple docstring''' A__ : Any = GPTNeoXModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case ) A__ : Optional[int] = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Union[str, Any] , snake_case : str , snake_case : Any , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = True A__ : str = GPTNeoXModel(snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Dict , snake_case : List[Any] , snake_case : str , snake_case : Optional[Any] , snake_case : Any ): '''simple docstring''' A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Tuple = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : Union[str, Any] , snake_case : Tuple ): '''simple docstring''' A__ : int = self.num_labels A__ : int = GPTNeoXForQuestionAnswering(snake_case ) model.to(snake_case ) model.eval() A__ : Optional[Any] = model(snake_case , attention_mask=snake_case ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def _UpperCamelCase ( self : str , snake_case : Tuple , snake_case : int , snake_case : int , snake_case : Dict ): '''simple docstring''' A__ : List[Any] = self.num_labels A__ : Tuple = GPTNeoXForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Any = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def _UpperCamelCase ( self : Any , snake_case : Union[str, Any] , snake_case : int , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Tuple = self.num_labels A__ : Any = GPTNeoXForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : Dict = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : List[str] , snake_case : List[str] , snake_case : Tuple , snake_case : Any ): '''simple docstring''' A__ : Optional[int] = True A__ : Any = GPTNeoXForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() # first forward pass A__ : Tuple = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ : str = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids A__ : Any = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : Tuple = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and A__ : Any = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Any = torch.cat([input_mask, next_mask] , dim=-1 ) A__ : Tuple = model(snake_case , attention_mask=snake_case , output_hidden_states=snake_case ) A__ : List[Any] = output_from_no_past["""hidden_states"""][0] A__ : List[str] = model( snake_case , attention_mask=snake_case , past_key_values=snake_case , output_hidden_states=snake_case , )["""hidden_states"""][0] # select random slice A__ : Tuple = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[Any] = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : Any = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : str = self.prepare_config_and_inputs() A__ , A__ , A__ , A__ : Dict = config_and_inputs A__ : Optional[Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( ( GPTNeoXModel, GPTNeoXForCausalLM, GPTNeoXForQuestionAnswering, GPTNeoXForSequenceClassification, GPTNeoXForTokenClassification, ) if is_torch_available() else () ) snake_case_ = (GPTNeoXForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': GPTNeoXModel, 'question-answering': GPTNeoXForQuestionAnswering, 'text-classification': GPTNeoXForSequenceClassification, 'text-generation': GPTNeoXForCausalLM, 'token-classification': GPTNeoXForTokenClassification, 'zero-shot': GPTNeoXForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False snake_case_ = False snake_case_ = False snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : Any = GPTNeoXModelTester(self ) A__ : Any = ConfigTester(self , config_class=snake_case , hidden_size=64 , num_attention_heads=8 ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ , A__ , A__ , A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ , A__ , A__ , A__ : List[str] = self.model_tester.prepare_config_and_inputs_for_decoder() A__ : Optional[Any] = None self.model_tester.create_and_check_model_as_decoder(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ , A__ , A__ : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_decoder_model_past_large_inputs(snake_case , snake_case , snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_causal_lm(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case ) @unittest.skip(reason="""Feed forward chunking is not implemented""" ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass @parameterized.expand([("""linear""",), ("""dynamic""",)] ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Optional[Any] ): '''simple docstring''' A__ , A__ : int = self.model_tester.prepare_config_and_inputs_for_common() A__ : List[Any] = ids_tensor([1, 10] , config.vocab_size ) A__ : str = ids_tensor([1, int(config.max_position_embeddings * 1.5 )] , config.vocab_size ) set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Union[str, Any] = GPTNeoXModel(snake_case ) original_model.to(snake_case ) original_model.eval() A__ : Optional[int] = original_model(snake_case ).last_hidden_state A__ : List[str] = original_model(snake_case ).last_hidden_state set_seed(42 ) # Fixed seed at init time so the two models get the same random weights A__ : Optional[int] = {"""type""": scaling_type, """factor""": 10.0} A__ : Optional[int] = GPTNeoXModel(snake_case ) scaled_model.to(snake_case ) scaled_model.eval() A__ : List[str] = scaled_model(snake_case ).last_hidden_state A__ : Tuple = scaled_model(snake_case ).last_hidden_state # Dynamic scaling does not change the RoPE embeddings until it receives an input longer than the original # maximum sequence length, so the outputs for the short input should match. if scaling_type == "dynamic": self.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) else: self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) # The output should be different for long inputs self.assertFalse(torch.allclose(snake_case , snake_case , atol=1e-5 ) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = AutoTokenizer.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) for checkpointing in [True, False]: A__ : Optional[Any] = GPTNeoXForCausalLM.from_pretrained("""EleutherAI/pythia-410m-deduped""" ) if checkpointing: model.gradient_checkpointing_enable() else: model.gradient_checkpointing_disable() model.to(snake_case ) A__ : Optional[Any] = tokenizer("""My favorite food is""" , return_tensors="""pt""" ).to(snake_case ) # The hub repo. is updated on 2023-04-04, resulting in poor outputs. # See: https://github.com/huggingface/transformers/pull/24193 A__ : Union[str, Any] = """My favorite food is a good old-fashioned, old-fashioned, old-fashioned.\n\nI'm not sure""" A__ : Tuple = model.generate(**snake_case , do_sample=snake_case , max_new_tokens=20 ) A__ : Tuple = tokenizer.batch_decode(snake_case )[0] self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import LevitImageProcessor class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def __init__( self : List[str] , snake_case : Optional[Any] , snake_case : str=7 , snake_case : Tuple=3 , snake_case : List[Any]=18 , snake_case : Any=30 , snake_case : Optional[Any]=400 , snake_case : Any=True , snake_case : List[Any]=None , snake_case : Optional[Any]=True , snake_case : int=None , snake_case : int=True , snake_case : List[str]=[0.5, 0.5, 0.5] , snake_case : Optional[int]=[0.5, 0.5, 0.5] , ): '''simple docstring''' A__ : Any = size if size is not None else {"""shortest_edge""": 18} A__ : Dict = crop_size if crop_size is not None else {"""height""": 18, """width""": 18} A__ : Dict = parent A__ : Optional[Any] = batch_size A__ : Tuple = num_channels A__ : List[Any] = image_size A__ : List[Any] = min_resolution A__ : List[str] = max_resolution A__ : Optional[int] = do_resize A__ : Optional[int] = size A__ : Optional[int] = do_center_crop A__ : Optional[int] = crop_size A__ : Tuple = do_normalize A__ : Union[str, Any] = image_mean A__ : Union[str, Any] = image_std def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "do_center_crop": self.do_center_crop, "size": self.size, "crop_size": self.crop_size, } @require_torch @require_vision class __SCREAMING_SNAKE_CASE ( UpperCamelCase , unittest.TestCase ): snake_case_ = LevitImageProcessor if is_vision_available() else None def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : List[str] = LevitImageProcessingTester(self ) @property def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' return self.image_processor_tester.prepare_image_processor_dict() def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : str = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case , """image_mean""" ) ) self.assertTrue(hasattr(snake_case , """image_std""" ) ) self.assertTrue(hasattr(snake_case , """do_normalize""" ) ) self.assertTrue(hasattr(snake_case , """do_resize""" ) ) self.assertTrue(hasattr(snake_case , """do_center_crop""" ) ) self.assertTrue(hasattr(snake_case , """size""" ) ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : Dict = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"""shortest_edge""": 18} ) self.assertEqual(image_processor.crop_size , {"""height""": 18, """width""": 18} ) A__ : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 , crop_size=84 ) self.assertEqual(image_processor.size , {"""shortest_edge""": 42} ) self.assertEqual(image_processor.crop_size , {"""height""": 84, """width""": 84} ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' pass def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.image_processing_class(**self.image_processor_dict ) # create random PIL images A__ : Tuple = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case , Image.Image ) # Test not batched input A__ : List[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Optional[Any] = image_processing(snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors A__ : List[str] = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case , numpify=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case , np.ndarray ) # Test not batched input A__ : Optional[Any] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Union[str, Any] = image_processing(snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : Any = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors A__ : Optional[int] = prepare_image_inputs(self.image_processor_tester , equal_resolution=snake_case , torchify=snake_case ) for image in image_inputs: self.assertIsInstance(snake_case , torch.Tensor ) # Test not batched input A__ : Optional[int] = image_processing(image_inputs[0] , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , ) # Test batched A__ : Dict = image_processing(snake_case , return_tensors="""pt""" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.crop_size["""height"""], self.image_processor_tester.crop_size["""width"""], ) , )
296
"""simple docstring""" from collections import defaultdict from math import gcd def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_5_0_0_0_0_0 ) ->int: A__ : defaultdict = defaultdict(UpperCAmelCase__ ) A__ : Any = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, UpperCAmelCase__, 2 ): if gcd(UpperCAmelCase__, UpperCAmelCase__ ) > 1: continue A__ : str = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(F'{solution() = }')
296
1
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTConfig, MobileViTForImageClassification, MobileViTForSemanticSegmentation, MobileViTImageProcessor, ) from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->Optional[int]: A__ : str = MobileViTConfig() # size of the architecture if "mobilevit_s" in mobilevit_name: A__ : Union[str, Any] = [1_4_4, 1_9_2, 2_4_0] A__ : str = [1_6, 3_2, 6_4, 9_6, 1_2_8, 1_6_0, 6_4_0] elif "mobilevit_xs" in mobilevit_name: A__ : Optional[Any] = [9_6, 1_2_0, 1_4_4] A__ : List[str] = [1_6, 3_2, 4_8, 6_4, 8_0, 9_6, 3_8_4] elif "mobilevit_xxs" in mobilevit_name: A__ : Optional[int] = [6_4, 8_0, 9_6] A__ : Optional[int] = [1_6, 1_6, 2_4, 4_8, 6_4, 8_0, 3_2_0] A__ : Optional[Any] = 0.05 A__ : Optional[int] = 2.0 if mobilevit_name.startswith("""deeplabv3_""" ): A__ : List[Any] = 5_1_2 A__ : Optional[Any] = 1_6 A__ : Union[str, Any] = 2_1 A__ : Tuple = """pascal-voc-id2label.json""" else: A__ : Union[str, Any] = 1_0_0_0 A__ : Optional[int] = """imagenet-1k-id2label.json""" A__ : Optional[Any] = """huggingface/label-files""" A__ : Any = json.load(open(hf_hub_download(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ), """r""" ) ) A__ : Optional[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Tuple = idalabel A__ : Dict = {v: k for k, v in idalabel.items()} return config def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any], UpperCAmelCase__ : str=False ) ->Optional[int]: for i in range(1, 6 ): if f'layer_{i}.' in name: A__ : Any = name.replace(f'layer_{i}.', f'encoder.layer.{i - 1}.' ) if "conv_1." in name: A__ : Union[str, Any] = name.replace("""conv_1.""", """conv_stem.""" ) if ".block." in name: A__ : Dict = name.replace(""".block.""", """.""" ) if "exp_1x1" in name: A__ : Optional[int] = name.replace("""exp_1x1""", """expand_1x1""" ) if "red_1x1" in name: A__ : List[Any] = name.replace("""red_1x1""", """reduce_1x1""" ) if ".local_rep.conv_3x3." in name: A__ : Any = name.replace(""".local_rep.conv_3x3.""", """.conv_kxk.""" ) if ".local_rep.conv_1x1." in name: A__ : List[str] = name.replace(""".local_rep.conv_1x1.""", """.conv_1x1.""" ) if ".norm." in name: A__ : Tuple = name.replace(""".norm.""", """.normalization.""" ) if ".conv." in name: A__ : int = name.replace(""".conv.""", """.convolution.""" ) if ".conv_proj." in name: A__ : int = name.replace(""".conv_proj.""", """.conv_projection.""" ) for i in range(0, 2 ): for j in range(0, 4 ): if f'.{i}.{j}.' in name: A__ : Tuple = name.replace(f'.{i}.{j}.', f'.{i}.layer.{j}.' ) for i in range(2, 6 ): for j in range(0, 4 ): if f'.{i}.{j}.' in name: A__ : str = name.replace(f'.{i}.{j}.', f'.{i}.' ) if "expand_1x1" in name: A__ : Optional[int] = name.replace("""expand_1x1""", """downsampling_layer.expand_1x1""" ) if "conv_3x3" in name: A__ : Optional[int] = name.replace("""conv_3x3""", """downsampling_layer.conv_3x3""" ) if "reduce_1x1" in name: A__ : Dict = name.replace("""reduce_1x1""", """downsampling_layer.reduce_1x1""" ) for i in range(2, 5 ): if f'.global_rep.{i}.weight' in name: A__ : int = name.replace(f'.global_rep.{i}.weight', """.layernorm.weight""" ) if f'.global_rep.{i}.bias' in name: A__ : Optional[Any] = name.replace(f'.global_rep.{i}.bias', """.layernorm.bias""" ) if ".global_rep." in name: A__ : Any = name.replace(""".global_rep.""", """.transformer.""" ) if ".pre_norm_mha.0." in name: A__ : str = name.replace(""".pre_norm_mha.0.""", """.layernorm_before.""" ) if ".pre_norm_mha.1.out_proj." in name: A__ : List[str] = name.replace(""".pre_norm_mha.1.out_proj.""", """.attention.output.dense.""" ) if ".pre_norm_ffn.0." in name: A__ : List[Any] = name.replace(""".pre_norm_ffn.0.""", """.layernorm_after.""" ) if ".pre_norm_ffn.1." in name: A__ : Union[str, Any] = name.replace(""".pre_norm_ffn.1.""", """.intermediate.dense.""" ) if ".pre_norm_ffn.4." in name: A__ : List[Any] = name.replace(""".pre_norm_ffn.4.""", """.output.dense.""" ) if ".transformer." in name: A__ : List[str] = name.replace(""".transformer.""", """.transformer.layer.""" ) if ".aspp_layer." in name: A__ : Tuple = name.replace(""".aspp_layer.""", """.""" ) if ".aspp_pool." in name: A__ : Tuple = name.replace(""".aspp_pool.""", """.""" ) if "seg_head." in name: A__ : str = name.replace("""seg_head.""", """segmentation_head.""" ) if "segmentation_head.classifier.classifier." in name: A__ : int = name.replace("""segmentation_head.classifier.classifier.""", """segmentation_head.classifier.""" ) if "classifier.fc." in name: A__ : List[str] = name.replace("""classifier.fc.""", """classifier.""" ) elif (not base_model) and ("segmentation_head." not in name): A__ : int = """mobilevit.""" + name return name def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Union[str, Any]=False ) ->Tuple: if base_model: A__ : Union[str, Any] = """""" else: A__ : Dict = """mobilevit.""" for key in orig_state_dict.copy().keys(): A__ : Any = orig_state_dict.pop(UpperCAmelCase__ ) if key[:8] == "encoder.": A__ : int = key[8:] if "qkv" in key: A__ : List[Any] = key.split(""".""" ) A__ : Tuple = int(key_split[0][6:] ) - 1 A__ : Tuple = int(key_split[3] ) A__ : Optional[Any] = model.get_submodule(f'{model_prefix}encoder.layer.{layer_num}' ) A__ : int = layer.transformer.layer[transformer_num].attention.attention.all_head_size A__ : Optional[Any] = ( f'{model_prefix}encoder.layer.{layer_num}.transformer.layer.{transformer_num}.attention.attention.' ) if "weight" in key: A__ : int = val[:dim, :] A__ : str = val[dim : dim * 2, :] A__ : Optional[Any] = val[-dim:, :] else: A__ : Any = val[:dim] A__ : Optional[Any] = val[dim : dim * 2] A__ : Dict = val[-dim:] else: A__ : List[str] = val return orig_state_dict def _lowerCAmelCase ( ) ->Tuple: A__ : Optional[int] = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : Union[str, Any] = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Any=False ) ->Optional[int]: A__ : Tuple = get_mobilevit_config(UpperCAmelCase__ ) # load original state_dict A__ : int = torch.load(UpperCAmelCase__, map_location="""cpu""" ) # load 🤗 model if mobilevit_name.startswith("""deeplabv3_""" ): A__ : int = MobileViTForSemanticSegmentation(UpperCAmelCase__ ).eval() else: A__ : Tuple = MobileViTForImageClassification(UpperCAmelCase__ ).eval() A__ : Optional[Any] = convert_state_dict(UpperCAmelCase__, UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) # Check outputs on an image, prepared by MobileViTImageProcessor A__ : str = MobileViTImageProcessor(crop_size=config.image_size, size=config.image_size + 3_2 ) A__ : Dict = image_processor(images=prepare_img(), return_tensors="""pt""" ) A__ : List[str] = model(**UpperCAmelCase__ ) A__ : str = outputs.logits if mobilevit_name.startswith("""deeplabv3_""" ): assert logits.shape == (1, 2_1, 3_2, 3_2) if mobilevit_name == "deeplabv3_mobilevit_s": A__ : Any = torch.tensor( [ [[6.2065, 6.1292, 6.2070], [6.1079, 6.1254, 6.1747], [6.0042, 6.1071, 6.1034]], [[-6.9253, -6.8653, -7.0398], [-7.3218, -7.3983, -7.3670], [-7.1961, -7.2482, -7.1569]], [[-4.4723, -4.4348, -4.3769], [-5.3629, -5.4632, -5.4598], [-5.1587, -5.3402, -5.5059]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xs": A__ : Optional[Any] = torch.tensor( [ [[5.4449, 5.5733, 5.6314], [5.1815, 5.3930, 5.5963], [5.1656, 5.4333, 5.4853]], [[-9.4423, -9.7766, -9.6714], [-9.1581, -9.5720, -9.5519], [-9.1006, -9.6458, -9.5703]], [[-7.7721, -7.3716, -7.1583], [-8.4599, -8.0624, -7.7944], [-8.4172, -7.8366, -7.5025]], ] ) elif mobilevit_name == "deeplabv3_mobilevit_xxs": A__ : Optional[Any] = torch.tensor( [ [[6.9811, 6.9743, 7.3123], [7.1777, 7.1931, 7.3938], [7.5633, 7.8050, 7.8901]], [[-10.5536, -10.2332, -10.2924], [-10.2336, -9.8624, -9.5964], [-10.8840, -10.8158, -10.6659]], [[-3.4938, -3.0631, -2.8620], [-3.4205, -2.8135, -2.6875], [-3.4179, -2.7945, -2.8750]], ] ) else: raise ValueError(f'Unknown mobilevit_name: {mobilevit_name}' ) assert torch.allclose(logits[0, :3, :3, :3], UpperCAmelCase__, atol=1e-4 ) else: assert logits.shape == (1, 1_0_0_0) if mobilevit_name == "mobilevit_s": A__ : List[str] = torch.tensor([-0.9866, 0.2392, -1.1241] ) elif mobilevit_name == "mobilevit_xs": A__ : Dict = torch.tensor([-2.4761, -0.9399, -1.9587] ) elif mobilevit_name == "mobilevit_xxs": A__ : Optional[Any] = torch.tensor([-1.9364, -1.2327, -0.4653] ) else: raise ValueError(f'Unknown mobilevit_name: {mobilevit_name}' ) assert torch.allclose(logits[0, :3], UpperCAmelCase__, atol=1e-4 ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model {mobilevit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: A__ : Optional[Any] = { """mobilevit_s""": """mobilevit-small""", """mobilevit_xs""": """mobilevit-x-small""", """mobilevit_xxs""": """mobilevit-xx-small""", """deeplabv3_mobilevit_s""": """deeplabv3-mobilevit-small""", """deeplabv3_mobilevit_xs""": """deeplabv3-mobilevit-x-small""", """deeplabv3_mobilevit_xxs""": """deeplabv3-mobilevit-xx-small""", } print("""Pushing to the hub...""" ) A__ : List[str] = model_mapping[mobilevit_name] image_processor.push_to_hub(UpperCAmelCase__, organization="""apple""" ) model.push_to_hub(UpperCAmelCase__, organization="""apple""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--mobilevit_name''', default='''mobilevit_s''', type=str, help=( '''Name of the MobileViT model you\'d like to convert. Should be one of \'mobilevit_s\', \'mobilevit_xs\',''' ''' \'mobilevit_xxs\', \'deeplabv3_mobilevit_s\', \'deeplabv3_mobilevit_xs\', \'deeplabv3_mobilevit_xxs\'.''' ), ) parser.add_argument( '''--checkpoint_path''', required=True, type=str, help='''Path to the original state dict (.pt file).''' ) parser.add_argument( '''--pytorch_dump_folder_path''', required=True, type=str, help='''Path to the output PyTorch model directory.''' ) parser.add_argument( '''--push_to_hub''', action='''store_true''', help='''Whether or not to push the converted model to the 🤗 hub.''' ) A_ = parser.parse_args() convert_movilevit_checkpoint( args.mobilevit_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub )
296
"""simple docstring""" import os from distutils.util import strtobool def _lowerCAmelCase ( UpperCAmelCase__ : List[Any], UpperCAmelCase__ : Optional[Any] ) ->List[str]: for e in env_keys: A__ : List[Any] = int(os.environ.get(UpperCAmelCase__, -1 ) ) if val >= 0: return val return default def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : str=False ) ->List[str]: A__ : List[Any] = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return strtobool(UpperCAmelCase__ ) == 1 # As its name indicates `strtobool` actually returns an int... def _lowerCAmelCase ( UpperCAmelCase__ : Tuple, UpperCAmelCase__ : List[Any]="no" ) ->int: A__ : str = os.environ.get(UpperCAmelCase__, str(UpperCAmelCase__ ) ) return value
296
1
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
"""simple docstring""" import cva import numpy as np class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : float , snake_case : int ): '''simple docstring''' if k in (0.04, 0.06): A__ : Optional[int] = k A__ : int = window_size else: raise ValueError("""invalid k value""" ) def __str__( self : List[Any] ): '''simple docstring''' return str(self.k ) def _UpperCamelCase ( self : int , snake_case : str ): '''simple docstring''' A__ : List[str] = cva.imread(snake_case , 0 ) A__ , A__ : Union[str, Any] = img.shape A__ : list[list[int]] = [] A__ : Optional[Any] = img.copy() A__ : List[str] = cva.cvtColor(snake_case , cva.COLOR_GRAY2RGB ) A__ , A__ : List[Any] = np.gradient(snake_case ) A__ : List[Any] = dx**2 A__ : Any = dy**2 A__ : Dict = dx * dy A__ : Any = 0.04 A__ : Optional[Any] = self.window_size // 2 for y in range(snake_case , h - offset ): for x in range(snake_case , w - offset ): A__ : List[str] = ixx[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Tuple = iyy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : Optional[int] = ixy[ y - offset : y + offset + 1, x - offset : x + offset + 1 ].sum() A__ : int = (wxx * wyy) - (wxy**2) A__ : Any = wxx + wyy A__ : List[str] = det - k * (trace**2) # Can change the value if r > 0.5: corner_list.append([x, y, r] ) color_img.itemset((y, x, 0) , 0 ) color_img.itemset((y, x, 1) , 0 ) color_img.itemset((y, x, 2) , 255 ) return color_img, corner_list if __name__ == "__main__": A_ = HarrisCorner(0.04, 3) A_ , A_ = edge_detect.detect('''path_to_image''') cva.imwrite('''detect.png''', color_img)
296
1
"""simple docstring""" from math import asin, atan, cos, radians, sin, sqrt, tan A_ = 6378137.0 A_ = 6356752.314245 A_ = 637_8137 def _lowerCAmelCase ( UpperCAmelCase__ : float, UpperCAmelCase__ : float, UpperCAmelCase__ : float, UpperCAmelCase__ : float ) ->float: A__ : Dict = (AXIS_A - AXIS_B) / AXIS_A A__ : Tuple = atan((1 - flattening) * tan(radians(UpperCAmelCase__ ) ) ) A__ : Optional[int] = atan((1 - flattening) * tan(radians(UpperCAmelCase__ ) ) ) A__ : Optional[int] = radians(UpperCAmelCase__ ) A__ : Tuple = radians(UpperCAmelCase__ ) # Equation A__ : Optional[Any] = sin((phi_a - phi_a) / 2 ) A__ : Optional[Any] = sin((lambda_a - lambda_a) / 2 ) # Square both values sin_sq_phi *= sin_sq_phi sin_sq_lambda *= sin_sq_lambda A__ : Optional[int] = sqrt(sin_sq_phi + (cos(UpperCAmelCase__ ) * cos(UpperCAmelCase__ ) * sin_sq_lambda) ) return 2 * RADIUS * asin(UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" from typing import Any, Dict, List, Union from ..utils import add_end_docstrings, is_torch_available, is_vision_available, logging, requires_backends from .base import PIPELINE_INIT_ARGS, Pipeline if is_vision_available(): from ..image_utils import load_image if is_torch_available(): import torch from ..models.auto.modeling_auto import MODEL_FOR_OBJECT_DETECTION_MAPPING, MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING A_ = logging.get_logger(__name__) A_ = Dict[str, Any] A_ = List[Prediction] @add_end_docstrings(UpperCamelCase ) class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): def __init__( self : str , *snake_case : Tuple , **snake_case : Tuple ): '''simple docstring''' super().__init__(*snake_case , **snake_case ) if self.framework == "tf": raise ValueError(F'The {self.__class__} is only available in PyTorch.' ) requires_backends(self , """vision""" ) self.check_model_type( dict(MODEL_FOR_OBJECT_DETECTION_MAPPING.items() + MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING.items() ) ) def _UpperCamelCase ( self : List[Any] , **snake_case : Optional[int] ): '''simple docstring''' A__ : Dict = {} if "threshold" in kwargs: A__ : int = kwargs["""threshold"""] return {}, {}, postprocess_kwargs def __call__( self : Tuple , *snake_case : Union[str, Any] , **snake_case : Union[str, Any] ): '''simple docstring''' return super().__call__(*snake_case , **snake_case ) def _UpperCamelCase ( self : str , snake_case : int ): '''simple docstring''' A__ : List[str] = load_image(snake_case ) A__ : int = torch.IntTensor([[image.height, image.width]] ) A__ : Union[str, Any] = self.image_processor(images=[image] , return_tensors="""pt""" ) if self.tokenizer is not None: A__ : str = self.tokenizer(text=inputs["""words"""] , boxes=inputs["""boxes"""] , return_tensors="""pt""" ) A__ : List[str] = target_size return inputs def _UpperCamelCase ( self : Optional[int] , snake_case : List[Any] ): '''simple docstring''' A__ : str = model_inputs.pop("""target_size""" ) A__ : Dict = self.model(**snake_case ) A__ : Optional[Any] = outputs.__class__({"""target_size""": target_size, **outputs} ) if self.tokenizer is not None: A__ : str = model_inputs["""bbox"""] return model_outputs def _UpperCamelCase ( self : Tuple , snake_case : Optional[int] , snake_case : int=0.9 ): '''simple docstring''' A__ : Any = model_outputs["""target_size"""] if self.tokenizer is not None: # This is a LayoutLMForTokenClassification variant. # The OCR got the boxes and the model classified the words. A__ , A__ : Tuple = target_size[0].tolist() def unnormalize(snake_case : Optional[int] ): return self._get_bounding_box( torch.Tensor( [ (width * bbox[0] / 1000), (height * bbox[1] / 1000), (width * bbox[2] / 1000), (height * bbox[3] / 1000), ] ) ) A__ , A__ : Optional[int] = model_outputs["""logits"""].squeeze(0 ).softmax(dim=-1 ).max(dim=-1 ) A__ : Optional[Any] = [self.model.config.idalabel[prediction] for prediction in classes.tolist()] A__ : List[str] = [unnormalize(snake_case ) for bbox in model_outputs["""bbox"""].squeeze(0 )] A__ : Tuple = ["""score""", """label""", """box"""] A__ : Any = [dict(zip(snake_case , snake_case ) ) for vals in zip(scores.tolist() , snake_case , snake_case ) if vals[0] > threshold] else: # This is a regular ForObjectDetectionModel A__ : Union[str, Any] = self.image_processor.post_process_object_detection(snake_case , snake_case , snake_case ) A__ : str = raw_annotations[0] A__ : str = raw_annotation["""scores"""] A__ : List[Any] = raw_annotation["""labels"""] A__ : int = raw_annotation["""boxes"""] A__ : str = scores.tolist() A__ : Any = [self.model.config.idalabel[label.item()] for label in labels] A__ : int = [self._get_bounding_box(snake_case ) for box in boxes] # {"scores": [...], ...} --> [{"score":x, ...}, ...] A__ : str = ["""score""", """label""", """box"""] A__ : Dict = [ dict(zip(snake_case , snake_case ) ) for vals in zip(raw_annotation["""scores"""] , raw_annotation["""labels"""] , raw_annotation["""boxes"""] ) ] return annotation def _UpperCamelCase ( self : Union[str, Any] , snake_case : "torch.Tensor" ): '''simple docstring''' if self.framework != "pt": raise ValueError("""The ObjectDetectionPipeline is only available in PyTorch.""" ) A__ , A__ , A__ , A__ : Any = box.int().tolist() A__ : Any = { """xmin""": xmin, """ymin""": ymin, """xmax""": xmax, """ymax""": ymax, } return bbox
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : list ) ->bool: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(UpperCAmelCase__ ) == 0: raise ValueError("""Input list must be a non empty list""" ) if len(UpperCAmelCase__ ) == 1: return True A__ : Optional[Any] = series[1] - series[0] for index in range(len(UpperCAmelCase__ ) - 1 ): if series[index + 1] - series[index] != common_diff: return False return True def _lowerCAmelCase ( UpperCAmelCase__ : list ) ->float: if not isinstance(UpperCAmelCase__, UpperCAmelCase__ ): raise ValueError("""Input series is not valid, valid series - [2, 4, 6]""" ) if len(UpperCAmelCase__ ) == 0: raise ValueError("""Input list must be a non empty list""" ) A__ : Optional[int] = 0 for val in series: answer += val return answer / len(UpperCAmelCase__ ) if __name__ == "__main__": import doctest doctest.testmod()
296
"""simple docstring""" from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging from ..auto import CONFIG_MAPPING A_ = logging.get_logger(__name__) A_ = { '''microsoft/table-transformer-detection''': ( '''https://huggingface.co/microsoft/table-transformer-detection/resolve/main/config.json''' ), } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'table-transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'hidden_size': 'd_model', 'num_attention_heads': 'encoder_attention_heads', } def __init__( self : Dict , snake_case : int=True , snake_case : Dict=None , snake_case : Union[str, Any]=3 , snake_case : Dict=100 , snake_case : Tuple=6 , snake_case : Optional[int]=2048 , snake_case : int=8 , snake_case : Dict=6 , snake_case : Any=2048 , snake_case : str=8 , snake_case : Union[str, Any]=0.0 , snake_case : List[str]=0.0 , snake_case : List[str]=True , snake_case : Any="relu" , snake_case : str=256 , snake_case : int=0.1 , snake_case : Dict=0.0 , snake_case : str=0.0 , snake_case : Union[str, Any]=0.02 , snake_case : Union[str, Any]=1.0 , snake_case : Optional[Any]=False , snake_case : int="sine" , snake_case : Optional[Any]="resnet50" , snake_case : Optional[int]=True , snake_case : Any=False , snake_case : int=1 , snake_case : Tuple=5 , snake_case : Optional[int]=2 , snake_case : Tuple=1 , snake_case : Optional[Any]=1 , snake_case : Optional[Any]=5 , snake_case : Dict=2 , snake_case : Any=0.1 , **snake_case : Any , ): '''simple docstring''' if backbone_config is not None and use_timm_backbone: raise ValueError("""You can't specify both `backbone_config` and `use_timm_backbone`.""" ) if not use_timm_backbone: if backbone_config is None: logger.info("""`backbone_config` is `None`. Initializing the config with the default `ResNet` backbone.""" ) A__ : Optional[Any] = CONFIG_MAPPING["""resnet"""](out_features=["""stage4"""] ) elif isinstance(snake_case , snake_case ): A__ : Optional[int] = backbone_config.get("""model_type""" ) A__ : Optional[int] = CONFIG_MAPPING[backbone_model_type] A__ : List[str] = config_class.from_dict(snake_case ) # set timm attributes to None A__ , A__ , A__ : str = None, None, None A__ : Tuple = use_timm_backbone A__ : str = backbone_config A__ : str = num_channels A__ : List[Any] = num_queries A__ : Optional[Any] = d_model A__ : Tuple = encoder_ffn_dim A__ : Union[str, Any] = encoder_layers A__ : List[Any] = encoder_attention_heads A__ : Optional[int] = decoder_ffn_dim A__ : Any = decoder_layers A__ : int = decoder_attention_heads A__ : Any = dropout A__ : Dict = attention_dropout A__ : Dict = activation_dropout A__ : Tuple = activation_function A__ : List[str] = init_std A__ : List[str] = init_xavier_std A__ : Any = encoder_layerdrop A__ : Optional[Any] = decoder_layerdrop A__ : Union[str, Any] = encoder_layers A__ : Dict = auxiliary_loss A__ : List[Any] = position_embedding_type A__ : Optional[Any] = backbone A__ : str = use_pretrained_backbone A__ : Union[str, Any] = dilation # Hungarian matcher A__ : Tuple = class_cost A__ : Optional[Any] = bbox_cost A__ : Dict = giou_cost # Loss coefficients A__ : Any = mask_loss_coefficient A__ : str = dice_loss_coefficient A__ : str = bbox_loss_coefficient A__ : Union[str, Any] = giou_loss_coefficient A__ : List[str] = eos_coefficient super().__init__(is_encoder_decoder=snake_case , **snake_case ) @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return self.encoder_attention_heads @property def _UpperCamelCase ( self : Dict ): '''simple docstring''' return self.d_model class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = version.parse('1.11' ) @property def _UpperCamelCase ( self : Any ): '''simple docstring''' return OrderedDict( [ ("""pixel_values""", {0: """batch""", 1: """num_channels""", 2: """height""", 3: """width"""}), ("""pixel_mask""", {0: """batch"""}), ] ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return 1e-5 @property def _UpperCamelCase ( self : List[str] ): '''simple docstring''' return 12
296
1
"""simple docstring""" import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): snake_case_ = JukeboxTokenizer snake_case_ = { 'artist': 'Zac Brown Band', 'genres': 'Country', 'lyrics': 'I met a traveller from an antique land,\n Who said "Two vast and trunkless legs of stone\n Stand in the desert. . . . Near them, on the sand,\n Half sunk a shattered visage lies, whose frown,\n And wrinkled lip, and sneer of cold command,\n Tell that its sculptor well those passions read\n Which yet survive, stamped on these lifeless things,\n The hand that mocked them, and the heart that fed;\n And on the pedestal, these words appear:\n My name is Ozymandias, King of Kings;\n Look on my Works, ye Mighty, and despair!\n Nothing beside remains. Round the decay\n Of that colossal Wreck, boundless and bare\n The lone and level sands stretch far away\n ', } @require_torch def _UpperCamelCase ( self : Tuple ): '''simple docstring''' import torch A__ : Dict = JukeboxTokenizer.from_pretrained("""openai/jukebox-1b-lyrics""" ) A__ : Any = tokenizer(**self.metas )["""input_ids"""] # fmt: off A__ : List[Any] = [ torch.tensor([[ 0, 0, 0, 7169, 507, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), torch.tensor([[0, 0, 0, 1069, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' import torch A__ : Optional[int] = JukeboxTokenizer.from_pretrained("""openai/jukebox-5b-lyrics""" ) A__ : Tuple = tokenizer(**self.metas )["""input_ids"""] # fmt: off A__ : List[Any] = [ torch.tensor([[ 0, 0, 0, 1069, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 1069, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
296
"""simple docstring""" # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING from ..models.auto import AutoModelForVisionaSeq from ..utils import requires_backends from .base import PipelineTool if TYPE_CHECKING: from PIL import Image class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'Salesforce/blip-image-captioning-base' snake_case_ = ( 'This is a tool that generates a description of an image. It takes an input named `image` which should be the ' 'image to caption, and returns a text that contains the description in English.' ) snake_case_ = 'image_captioner' snake_case_ = AutoModelForVisionaSeq snake_case_ = ['image'] snake_case_ = ['text'] def __init__( self : int , *snake_case : Optional[int] , **snake_case : Optional[int] ): '''simple docstring''' requires_backends(self , ["""vision"""] ) super().__init__(*snake_case , **snake_case ) def _UpperCamelCase ( self : int , snake_case : "Image" ): '''simple docstring''' return self.pre_processor(images=snake_case , return_tensors="""pt""" ) def _UpperCamelCase ( self : int , snake_case : List[Any] ): '''simple docstring''' return self.model.generate(**snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' return self.pre_processor.batch_decode(snake_case , skip_special_tokens=snake_case )[0].strip()
296
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices A_ = logging.get_logger(__name__) A_ = { '''shi-labs/nat-mini-in1k-224''': '''https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json''', # See all Nat models at https://huggingface.co/models?filter=nat } class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase ): snake_case_ = 'nat' snake_case_ = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__( self : Union[str, Any] , snake_case : Any=4 , snake_case : Any=3 , snake_case : Optional[Any]=64 , snake_case : Any=[3, 4, 6, 5] , snake_case : List[str]=[2, 4, 8, 16] , snake_case : str=7 , snake_case : List[Any]=3.0 , snake_case : Optional[Any]=True , snake_case : Any=0.0 , snake_case : int=0.0 , snake_case : int=0.1 , snake_case : int="gelu" , snake_case : Optional[Any]=0.02 , snake_case : List[str]=1e-5 , snake_case : Tuple=0.0 , snake_case : Dict=None , snake_case : Tuple=None , **snake_case : Union[str, Any] , ): '''simple docstring''' super().__init__(**snake_case ) A__ : Optional[int] = patch_size A__ : Optional[int] = num_channels A__ : Dict = embed_dim A__ : str = depths A__ : Any = len(snake_case ) A__ : Tuple = num_heads A__ : Tuple = kernel_size A__ : int = mlp_ratio A__ : int = qkv_bias A__ : Any = hidden_dropout_prob A__ : List[str] = attention_probs_dropout_prob A__ : Any = drop_path_rate A__ : str = hidden_act A__ : Optional[int] = layer_norm_eps A__ : Dict = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model A__ : Optional[Any] = int(embed_dim * 2 ** (len(snake_case ) - 1) ) A__ : Tuple = layer_scale_init_value A__ : Optional[int] = ["""stem"""] + [F'stage{idx}' for idx in range(1 , len(snake_case ) + 1 )] A__ , A__ : Dict = get_aligned_output_features_output_indices( out_features=snake_case , out_indices=snake_case , stage_names=self.stage_names )
296
"""simple docstring""" import os import unittest from tempfile import TemporaryDirectory import torch import torch.nn as nn from accelerate.utils import ( OffloadedWeightsLoader, extract_submodules_state_dict, load_offloaded_weight, offload_state_dict, offload_weight, ) class __SCREAMING_SNAKE_CASE ( nn.Module ): def __init__( self : List[Any] ): '''simple docstring''' super().__init__() A__ : int = nn.Linear(3 , 4 ) A__ : Union[str, Any] = nn.BatchNormad(4 ) A__ : Union[str, Any] = nn.Linear(4 , 5 ) def _UpperCamelCase ( self : str , snake_case : List[str] ): '''simple docstring''' return self.lineara(self.batchnorm(self.lineara(snake_case ) ) ) class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : int = ModelForTest() with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , model.state_dict() ) A__ : List[str] = os.path.join(snake_case , """index.json""" ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on what is inside the index for key in ["linear1.weight", "linear1.bias", "linear2.weight", "linear2.bias"]: A__ : List[str] = os.path.join(snake_case , F'{key}.dat' ) self.assertTrue(os.path.isfile(snake_case ) ) # TODO: add tests on the fact weights are properly loaded def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Union[str, Any] = [torch.floataa, torch.floataa, torch.bfloataa] for dtype in dtypes: A__ : str = torch.randn(2 , 3 , dtype=snake_case ) with TemporaryDirectory() as tmp_dir: A__ : List[str] = offload_weight(snake_case , """weight""" , snake_case , {} ) A__ : Union[str, Any] = os.path.join(snake_case , """weight.dat""" ) self.assertTrue(os.path.isfile(snake_case ) ) self.assertDictEqual(snake_case , {"""weight""": {"""shape""": [2, 3], """dtype""": str(snake_case ).split(""".""" )[1]}} ) A__ : str = load_offloaded_weight(snake_case , index["""weight"""] ) self.assertTrue(torch.equal(snake_case , snake_case ) ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : str = ModelForTest() A__ : Union[str, Any] = model.state_dict() A__ : Optional[int] = {k: v for k, v in state_dict.items() if """linear2""" not in k} A__ : List[Any] = {k: v for k, v in state_dict.items() if """linear2""" in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Dict = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) A__ : int = {k: v for k, v in state_dict.items() if """weight""" in k} A__ : Tuple = {k: v for k, v in state_dict.items() if """weight""" not in k} with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) A__ : Optional[Any] = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) with TemporaryDirectory() as tmp_dir: offload_state_dict(snake_case , snake_case ) # Duplicates are removed A__ : int = OffloadedWeightsLoader(state_dict=snake_case , save_folder=snake_case ) # Every key is there with the right value self.assertEqual(sorted(snake_case ) , sorted(state_dict.keys() ) ) for key, param in state_dict.items(): self.assertTrue(torch.allclose(snake_case , weight_map[key] ) ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' A__ : List[str] = {"""a.1""": 0, """a.10""": 1, """a.2""": 2} A__ : str = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1""": 0, """a.2""": 2} ) A__ : Dict = {"""a.1.a""": 0, """a.10.a""": 1, """a.2.a""": 2} A__ : int = extract_submodules_state_dict(snake_case , ["""a.1""", """a.2"""] ) self.assertDictEqual(snake_case , {"""a.1.a""": 0, """a.2.a""": 2} )
296
1
"""simple docstring""" def _lowerCAmelCase ( UpperCAmelCase__ : int = 1_0_0_0_0_0_0 ) ->int: A__ : int = set(range(3, UpperCAmelCase__, 2 ) ) primes.add(2 ) for p in range(3, UpperCAmelCase__, 2 ): if p not in primes: continue primes.difference_update(set(range(p * p, UpperCAmelCase__, UpperCAmelCase__ ) ) ) A__ : str = [float(UpperCAmelCase__ ) for n in range(limit + 1 )] for p in primes: for n in range(UpperCAmelCase__, limit + 1, UpperCAmelCase__ ): phi[n] *= 1 - 1 / p return int(sum(phi[2:] ) ) if __name__ == "__main__": print(F'{solution() = }')
296
"""simple docstring""" import math import unittest from transformers import BioGptConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptTokenizer, ) from transformers.models.biogpt.modeling_biogpt import BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST class __SCREAMING_SNAKE_CASE : def __init__( self : Union[str, Any] , snake_case : str , snake_case : List[Any]=13 , snake_case : Union[str, Any]=7 , snake_case : Optional[Any]=True , snake_case : str=True , snake_case : Dict=False , snake_case : Union[str, Any]=True , snake_case : Optional[Any]=99 , snake_case : str=32 , snake_case : Tuple=5 , snake_case : List[str]=4 , snake_case : Optional[int]=37 , snake_case : str="gelu" , snake_case : Tuple=0.1 , snake_case : Optional[int]=0.1 , snake_case : int=512 , snake_case : List[str]=16 , snake_case : str=2 , snake_case : Optional[int]=0.02 , snake_case : str=3 , snake_case : Dict=4 , snake_case : Optional[Any]=None , ): '''simple docstring''' A__ : int = parent A__ : Union[str, Any] = batch_size A__ : Optional[int] = seq_length A__ : List[Any] = is_training A__ : List[str] = use_input_mask A__ : Optional[Any] = use_token_type_ids A__ : List[Any] = use_labels A__ : Union[str, Any] = vocab_size A__ : List[Any] = hidden_size A__ : Any = num_hidden_layers A__ : Any = num_attention_heads A__ : Optional[int] = intermediate_size A__ : Any = hidden_act A__ : Tuple = hidden_dropout_prob A__ : Dict = attention_probs_dropout_prob A__ : Optional[int] = max_position_embeddings A__ : Tuple = type_vocab_size A__ : Union[str, Any] = type_sequence_label_size A__ : List[str] = initializer_range A__ : Any = num_labels A__ : Any = num_choices A__ : int = scope def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) A__ : Tuple = None if self.use_input_mask: A__ : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) A__ : Union[str, Any] = None if self.use_token_type_ids: A__ : str = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) A__ : int = None A__ : int = None A__ : List[str] = None if self.use_labels: A__ : List[Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) A__ : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) A__ : Dict = ids_tensor([self.batch_size] , self.num_choices ) A__ : Union[str, Any] = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def _UpperCamelCase ( self : Tuple ): '''simple docstring''' return BioGptConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case , initializer_range=self.initializer_range , ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Any , snake_case : Dict , snake_case : Any , snake_case : Union[str, Any] , snake_case : Tuple , snake_case : Optional[Any] ): '''simple docstring''' A__ : Optional[Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() A__ : List[Any] = model(snake_case , attention_mask=snake_case ) A__ : Dict = model(snake_case ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def _UpperCamelCase ( self : Optional[int] , snake_case : Dict , snake_case : Optional[int] , snake_case : List[str] , snake_case : str , snake_case : Optional[Any] , snake_case : List[str] , snake_case : List[Any] , snake_case : Tuple , snake_case : Optional[Any] , ): '''simple docstring''' A__ : List[str] = BioGptForCausalLM(config=snake_case ) model.to(snake_case ) model.eval() A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case , labels=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def _UpperCamelCase ( self : Any , snake_case : str , snake_case : Tuple , snake_case : int , snake_case : Optional[Any] , snake_case : Any , *snake_case : Dict ): '''simple docstring''' A__ : Union[str, Any] = BioGptModel(config=snake_case ) model.to(snake_case ) model.eval() # create attention mask A__ : List[Any] = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) A__ : Any = self.seq_length // 2 A__ : str = 0 # first forward pass A__ , A__ : List[Any] = model(snake_case , attention_mask=snake_case ).to_tuple() # create hypothetical next token and extent to next_input_ids A__ : int = ids_tensor((self.batch_size, 1) , config.vocab_size ) # change a random masked slice from input_ids A__ : List[str] = ids_tensor((1,) , snake_case ).item() + 1 A__ : Optional[int] = ids_tensor((self.batch_size, 1) , config.vocab_size ).squeeze(-1 ) A__ : int = random_other_next_tokens # append to next input_ids and attn_mask A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : List[Any] = torch.cat( [attn_mask, torch.ones((attn_mask.shape[0], 1) , dtype=torch.long , device=snake_case )] , dim=1 , ) # get two different outputs A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Optional[int] = model(snake_case , past_key_values=snake_case , attention_mask=snake_case )["""last_hidden_state"""] # select random slice A__ : List[Any] = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : List[str] = output_from_no_past[:, -1, random_slice_idx].detach() A__ : Any = output_from_past[:, 0, random_slice_idx].detach() # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : int , snake_case : Optional[Any] , snake_case : Optional[Any] , snake_case : str , snake_case : int , snake_case : Optional[Any] , *snake_case : str ): '''simple docstring''' A__ : Dict = BioGptModel(config=snake_case ).to(snake_case ).eval() A__ : Tuple = torch.ones(input_ids.shape , dtype=torch.long , device=snake_case ) # first forward pass A__ : Dict = model(snake_case , attention_mask=snake_case , use_cache=snake_case ) A__ , A__ : List[Any] = outputs.to_tuple() # create hypothetical multiple next token and extent to next_input_ids A__ : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) A__ : int = ids_tensor((self.batch_size, 3) , 2 ) # append to next input_ids and A__ : str = torch.cat([input_ids, next_tokens] , dim=-1 ) A__ : Optional[int] = torch.cat([attention_mask, next_attn_mask] , dim=-1 ) A__ : Any = model(snake_case , attention_mask=snake_case )["""last_hidden_state"""] A__ : Union[str, Any] = model(snake_case , attention_mask=snake_case , past_key_values=snake_case )[ """last_hidden_state""" ] # select random slice A__ : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() A__ : Any = output_from_no_past[:, -3:, random_slice_idx].detach() A__ : List[Any] = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case , snake_case , atol=1e-3 ) ) def _UpperCamelCase ( self : Optional[Any] , snake_case : Tuple , snake_case : Union[str, Any] , snake_case : Optional[Any] , snake_case : Any , snake_case : Tuple , *snake_case : Union[str, Any] , snake_case : Union[str, Any]=False ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM(snake_case ) model.to(snake_case ) if gradient_checkpointing: model.gradient_checkpointing_enable() A__ : Optional[Any] = model(snake_case , labels=snake_case ) self.parent.assertEqual(result.loss.shape , () ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) result.loss.backward() def _UpperCamelCase ( self : int , snake_case : Optional[Any] , *snake_case : Optional[int] ): '''simple docstring''' A__ : int = BioGptModel(snake_case ) A__ : Union[str, Any] = model.config.initializer_range / math.sqrt(2 * model.config.num_hidden_layers ) for key in model.state_dict().keys(): if "c_proj" in key and "weight" in key: self.parent.assertLessEqual(abs(torch.std(model.state_dict()[key] ) - model_std ) , 0.001 ) self.parent.assertLessEqual(abs(torch.mean(model.state_dict()[key] ) - 0.0 ) , 0.01 ) def _UpperCamelCase ( self : Any , snake_case : Dict , snake_case : Tuple , snake_case : int , snake_case : Union[str, Any] , snake_case : Dict , *snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = self.num_labels A__ : int = BioGptForTokenClassification(snake_case ) model.to(snake_case ) model.eval() A__ : str = model(snake_case , attention_mask=snake_case , token_type_ids=snake_case ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : List[str] = self.prepare_config_and_inputs() ( ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ( A__ ) , ) : str = config_and_inputs A__ : Union[str, Any] = {"""input_ids""": input_ids, """attention_mask""": input_mask} return config, inputs_dict @require_torch class __SCREAMING_SNAKE_CASE ( UpperCamelCase , UpperCamelCase , UpperCamelCase , unittest.TestCase ): snake_case_ = ( (BioGptModel, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification) if is_torch_available() else () ) snake_case_ = (BioGptForCausalLM,) if is_torch_available() else () snake_case_ = ( { 'feature-extraction': BioGptModel, 'text-classification': BioGptForSequenceClassification, 'text-generation': BioGptForCausalLM, 'token-classification': BioGptForTokenClassification, 'zero-shot': BioGptForSequenceClassification, } if is_torch_available() else {} ) snake_case_ = False def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : List[str] = BioGptModelTester(self ) A__ : List[Any] = ConfigTester(self , config_class=snake_case , hidden_size=37 ) def _UpperCamelCase ( self : int ): '''simple docstring''' self.config_tester.run_common_tests() def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Any ): '''simple docstring''' A__ : int = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: A__ : str = type self.model_tester.create_and_check_model(*snake_case ) def _UpperCamelCase ( self : Union[str, Any] ): '''simple docstring''' A__ : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_attention_mask_past(*snake_case ) def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' A__ : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_forward_and_backwards(*snake_case , gradient_checkpointing=snake_case ) def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' A__ : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_model_past_large_inputs(*snake_case ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_weight_initialization(*snake_case ) def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_biogpt_for_token_classification(*snake_case ) @slow def _UpperCamelCase ( self : str ): '''simple docstring''' A__ : Tuple = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) A__ : Optional[int] = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = """left""" # Define PAD Token = EOS Token = 50256 A__ : Optional[int] = tokenizer.eos_token A__ : Dict = model.config.eos_token_id # use different length sentences to test batching A__ : Union[str, Any] = [ """Hello, my dog is a little""", """Today, I""", ] A__ : List[str] = tokenizer(snake_case , return_tensors="""pt""" , padding=snake_case ) A__ : str = inputs["""input_ids"""].to(snake_case ) A__ : Dict = model.generate( input_ids=snake_case , attention_mask=inputs["""attention_mask"""].to(snake_case ) , ) A__ : Optional[int] = tokenizer(sentences[0] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Any = model.generate(input_ids=snake_case ) A__ : List[str] = inputs_non_padded.shape[-1] - inputs["""attention_mask"""][-1].long().sum().cpu().item() A__ : str = tokenizer(sentences[1] , return_tensors="""pt""" ).input_ids.to(snake_case ) A__ : Dict = model.generate(input_ids=snake_case , max_length=model.config.max_length - num_paddings ) A__ : Optional[Any] = tokenizer.batch_decode(snake_case , skip_special_tokens=snake_case ) A__ : List[Any] = tokenizer.decode(output_non_padded[0] , skip_special_tokens=snake_case ) A__ : str = tokenizer.decode(output_padded[0] , skip_special_tokens=snake_case ) A__ : Optional[int] = [ """Hello, my dog is a little bit bigger than a little bit.""", """Today, I have a good idea of how to use the information""", ] self.assertListEqual(snake_case , snake_case ) self.assertListEqual(snake_case , [non_padded_sentence, padded_sentence] ) @slow def _UpperCamelCase ( self : Optional[Any] ): '''simple docstring''' for model_name in BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: A__ : Optional[Any] = BioGptModel.from_pretrained(snake_case ) self.assertIsNotNone(snake_case ) def _UpperCamelCase ( self : str ): '''simple docstring''' A__ , A__ : Dict = self.model_tester.prepare_config_and_inputs_for_common() A__ : Optional[int] = 3 A__ : List[Any] = input_dict["""input_ids"""] A__ : Dict = input_ids.ne(1 ).to(snake_case ) A__ : Optional[Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) A__ : Union[str, Any] = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : int = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def _UpperCamelCase ( self : int ): '''simple docstring''' A__ , A__ : str = self.model_tester.prepare_config_and_inputs_for_common() A__ : Any = 3 A__ : List[Any] = """multi_label_classification""" A__ : Dict = input_dict["""input_ids"""] A__ : Tuple = input_ids.ne(1 ).to(snake_case ) A__ : Any = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) A__ : Tuple = BioGptForSequenceClassification(snake_case ) model.to(snake_case ) model.eval() A__ : List[str] = model(snake_case , attention_mask=snake_case , labels=snake_case ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) @require_torch class __SCREAMING_SNAKE_CASE ( unittest.TestCase ): @slow def _UpperCamelCase ( self : List[str] ): '''simple docstring''' A__ : Optional[Any] = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) A__ : str = torch.tensor([[2, 4805, 9, 656, 21]] ) A__ : Dict = model(snake_case )[0] A__ : Tuple = 4_2384 A__ : str = torch.Size((1, 5, vocab_size) ) self.assertEqual(output.shape , snake_case ) A__ : str = torch.tensor( [[[-9.5236, -9.8918, 10.4557], [-11.0469, -9.6423, 8.1022], [-8.8664, -7.8826, 5.5325]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case , atol=1e-4 ) ) @slow def _UpperCamelCase ( self : Dict ): '''simple docstring''' A__ : Tuple = BioGptTokenizer.from_pretrained("""microsoft/biogpt""" ) A__ : Any = BioGptForCausalLM.from_pretrained("""microsoft/biogpt""" ) model.to(snake_case ) torch.manual_seed(0 ) A__ : Tuple = tokenizer("""COVID-19 is""" , return_tensors="""pt""" ).to(snake_case ) A__ : Optional[int] = model.generate( **snake_case , min_length=100 , max_length=1024 , num_beams=5 , early_stopping=snake_case , ) A__ : Optional[int] = tokenizer.decode(output_ids[0] , skip_special_tokens=snake_case ) A__ : List[str] = ( """COVID-19 is a global pandemic caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), the""" """ causative agent of coronavirus disease 2019 (COVID-19), which has spread to more than 200 countries and""" """ territories, including the United States (US), Canada, Australia, New Zealand, the United Kingdom (UK),""" """ and the United States of America (USA), as of March 11, 2020, with more than 800,000 confirmed cases and""" """ more than 800,000 deaths.""" ) self.assertEqual(snake_case , snake_case )
296
1
"""simple docstring""" import random class __SCREAMING_SNAKE_CASE : @staticmethod def _UpperCamelCase ( snake_case : str ): '''simple docstring''' A__ : int = [ord(snake_case ) for i in text] A__ : Any = [] A__ : str = [] for i in plain: A__ : Optional[int] = random.randint(1 , 300 ) A__ : Optional[int] = (i + k) * k cipher.append(snake_case ) key.append(snake_case ) return cipher, key @staticmethod def _UpperCamelCase ( snake_case : list[int] , snake_case : list[int] ): '''simple docstring''' A__ : List[Any] = [] for i in range(len(snake_case ) ): A__ : int = int((cipher[i] - (key[i]) ** 2) / key[i] ) plain.append(chr(snake_case ) ) return "".join(snake_case ) if __name__ == "__main__": A_ , A_ = Onepad().encrypt('''Hello''') print(c, k) print(Onepad().decrypt(c, k))
296
"""simple docstring""" import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging A_ = logging.get_logger(__name__) A_ = {'''vocab_file''': '''spiece.model'''} A_ = { '''vocab_file''': { '''xlnet-base-cased''': '''https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model''', '''xlnet-large-cased''': '''https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model''', } } A_ = { '''xlnet-base-cased''': None, '''xlnet-large-cased''': None, } # Segments (not really needed) A_ = 0 A_ = 1 A_ = 2 A_ = 3 A_ = 4 class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = VOCAB_FILES_NAMES snake_case_ = PRETRAINED_VOCAB_FILES_MAP snake_case_ = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES snake_case_ = 'left' def __init__( self : Dict , snake_case : int , snake_case : List[Any]=False , snake_case : List[str]=True , snake_case : Dict=False , snake_case : Optional[Any]="<s>" , snake_case : List[str]="</s>" , snake_case : Tuple="<unk>" , snake_case : Tuple="<sep>" , snake_case : Union[str, Any]="<pad>" , snake_case : Dict="<cls>" , snake_case : Optional[Any]="<mask>" , snake_case : Optional[int]=["<eop>", "<eod>"] , snake_case : Optional[Dict[str, Any]] = None , **snake_case : Dict , ): '''simple docstring''' A__ : Optional[int] = AddedToken(snake_case , lstrip=snake_case , rstrip=snake_case ) if isinstance(snake_case , snake_case ) else mask_token A__ : Union[str, Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=snake_case , remove_space=snake_case , keep_accents=snake_case , bos_token=snake_case , eos_token=snake_case , unk_token=snake_case , sep_token=snake_case , pad_token=snake_case , cls_token=snake_case , mask_token=snake_case , additional_special_tokens=snake_case , sp_model_kwargs=self.sp_model_kwargs , **snake_case , ) A__ : str = 3 A__ : str = do_lower_case A__ : Optional[Any] = remove_space A__ : List[Any] = keep_accents A__ : Union[str, Any] = vocab_file A__ : Dict = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(snake_case ) @property def _UpperCamelCase ( self : Optional[int] ): '''simple docstring''' return len(self.sp_model ) def _UpperCamelCase ( self : List[Any] ): '''simple docstring''' A__ : int = {self.convert_ids_to_tokens(snake_case ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__( self : str ): '''simple docstring''' A__ : int = self.__dict__.copy() A__ : int = None return state def __setstate__( self : Tuple , snake_case : Union[str, Any] ): '''simple docstring''' A__ : int = d # for backward compatibility if not hasattr(self , """sp_model_kwargs""" ): A__ : Optional[int] = {} A__ : Optional[int] = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def _UpperCamelCase ( self : List[str] , snake_case : Optional[Any] ): '''simple docstring''' if self.remove_space: A__ : Optional[Any] = """ """.join(inputs.strip().split() ) else: A__ : Dict = inputs A__ : str = outputs.replace("""``""" , """\"""" ).replace("""''""" , """\"""" ) if not self.keep_accents: A__ : Any = unicodedata.normalize("""NFKD""" , snake_case ) A__ : Optional[int] = """""".join([c for c in outputs if not unicodedata.combining(snake_case )] ) if self.do_lower_case: A__ : Any = outputs.lower() return outputs def _UpperCamelCase ( self : Union[str, Any] , snake_case : str ): '''simple docstring''' A__ : Dict = self.preprocess_text(snake_case ) A__ : Dict = self.sp_model.encode(snake_case , out_type=snake_case ) A__ : Optional[int] = [] for piece in pieces: if len(snake_case ) > 1 and piece[-1] == str(""",""" ) and piece[-2].isdigit(): A__ : Optional[Any] = self.sp_model.EncodeAsPieces(piece[:-1].replace(snake_case , """""" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: A__ : int = cur_pieces[1:] else: A__ : Any = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(snake_case ) else: new_pieces.append(snake_case ) return new_pieces def _UpperCamelCase ( self : List[str] , snake_case : Tuple ): '''simple docstring''' return self.sp_model.PieceToId(snake_case ) def _UpperCamelCase ( self : List[str] , snake_case : Any ): '''simple docstring''' return self.sp_model.IdToPiece(snake_case ) def _UpperCamelCase ( self : Optional[int] , snake_case : Any ): '''simple docstring''' A__ : Union[str, Any] = """""".join(snake_case ).replace(snake_case , """ """ ).strip() return out_string def _UpperCamelCase ( self : int , snake_case : List[int] , snake_case : bool = False , snake_case : bool = None , snake_case : bool = True , **snake_case : Union[str, Any] , ): '''simple docstring''' A__ : List[str] = kwargs.pop("""use_source_tokenizer""" , snake_case ) A__ : Any = self.convert_ids_to_tokens(snake_case , skip_special_tokens=snake_case ) # To avoid mixing byte-level and unicode for byte-level BPT # we need to build string separately for added tokens and byte-level tokens # cf. https://github.com/huggingface/transformers/issues/1133 A__ : Any = [] A__ : Any = [] for token in filtered_tokens: if skip_special_tokens and token in self.all_special_ids: continue if token in self.added_tokens_encoder: if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) A__ : str = [] sub_texts.append(snake_case ) else: current_sub_text.append(snake_case ) if current_sub_text: sub_texts.append(self.convert_tokens_to_string(snake_case ) ) # Mimic the behavior of the Rust tokenizer: # By default, there are no spaces between special tokens A__ : Dict = """""".join(snake_case ) A__ : int = ( clean_up_tokenization_spaces if clean_up_tokenization_spaces is not None else self.clean_up_tokenization_spaces ) if clean_up_tokenization_spaces: A__ : Tuple = self.clean_up_tokenization(snake_case ) return clean_text else: return text def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Tuple = [self.sep_token_id] A__ : Dict = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def _UpperCamelCase ( self : Dict , snake_case : List[int] , snake_case : Optional[List[int]] = None , snake_case : bool = False ): '''simple docstring''' if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case , token_ids_a=snake_case , already_has_special_tokens=snake_case ) if token_ids_a is not None: return ([0] * len(snake_case )) + [1] + ([0] * len(snake_case )) + [1, 1] return ([0] * len(snake_case )) + [1, 1] def _UpperCamelCase ( self : str , snake_case : List[int] , snake_case : Optional[List[int]] = None ): '''simple docstring''' A__ : Any = [self.sep_token_id] A__ : int = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def _UpperCamelCase ( self : Optional[Any] , snake_case : str , snake_case : Optional[str] = None ): '''simple docstring''' if not os.path.isdir(snake_case ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return A__ : List[Any] = os.path.join( snake_case , (filename_prefix + """-""" if filename_prefix else """""") + VOCAB_FILES_NAMES["""vocab_file"""] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case ) elif not os.path.isfile(self.vocab_file ): with open(snake_case , """wb""" ) as fi: A__ : Optional[Any] = self.sp_model.serialized_model_proto() fi.write(snake_case ) return (out_vocab_file,)
296
1
"""simple docstring""" from sklearn.metrics import mean_squared_error import datasets A_ = '''\ @article{scikit-learn, title={Scikit-learn: Machine Learning in {P}ython}, author={Pedregosa, F. and Varoquaux, G. and Gramfort, A. and Michel, V. and Thirion, B. and Grisel, O. and Blondel, M. and Prettenhofer, P. and Weiss, R. and Dubourg, V. and Vanderplas, J. and Passos, A. and Cournapeau, D. and Brucher, M. and Perrot, M. and Duchesnay, E.}, journal={Journal of Machine Learning Research}, volume={12}, pages={2825--2830}, year={2011} } ''' A_ = '''\ Mean Squared Error(MSE) is the average of the square of difference between the predicted and actual values. ''' A_ = ''' Args: predictions: array-like of shape (n_samples,) or (n_samples, n_outputs) Estimated target values. references: array-like of shape (n_samples,) or (n_samples, n_outputs) Ground truth (correct) target values. sample_weight: array-like of shape (n_samples,), default=None Sample weights. multioutput: {"raw_values", "uniform_average"} or array-like of shape (n_outputs,), default="uniform_average" Defines aggregating of multiple output values. Array-like value defines weights used to average errors. "raw_values" : Returns a full set of errors in case of multioutput input. "uniform_average" : Errors of all outputs are averaged with uniform weight. squared : bool, default=True If True returns MSE value, if False returns RMSE (Root Mean Squared Error) value. Returns: mse : mean squared error. Examples: >>> mse_metric = datasets.load_metric("mse") >>> predictions = [2.5, 0.0, 2, 8] >>> references = [3, -0.5, 2, 7] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.375} >>> rmse_result = mse_metric.compute(predictions=predictions, references=references, squared=False) >>> print(rmse_result) {\'mse\': 0.6123724356957945} If you\'re using multi-dimensional lists, then set the config as follows : >>> mse_metric = datasets.load_metric("mse", "multilist") >>> predictions = [[0.5, 1], [-1, 1], [7, -6]] >>> references = [[0, 2], [-1, 2], [8, -5]] >>> results = mse_metric.compute(predictions=predictions, references=references) >>> print(results) {\'mse\': 0.7083333333333334} >>> results = mse_metric.compute(predictions=predictions, references=references, multioutput=\'raw_values\') >>> print(results) # doctest: +NORMALIZE_WHITESPACE {\'mse\': array([0.41666667, 1. ])} ''' @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION ) class __SCREAMING_SNAKE_CASE ( datasets.Metric ): def _UpperCamelCase ( self : Dict ): '''simple docstring''' return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features(self._get_feature_types() ) , reference_urls=[ """https://scikit-learn.org/stable/modules/generated/sklearn.metrics.mean_squared_error.html""" ] , ) def _UpperCamelCase ( self : Tuple ): '''simple docstring''' if self.config_name == "multilist": return { "predictions": datasets.Sequence(datasets.Value("""float""" ) ), "references": datasets.Sequence(datasets.Value("""float""" ) ), } else: return { "predictions": datasets.Value("""float""" ), "references": datasets.Value("""float""" ), } def _UpperCamelCase ( self : List[str] , snake_case : Dict , snake_case : List[Any] , snake_case : List[str]=None , snake_case : List[Any]="uniform_average" , snake_case : int=True ): '''simple docstring''' A__ : Optional[int] = mean_squared_error( snake_case , snake_case , sample_weight=snake_case , multioutput=snake_case , squared=snake_case ) return {"mse": mse}
296
"""simple docstring""" import argparse import json from pathlib import Path import requests import torch from huggingface_hub import cached_download, hf_hub_url from PIL import Image from transformers import DPTConfig, DPTForDepthEstimation, DPTForSemanticSegmentation, DPTImageProcessor from transformers.utils import logging logging.set_verbosity_info() A_ = logging.get_logger(__name__) def _lowerCAmelCase ( UpperCAmelCase__ : List[Any] ) ->List[str]: A__ : Union[str, Any] = DPTConfig() if "large" in checkpoint_url: A__ : int = 1_0_2_4 A__ : Union[str, Any] = 4_0_9_6 A__ : Optional[int] = 2_4 A__ : int = 1_6 A__ : Union[str, Any] = [5, 1_1, 1_7, 2_3] A__ : Tuple = [2_5_6, 5_1_2, 1_0_2_4, 1_0_2_4] A__ : Tuple = (1, 3_8_4, 3_8_4) if "ade" in checkpoint_url: A__ : Optional[int] = True A__ : int = 1_5_0 A__ : Union[str, Any] = """huggingface/label-files""" A__ : List[Any] = """ade20k-id2label.json""" A__ : Union[str, Any] = json.load(open(cached_download(hf_hub_url(UpperCAmelCase__, UpperCAmelCase__, repo_type="""dataset""" ) ), """r""" ) ) A__ : List[Any] = {int(UpperCAmelCase__ ): v for k, v in idalabel.items()} A__ : Dict = idalabel A__ : List[Any] = {v: k for k, v in idalabel.items()} A__ : Optional[Any] = [1, 1_5_0, 4_8_0, 4_8_0] return config, expected_shape def _lowerCAmelCase ( UpperCAmelCase__ : int ) ->Any: A__ : List[Any] = ["""pretrained.model.head.weight""", """pretrained.model.head.bias"""] for k in ignore_keys: state_dict.pop(UpperCAmelCase__, UpperCAmelCase__ ) def _lowerCAmelCase ( UpperCAmelCase__ : Union[str, Any] ) ->List[str]: if ( "pretrained.model" in name and "cls_token" not in name and "pos_embed" not in name and "patch_embed" not in name ): A__ : str = name.replace("""pretrained.model""", """dpt.encoder""" ) if "pretrained.model" in name: A__ : Dict = name.replace("""pretrained.model""", """dpt.embeddings""" ) if "patch_embed" in name: A__ : List[Any] = name.replace("""patch_embed""", """patch_embeddings""" ) if "pos_embed" in name: A__ : int = name.replace("""pos_embed""", """position_embeddings""" ) if "attn.proj" in name: A__ : Tuple = name.replace("""attn.proj""", """attention.output.dense""" ) if "proj" in name and "project" not in name: A__ : List[Any] = name.replace("""proj""", """projection""" ) if "blocks" in name: A__ : Optional[Any] = name.replace("""blocks""", """layer""" ) if "mlp.fc1" in name: A__ : int = name.replace("""mlp.fc1""", """intermediate.dense""" ) if "mlp.fc2" in name: A__ : List[str] = name.replace("""mlp.fc2""", """output.dense""" ) if "norm1" in name: A__ : Any = name.replace("""norm1""", """layernorm_before""" ) if "norm2" in name: A__ : List[str] = name.replace("""norm2""", """layernorm_after""" ) if "scratch.output_conv" in name: A__ : Optional[int] = name.replace("""scratch.output_conv""", """head""" ) if "scratch" in name: A__ : List[str] = name.replace("""scratch""", """neck""" ) if "layer1_rn" in name: A__ : List[str] = name.replace("""layer1_rn""", """convs.0""" ) if "layer2_rn" in name: A__ : Optional[int] = name.replace("""layer2_rn""", """convs.1""" ) if "layer3_rn" in name: A__ : Any = name.replace("""layer3_rn""", """convs.2""" ) if "layer4_rn" in name: A__ : Any = name.replace("""layer4_rn""", """convs.3""" ) if "refinenet" in name: A__ : Union[str, Any] = int(name[len("""neck.refinenet""" ) : len("""neck.refinenet""" ) + 1] ) # tricky here: we need to map 4 to 0, 3 to 1, 2 to 2 and 1 to 3 A__ : str = name.replace(f'refinenet{layer_idx}', f'fusion_stage.layers.{abs(layer_idx-4 )}' ) if "out_conv" in name: A__ : Optional[Any] = name.replace("""out_conv""", """projection""" ) if "resConfUnit1" in name: A__ : List[Any] = name.replace("""resConfUnit1""", """residual_layer1""" ) if "resConfUnit2" in name: A__ : Tuple = name.replace("""resConfUnit2""", """residual_layer2""" ) if "conv1" in name: A__ : Tuple = name.replace("""conv1""", """convolution1""" ) if "conv2" in name: A__ : List[Any] = name.replace("""conv2""", """convolution2""" ) # readout blocks if "pretrained.act_postprocess1.0.project.0" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess1.0.project.0""", """neck.reassemble_stage.readout_projects.0.0""" ) if "pretrained.act_postprocess2.0.project.0" in name: A__ : Tuple = name.replace("""pretrained.act_postprocess2.0.project.0""", """neck.reassemble_stage.readout_projects.1.0""" ) if "pretrained.act_postprocess3.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess3.0.project.0""", """neck.reassemble_stage.readout_projects.2.0""" ) if "pretrained.act_postprocess4.0.project.0" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess4.0.project.0""", """neck.reassemble_stage.readout_projects.3.0""" ) # resize blocks if "pretrained.act_postprocess1.3" in name: A__ : Any = name.replace("""pretrained.act_postprocess1.3""", """neck.reassemble_stage.layers.0.projection""" ) if "pretrained.act_postprocess1.4" in name: A__ : List[Any] = name.replace("""pretrained.act_postprocess1.4""", """neck.reassemble_stage.layers.0.resize""" ) if "pretrained.act_postprocess2.3" in name: A__ : Dict = name.replace("""pretrained.act_postprocess2.3""", """neck.reassemble_stage.layers.1.projection""" ) if "pretrained.act_postprocess2.4" in name: A__ : Optional[Any] = name.replace("""pretrained.act_postprocess2.4""", """neck.reassemble_stage.layers.1.resize""" ) if "pretrained.act_postprocess3.3" in name: A__ : Union[str, Any] = name.replace("""pretrained.act_postprocess3.3""", """neck.reassemble_stage.layers.2.projection""" ) if "pretrained.act_postprocess4.3" in name: A__ : Optional[int] = name.replace("""pretrained.act_postprocess4.3""", """neck.reassemble_stage.layers.3.projection""" ) if "pretrained.act_postprocess4.4" in name: A__ : Dict = name.replace("""pretrained.act_postprocess4.4""", """neck.reassemble_stage.layers.3.resize""" ) if "pretrained" in name: A__ : Union[str, Any] = name.replace("""pretrained""", """dpt""" ) if "bn" in name: A__ : Union[str, Any] = name.replace("""bn""", """batch_norm""" ) if "head" in name: A__ : Dict = name.replace("""head""", """head.head""" ) if "encoder.norm" in name: A__ : Optional[int] = name.replace("""encoder.norm""", """layernorm""" ) if "auxlayer" in name: A__ : List[str] = name.replace("""auxlayer""", """auxiliary_head.head""" ) return name def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Dict ) ->str: for i in range(config.num_hidden_layers ): # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) A__ : Any = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.weight' ) A__ : Tuple = state_dict.pop(f'dpt.encoder.layer.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict A__ : List[str] = in_proj_weight[: config.hidden_size, :] A__ : int = in_proj_bias[: config.hidden_size] A__ : Tuple = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] A__ : Any = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] A__ : str = in_proj_weight[ -config.hidden_size :, : ] A__ : Optional[Any] = in_proj_bias[-config.hidden_size :] def _lowerCAmelCase ( ) ->List[str]: A__ : int = """http://images.cocodataset.org/val2017/000000039769.jpg""" A__ : int = Image.open(requests.get(UpperCAmelCase__, stream=UpperCAmelCase__ ).raw ) return im @torch.no_grad() def _lowerCAmelCase ( UpperCAmelCase__ : int, UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : str, UpperCAmelCase__ : int ) ->str: A__ , A__ : Dict = get_dpt_config(UpperCAmelCase__ ) # load original state_dict from URL A__ : Any = torch.hub.load_state_dict_from_url(UpperCAmelCase__, map_location="""cpu""" ) # remove certain keys remove_ignore_keys_(UpperCAmelCase__ ) # rename keys for key in state_dict.copy().keys(): A__ : int = state_dict.pop(UpperCAmelCase__ ) A__ : str = val # read in qkv matrices read_in_q_k_v(UpperCAmelCase__, UpperCAmelCase__ ) # load HuggingFace model A__ : Optional[Any] = DPTForSemanticSegmentation(UpperCAmelCase__ ) if """ade""" in checkpoint_url else DPTForDepthEstimation(UpperCAmelCase__ ) model.load_state_dict(UpperCAmelCase__ ) model.eval() # Check outputs on an image A__ : Optional[Any] = 4_8_0 if """ade""" in checkpoint_url else 3_8_4 A__ : Dict = DPTImageProcessor(size=UpperCAmelCase__ ) A__ : Optional[int] = prepare_img() A__ : Any = image_processor(UpperCAmelCase__, return_tensors="""pt""" ) # forward pass A__ : List[str] = model(**UpperCAmelCase__ ).logits if """ade""" in checkpoint_url else model(**UpperCAmelCase__ ).predicted_depth # Assert logits A__ : Optional[Any] = torch.tensor([[6.3199, 6.3629, 6.4148], [6.3850, 6.3615, 6.4166], [6.3519, 6.3176, 6.3575]] ) if "ade" in checkpoint_url: A__ : Optional[int] = torch.tensor([[4.0480, 4.2420, 4.4360], [4.3124, 4.5693, 4.8261], [4.5768, 4.8965, 5.2163]] ) assert outputs.shape == torch.Size(UpperCAmelCase__ ) assert ( torch.allclose(outputs[0, 0, :3, :3], UpperCAmelCase__, atol=1e-4 ) if "ade" in checkpoint_url else torch.allclose(outputs[0, :3, :3], UpperCAmelCase__ ) ) Path(UpperCAmelCase__ ).mkdir(exist_ok=UpperCAmelCase__ ) print(f'Saving model to {pytorch_dump_folder_path}' ) model.save_pretrained(UpperCAmelCase__ ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(UpperCAmelCase__ ) if push_to_hub: print("""Pushing model to hub...""" ) model.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add model""", use_temp_dir=UpperCAmelCase__, ) image_processor.push_to_hub( repo_path_or_name=Path(UpperCAmelCase__, UpperCAmelCase__ ), organization="""nielsr""", commit_message="""Add image processor""", use_temp_dir=UpperCAmelCase__, ) if __name__ == "__main__": A_ = argparse.ArgumentParser() # Required parameters parser.add_argument( '''--checkpoint_url''', default='''https://github.com/intel-isl/DPT/releases/download/1_0/dpt_large-midas-2f21e586.pt''', type=str, help='''URL of the original DPT checkpoint you\'d like to convert.''', ) parser.add_argument( '''--pytorch_dump_folder_path''', default=None, type=str, required=True, help='''Path to the output PyTorch model directory.''', ) parser.add_argument( '''--push_to_hub''', action='''store_true''', ) parser.add_argument( '''--model_name''', default='''dpt-large''', type=str, help='''Name of the model, in case you\'re pushing to the hub.''', ) A_ = parser.parse_args() convert_dpt_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub, args.model_name)
296
1
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_speech_available, is_tf_available, is_torch_available, ) A_ = { '''configuration_speech_to_text''': ['''SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''Speech2TextConfig'''], '''processing_speech_to_text''': ['''Speech2TextProcessor'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''Speech2TextTokenizer'''] try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''Speech2TextFeatureExtractor'''] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''TFSpeech2TextForConditionalGeneration''', '''TFSpeech2TextModel''', '''TFSpeech2TextPreTrainedModel''', ] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST''', '''Speech2TextForConditionalGeneration''', '''Speech2TextModel''', '''Speech2TextPreTrainedModel''', ] if TYPE_CHECKING: from .configuration_speech_to_text import SPEECH_TO_TEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, SpeechaTextConfig from .processing_speech_to_text import SpeechaTextProcessor try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_speech_to_text import SpeechaTextTokenizer try: if not is_speech_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_speech_to_text import SpeechaTextFeatureExtractor try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_speech_to_text import ( TF_SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, TFSpeechaTextForConditionalGeneration, TFSpeechaTextModel, TFSpeechaTextPreTrainedModel, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_speech_to_text import ( SPEECH_TO_TEXT_PRETRAINED_MODEL_ARCHIVE_LIST, SpeechaTextForConditionalGeneration, SpeechaTextModel, SpeechaTextPreTrainedModel, ) else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
"""simple docstring""" import argparse import glob import importlib.util import os import re import black from doc_builder.style_doc import style_docstrings_in_code # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_copies.py A_ = '''src/diffusers''' A_ = '''.''' # This is to make sure the diffusers module imported is the one in the repo. A_ = importlib.util.spec_from_file_location( '''diffusers''', os.path.join(DIFFUSERS_PATH, '''__init__.py'''), submodule_search_locations=[DIFFUSERS_PATH], ) A_ = spec.loader.load_module() def _lowerCAmelCase ( UpperCAmelCase__ : Optional[int], UpperCAmelCase__ : Optional[Any] ) ->Any: return line.startswith(UpperCAmelCase__ ) or len(UpperCAmelCase__ ) <= 1 or re.search(R"""^\s*\)(\s*->.*:|:)\s*$""", UpperCAmelCase__ ) is not None def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Union[str, Any]: A__ : Any = object_name.split(""".""" ) A__ : int = 0 # First let's find the module where our object lives. A__ : str = parts[i] while i < len(UpperCAmelCase__ ) and not os.path.isfile(os.path.join(UpperCAmelCase__, f'{module}.py' ) ): i += 1 if i < len(UpperCAmelCase__ ): A__ : Union[str, Any] = os.path.join(UpperCAmelCase__, parts[i] ) if i >= len(UpperCAmelCase__ ): raise ValueError(f'`object_name` should begin with the name of a module of diffusers but got {object_name}.' ) with open(os.path.join(UpperCAmelCase__, f'{module}.py' ), """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : List[Any] = f.readlines() # Now let's find the class / func in the code! A__ : Optional[Any] = """""" A__ : Any = 0 for name in parts[i + 1 :]: while ( line_index < len(UpperCAmelCase__ ) and re.search(Rf'^{indent}(class|def)\s+{name}(\(|\:)', lines[line_index] ) is None ): line_index += 1 indent += " " line_index += 1 if line_index >= len(UpperCAmelCase__ ): raise ValueError(f' {object_name} does not match any function or class in {module}.' ) # We found the beginning of the class / func, now let's find the end (when the indent diminishes). A__ : List[Any] = line_index while line_index < len(UpperCAmelCase__ ) and _should_continue(lines[line_index], UpperCAmelCase__ ): line_index += 1 # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : List[Any] = lines[start_index:line_index] return "".join(UpperCAmelCase__ ) A_ = re.compile(r'''^(\s*)#\s*Copied from\s+diffusers\.(\S+\.\S+)\s*($|\S.*$)''') A_ = re.compile(r'''^\s*(\S+)->(\S+)(\s+.*|$)''') A_ = re.compile(r'''<FILL\s+[^>]*>''') def _lowerCAmelCase ( UpperCAmelCase__ : List[str] ) ->Optional[Any]: A__ : Dict = code.split("""\n""" ) A__ : List[Any] = 0 while idx < len(UpperCAmelCase__ ) and len(lines[idx] ) == 0: idx += 1 if idx < len(UpperCAmelCase__ ): return re.search(R"""^(\s*)\S""", lines[idx] ).groups()[0] return "" def _lowerCAmelCase ( UpperCAmelCase__ : Optional[Any] ) ->int: A__ : str = len(get_indent(UpperCAmelCase__ ) ) > 0 if has_indent: A__ : Union[str, Any] = f'class Bla:\n{code}' A__ : Optional[Any] = black.Mode(target_versions={black.TargetVersion.PYaa}, line_length=1_1_9, preview=UpperCAmelCase__ ) A__ : Tuple = black.format_str(UpperCAmelCase__, mode=UpperCAmelCase__ ) A__ , A__ : List[Any] = style_docstrings_in_code(UpperCAmelCase__ ) return result[len("""class Bla:\n""" ) :] if has_indent else result def _lowerCAmelCase ( UpperCAmelCase__ : Any, UpperCAmelCase__ : Dict=False ) ->List[Any]: with open(UpperCAmelCase__, """r""", encoding="""utf-8""", newline="""\n""" ) as f: A__ : int = f.readlines() A__ : Dict = [] A__ : List[str] = 0 # Not a for loop cause `lines` is going to change (if `overwrite=True`). while line_index < len(UpperCAmelCase__ ): A__ : Dict = _re_copy_warning.search(lines[line_index] ) if search is None: line_index += 1 continue # There is some copied code here, let's retrieve the original. A__ , A__ , A__ : Dict = search.groups() A__ : Tuple = find_code_in_diffusers(UpperCAmelCase__ ) A__ : int = get_indent(UpperCAmelCase__ ) A__ : List[str] = line_index + 1 if indent == theoretical_indent else line_index + 2 A__ : Tuple = theoretical_indent A__ : Optional[Any] = start_index # Loop to check the observed code, stop when indentation diminishes or if we see a End copy comment. A__ : Tuple = True while line_index < len(UpperCAmelCase__ ) and should_continue: line_index += 1 if line_index >= len(UpperCAmelCase__ ): break A__ : Optional[int] = lines[line_index] A__ : Tuple = _should_continue(UpperCAmelCase__, UpperCAmelCase__ ) and re.search(f'^{indent}# End copy', UpperCAmelCase__ ) is None # Clean up empty lines at the end (if any). while len(lines[line_index - 1] ) <= 1: line_index -= 1 A__ : Dict = lines[start_index:line_index] A__ : Tuple = """""".join(UpperCAmelCase__ ) # Remove any nested `Copied from` comments to avoid circular copies A__ : Optional[int] = [line for line in theoretical_code.split("""\n""" ) if _re_copy_warning.search(UpperCAmelCase__ ) is None] A__ : Optional[Any] = """\n""".join(UpperCAmelCase__ ) # Before comparing, use the `replace_pattern` on the original code. if len(UpperCAmelCase__ ) > 0: A__ : int = replace_pattern.replace("""with""", """""" ).split(""",""" ) A__ : List[Any] = [_re_replace_pattern.search(UpperCAmelCase__ ) for p in patterns] for pattern in patterns: if pattern is None: continue A__ , A__ , A__ : Union[str, Any] = pattern.groups() A__ : Union[str, Any] = re.sub(UpperCAmelCase__, UpperCAmelCase__, UpperCAmelCase__ ) if option.strip() == "all-casing": A__ : List[Any] = re.sub(obja.lower(), obja.lower(), UpperCAmelCase__ ) A__ : Tuple = re.sub(obja.upper(), obja.upper(), UpperCAmelCase__ ) # Blackify after replacement. To be able to do that, we need the header (class or function definition) # from the previous line A__ : Optional[int] = blackify(lines[start_index - 1] + theoretical_code ) A__ : List[Any] = theoretical_code[len(lines[start_index - 1] ) :] # Test for a diff and act accordingly. if observed_code != theoretical_code: diffs.append([object_name, start_index] ) if overwrite: A__ : List[Any] = lines[:start_index] + [theoretical_code] + lines[line_index:] A__ : Tuple = start_index + 1 if overwrite and len(UpperCAmelCase__ ) > 0: # Warn the user a file has been modified. print(f'Detected changes, rewriting {filename}.' ) with open(UpperCAmelCase__, """w""", encoding="""utf-8""", newline="""\n""" ) as f: f.writelines(UpperCAmelCase__ ) return diffs def _lowerCAmelCase ( UpperCAmelCase__ : bool = False ) ->Any: A__ : Dict = glob.glob(os.path.join(UpperCAmelCase__, """**/*.py""" ), recursive=UpperCAmelCase__ ) A__ : str = [] for filename in all_files: A__ : Any = is_copy_consistent(UpperCAmelCase__, UpperCAmelCase__ ) diffs += [f'- {filename}: copy does not match {d[0]} at line {d[1]}' for d in new_diffs] if not overwrite and len(UpperCAmelCase__ ) > 0: A__ : Any = """\n""".join(UpperCAmelCase__ ) raise Exception( """Found the following copy inconsistencies:\n""" + diff + """\nRun `make fix-copies` or `python utils/check_copies.py --fix_and_overwrite` to fix them.""" ) if __name__ == "__main__": A_ = argparse.ArgumentParser() parser.add_argument('''--fix_and_overwrite''', action='''store_true''', help='''Whether to fix inconsistencies.''') A_ = parser.parse_args() check_copies(args.fix_and_overwrite)
296
1
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging A_ = logging.get_logger(__name__) A_ = { '''edbeeching/decision-transformer-gym-hopper-medium''': ( '''https://huggingface.co/edbeeching/decision-transformer-gym-hopper-medium/resolve/main/config.json''' ), # See all DecisionTransformer models at https://huggingface.co/models?filter=decision_transformer } class __SCREAMING_SNAKE_CASE ( UpperCamelCase ): snake_case_ = 'decision_transformer' snake_case_ = ['past_key_values'] snake_case_ = { 'max_position_embeddings': 'n_positions', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__( self : Tuple , snake_case : Union[str, Any]=17 , snake_case : str=4 , snake_case : int=128 , snake_case : Tuple=4096 , snake_case : str=True , snake_case : Optional[Any]=1 , snake_case : str=1024 , snake_case : Union[str, Any]=3 , snake_case : Union[str, Any]=1 , snake_case : Tuple=None , snake_case : Optional[Any]="relu" , snake_case : Optional[Any]=0.1 , snake_case : int=0.1 , snake_case : Optional[Any]=0.1 , snake_case : Union[str, Any]=1e-5 , snake_case : Optional[int]=0.02 , snake_case : Optional[int]=True , snake_case : Any=True , snake_case : Optional[int]=5_0256 , snake_case : List[Any]=5_0256 , snake_case : Tuple=False , snake_case : Tuple=False , **snake_case : Any , ): '''simple docstring''' A__ : Union[str, Any] = state_dim A__ : Tuple = act_dim A__ : Optional[Any] = hidden_size A__ : str = max_ep_len A__ : Any = action_tanh A__ : Any = vocab_size A__ : Dict = n_positions A__ : Optional[Any] = n_layer A__ : int = n_head A__ : List[str] = n_inner A__ : Tuple = activation_function A__ : Any = resid_pdrop A__ : List[str] = embd_pdrop A__ : Any = attn_pdrop A__ : List[Any] = layer_norm_epsilon A__ : Tuple = initializer_range A__ : Union[str, Any] = scale_attn_weights A__ : Tuple = use_cache A__ : int = scale_attn_by_inverse_layer_idx A__ : Optional[int] = reorder_and_upcast_attn A__ : Optional[int] = bos_token_id A__ : Tuple = eos_token_id super().__init__(bos_token_id=snake_case , eos_token_id=snake_case , **snake_case )
296
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tokenizers_available, is_torch_available, ) A_ = { '''configuration_llama''': ['''LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP''', '''LlamaConfig'''], } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizer'''] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = ['''LlamaTokenizerFast'''] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: A_ = [ '''LlamaForCausalLM''', '''LlamaModel''', '''LlamaPreTrainedModel''', '''LlamaForSequenceClassification''', ] if TYPE_CHECKING: from .configuration_llama import LLAMA_PRETRAINED_CONFIG_ARCHIVE_MAP, LlamaConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama import LlamaTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_llama_fast import LlamaTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_llama import LlamaForCausalLM, LlamaForSequenceClassification, LlamaModel, LlamaPreTrainedModel else: import sys A_ = _LazyModule(__name__, globals()['''__file__'''], _import_structure, module_spec=__spec__)
296
1