code
stringlengths
81
54k
code_codestyle
int64
0
721
style_context
stringlengths
91
41.9k
style_context_codestyle
int64
0
699
label
int64
0
1
"""simple docstring""" def a__ ( a : Dict , a : List[Any] ): """simple docstring""" _snake_case : Tuple = "" for word_or_phrase in separated: if not isinstance(UpperCAmelCase__ , UpperCAmelCase__ ): raise Exception("join() accepts only strings to be joined" ) joined += word_or_phrase + separator return joined.strip(UpperCAmelCase__ ) if __name__ == "__main__": from doctest import testmod testmod()
713
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def a__ ( a : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) _a : int = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : Dict = parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=snake_case_ , required=snake_case_ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=snake_case_ , required=snake_case_ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=snake_case_ , required=snake_case_ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=snake_case_ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=snake_case_ , default=snake_case_ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : str = logging.get_logger("transformers-cli/converting" ) self._logger.info(F'Loading model {model_type}' ) _snake_case : Optional[int] = model_type _snake_case : Any = tf_checkpoint _snake_case : Optional[int] = pytorch_dump_output _snake_case : Tuple = config _snake_case : Tuple = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) if "ckpt" in self._tf_checkpoint.lower(): _snake_case : int = self._tf_checkpoint _snake_case : Optional[Any] = "" else: _snake_case : Optional[int] = self._tf_checkpoint _snake_case : List[str] = "" convert_transfo_xl_checkpoint_to_pytorch( snake_case_ , self._config , self._pytorch_dump_output , snake_case_ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
87
0
import copy from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import TensorType, logging _a : int = logging.get_logger(__name__) class _UpperCAmelCase ( _snake_case): __lowercase : Any = ["""input_features"""] def __init__( self , snake_case_=80 , snake_case_=1_60_00 , snake_case_=1_60 , snake_case_=30 , snake_case_=4_00 , snake_case_=0.0 , snake_case_=False , **snake_case_ , ): super().__init__( feature_size=__UpperCamelCase , sampling_rate=__UpperCamelCase , padding_value=__UpperCamelCase , return_attention_mask=__UpperCamelCase , **__UpperCamelCase , ) _snake_case : int = n_fft _snake_case : Optional[Any] = hop_length _snake_case : List[Any] = chunk_length _snake_case : Optional[Any] = chunk_length * sampling_rate _snake_case : List[str] = self.n_samples // hop_length _snake_case : Tuple = sampling_rate _snake_case : Any = mel_filter_bank( num_frequency_bins=1 + n_fft // 2 , num_mel_filters=__UpperCamelCase , min_frequency=0.0 , max_frequency=8000.0 , sampling_rate=__UpperCamelCase , norm="slaney" , mel_scale="slaney" , ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Any = spectrogram( __UpperCamelCase , window_function(self.n_fft , "hann" ) , frame_length=self.n_fft , hop_length=self.hop_length , power=2.0 , mel_filters=self.mel_filters , log_mel="log10" , ) _snake_case : Union[str, Any] = log_spec[:, :-1] _snake_case : List[str] = np.maximum(__UpperCamelCase , log_spec.max() - 8.0 ) _snake_case : str = (log_spec + 4.0) / 4.0 return log_spec @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowerCamelCase__ ( snake_case_ , snake_case_ , snake_case_ = 0.0 ): if attention_mask is not None: _snake_case : str = np.array(__UpperCamelCase , np.intaa ) _snake_case : Optional[int] = [] for vector, length in zip(__UpperCamelCase , attention_mask.sum(-1 ) ): _snake_case : Tuple = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: _snake_case : Dict = padding_value normed_input_values.append(__UpperCamelCase ) else: _snake_case : List[Any] = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def __call__( self , snake_case_ , snake_case_ = True , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = "max_length" , snake_case_ = None , snake_case_ = None , snake_case_ = None , **snake_case_ , ): if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( F'The model corresponding to this feature extractor: {self.__class__.__name__} was trained using a' F' sampling rate of {self.sampling_rate}. Please make sure that the provided `raw_speech` input' F' was sampled with {self.sampling_rate} and not {sampling_rate}.' ) else: logger.warning( "It is strongly recommended to pass the `sampling_rate` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) _snake_case : Any = isinstance(__UpperCamelCase , np.ndarray ) and len(raw_speech.shape ) > 1 if is_batched_numpy and len(raw_speech.shape ) > 2: raise ValueError(F'Only mono-channel audio is supported for input to {self}' ) _snake_case : Optional[int] = is_batched_numpy or ( isinstance(__UpperCamelCase , (list, tuple) ) and (isinstance(raw_speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: _snake_case : Tuple = [np.asarray([speech] , dtype=np.floataa ).T for speech in raw_speech] elif not is_batched and not isinstance(__UpperCamelCase , np.ndarray ): _snake_case : Optional[int] = np.asarray(__UpperCamelCase , dtype=np.floataa ) elif isinstance(__UpperCamelCase , np.ndarray ) and raw_speech.dtype is np.dtype(np.floataa ): _snake_case : Optional[Any] = raw_speech.astype(np.floataa ) # always return batch if not is_batched: _snake_case : Union[str, Any] = [np.asarray([raw_speech] ).T] _snake_case : Tuple = BatchFeature({"input_features": raw_speech} ) # convert into correct format for padding _snake_case : Optional[Any] = self.pad( __UpperCamelCase , padding=__UpperCamelCase , max_length=max_length if max_length else self.n_samples , truncation=__UpperCamelCase , pad_to_multiple_of=__UpperCamelCase , return_attention_mask=return_attention_mask or do_normalize , ) # zero-mean and unit-variance normalization if do_normalize: _snake_case : Any = self.zero_mean_unit_var_norm( padded_inputs["input_features"] , attention_mask=padded_inputs["attention_mask"] , padding_value=self.padding_value , ) _snake_case : Any = np.stack(padded_inputs["input_features"] , axis=0 ) # make sure list is in array format _snake_case : Any = padded_inputs.get("input_features" ).transpose(2 , 0 , 1 ) _snake_case : int = [self._np_extract_fbank_features(__UpperCamelCase ) for waveform in input_features[0]] if isinstance(input_features[0] , __UpperCamelCase ): _snake_case : List[str] = [np.asarray(__UpperCamelCase , dtype=np.floataa ) for feature in input_features] else: _snake_case : Optional[Any] = input_features if return_attention_mask: # rescale from sample (48000) to feature (3000) _snake_case : Dict = padded_inputs["attention_mask"][:, :: self.hop_length] if return_tensors is not None: _snake_case : Union[str, Any] = padded_inputs.convert_to_tensors(__UpperCamelCase ) return padded_inputs def lowerCamelCase__ ( self ): _snake_case : Tuple = copy.deepcopy(self.__dict__ ) _snake_case : str = self.__class__.__name__ if "mel_filters" in output: del output["mel_filters"] return output
714
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def a__ ( a : List[str] , a : Any ): """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _snake_case : Any = flax_key_tuple[:-1] + ("weight",) _snake_case : str = torch.permute(a , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(a ): # linear layer _snake_case : Optional[int] = flax_key_tuple[:-1] + ("weight",) _snake_case : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _snake_case : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) return flax_key_tuple, flax_tensor def a__ ( a : List[Any] , a : Union[str, Any] , a : List[str] ): """simple docstring""" if "metadata" in layer: _snake_case : Optional[int] = layer.split("metadata" ) _snake_case : Optional[int] = "".join(split_layer[0] )[:-1] _snake_case : int = [tuple(("metadata" + split_layer[1]).split("/" ) )] elif "kvstore" in layer: _snake_case : Any = layer.split("kvstore" ) _snake_case : str = "".join(split_layer[0] )[:-1] _snake_case : Any = [tuple(("kvstore" + split_layer[1]).split("/" ) )] else: _snake_case : List[Any] = layer.split("/" ) _snake_case : Tuple = "/".join(split_layer[:-1] ) _snake_case : int = (split_layer[-1],) if "kvstore/path" in layer: _snake_case : Optional[Any] = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _snake_case : Tuple = "file" else: _snake_case : Optional[int] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def a__ ( a : List[Any] , a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = rename_keys(a ) _snake_case : int = {} for k, v in current_block.items(): _snake_case : Optional[int] = v _snake_case : Optional[int] = new_current_block torch.save(a , a ) def a__ ( a : Dict , a : Tuple , a : List[str] , a : int , a : str = WEIGHTS_NAME ): """simple docstring""" _snake_case : Any = convert_file_size_to_int(a ) _snake_case : Tuple = [] _snake_case : Optional[int] = {} _snake_case : Tuple = 0 _snake_case : Optional[Any] = 0 os.makedirs(a , exist_ok=a ) with gfile.GFile(switch_checkpoint_path + "/checkpoint" , "rb" ) as fp: _snake_case : Any = serialization.msgpack_restore(fp.read() )["optimizer"]["target"] _snake_case : Optional[Any] = flatten_dict(a , sep="/" ) _snake_case : Optional[Any] = {} for layer in checkpoint_info.keys(): _snake_case , _snake_case , _snake_case : int = get_key_and_tensorstore_dict( a , a , a ) if curr_real_layer_name in all_layers: _snake_case : Dict = content else: _snake_case : Tuple = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _snake_case : List[str] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _snake_case : Dict = torch.tensor(a ) _snake_case : Dict = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _snake_case , _snake_case : Optional[int] = rename_base_flax_keys(tuple(key.split("/" ) ) , a ) _snake_case : Optional[Any] = "/".join(a ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _snake_case : Any = os.path.join( a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) del current_block _snake_case : List[Any] = {} _snake_case : str = 0 _snake_case : List[str] = raw_weights.to(getattr(a , a ) ) current_block_size += weight_size total_size += weight_size # Add the last block _snake_case : int = os.path.join(a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(a ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _snake_case : str = {} _snake_case : Any = {} for idx, shard in enumerate(a ): _snake_case : Optional[int] = weights_name.replace( ".bin" , f'-{idx+1:05d}-of-{len(a ):05d}.bin' ) # len(sharded_state_dicts):05d} _snake_case : Dict = os.path.join(a , weights_name.replace(".bin" , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(a , os.path.join(a , a ) ) _snake_case : Dict = shard for key in shard: _snake_case : int = shard_file # Add the metadata _snake_case : List[Any] = {"total_size": total_size} _snake_case : Any = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(a , a ) , "w" , encoding="utf-8" ) as f: _snake_case : Union[str, Any] = json.dumps(a , indent=2 , sort_keys=a ) + "\n" f.write(a ) return metadata, index if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) _a : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def a__ ( ): """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _snake_case : List[str] = SwitchTransformersConfig.from_pretrained("google/switch-base-8" ) config.save_pretrained("/home/arthur_huggingface_co/transformers/switch_converted" ) _snake_case : str = SwitchTransformersForConditionalGeneration.from_pretrained( "/home/arthur_huggingface_co/transformers/switch_converted" , device_map="auto" ) _snake_case : List[Any] = TaTokenizer.from_pretrained("t5-small" ) _snake_case : Optional[Any] = "A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>." _snake_case : Dict = tokenizer(a , return_tensors="pt" ).input_ids _snake_case : List[Any] = model.generate(a , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
87
0
"""simple docstring""" import logging import os from typing import List, TextIO, Union from conllu import parse_incr from utils_ner import InputExample, Split, TokenClassificationTask _a : Tuple = logging.getLogger(__name__) class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_=-1 ): _snake_case : Dict = label_idx def lowerCamelCase__ ( self , snake_case_ , snake_case_ ): if isinstance(snake_case_ , snake_case_ ): _snake_case : List[Any] = mode.value _snake_case : str = os.path.join(snake_case_ , F'{mode}.txt' ) _snake_case : int = 1 _snake_case : List[Any] = [] with open(snake_case_ , encoding="utf-8" ) as f: _snake_case : Optional[int] = [] _snake_case : Tuple = [] for line in f: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": if words: examples.append(InputExample(guid=F'{mode}-{guid_index}' , words=snake_case_ , labels=snake_case_ ) ) guid_index += 1 _snake_case : Optional[int] = [] _snake_case : Any = [] else: _snake_case : Dict = line.split(" " ) words.append(splits[0] ) if len(snake_case_ ) > 1: labels.append(splits[self.label_idx].replace("\n" , "" ) ) else: # Examples could have no label for mode = "test" labels.append("O" ) if words: examples.append(InputExample(guid=F'{mode}-{guid_index}' , words=snake_case_ , labels=snake_case_ ) ) return examples def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[str] = 0 for line in test_input_reader: if line.startswith("-DOCSTART-" ) or line == "" or line == "\n": writer.write(snake_case_ ) if not preds_list[example_id]: example_id += 1 elif preds_list[example_id]: _snake_case : str = line.split()[0] + " " + preds_list[example_id].pop(0 ) + "\n" writer.write(snake_case_ ) else: logger.warning("Maximum sequence length exceeded: No prediction for \'%s\'." , line.split()[0] ) def lowerCamelCase__ ( self , snake_case_ ): if path: with open(snake_case_ , "r" ) as f: _snake_case : Optional[Any] = f.read().splitlines() if "O" not in labels: _snake_case : Any = ["O"] + labels return labels else: return ["O", "B-MISC", "I-MISC", "B-PER", "I-PER", "B-ORG", "I-ORG", "B-LOC", "I-LOC"] class _UpperCAmelCase ( _snake_case): def __init__( self ): super().__init__(label_idx=-2 ) def lowerCamelCase__ ( self , snake_case_ ): if path: with open(snake_case_ , "r" ) as f: _snake_case : Optional[Any] = f.read().splitlines() if "O" not in labels: _snake_case : List[Any] = ["O"] + labels return labels else: return [ "O", "B-ADVP", "B-INTJ", "B-LST", "B-PRT", "B-NP", "B-SBAR", "B-VP", "B-ADJP", "B-CONJP", "B-PP", "I-ADVP", "I-INTJ", "I-LST", "I-PRT", "I-NP", "I-SBAR", "I-VP", "I-ADJP", "I-CONJP", "I-PP", ] class _UpperCAmelCase ( _snake_case): def lowerCamelCase__ ( self , snake_case_ , snake_case_ ): if isinstance(snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = mode.value _snake_case : Tuple = os.path.join(snake_case_ , F'{mode}.txt' ) _snake_case : Any = 1 _snake_case : List[str] = [] with open(snake_case_ , encoding="utf-8" ) as f: for sentence in parse_incr(snake_case_ ): _snake_case : Tuple = [] _snake_case : int = [] for token in sentence: words.append(token["form"] ) labels.append(token["upos"] ) assert len(snake_case_ ) == len(snake_case_ ) if words: examples.append(InputExample(guid=F'{mode}-{guid_index}' , words=snake_case_ , labels=snake_case_ ) ) guid_index += 1 return examples def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = 0 for sentence in parse_incr(snake_case_ ): _snake_case : List[str] = preds_list[example_id] _snake_case : int = "" for token in sentence: out += F'{token["form"]} ({token["upos"]}|{s_p.pop(0 )}) ' out += "\n" writer.write(snake_case_ ) example_id += 1 def lowerCamelCase__ ( self , snake_case_ ): if path: with open(snake_case_ , "r" ) as f: return f.read().splitlines() else: return [ "ADJ", "ADP", "ADV", "AUX", "CCONJ", "DET", "INTJ", "NOUN", "NUM", "PART", "PRON", "PROPN", "PUNCT", "SCONJ", "SYM", "VERB", "X", ]
715
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
87
0
"""simple docstring""" from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxSeqaSeqConfigWithPast from ...utils import logging _a : Tuple = logging.get_logger(__name__) _a : Optional[int] = { """google/umt5-small""": """https://huggingface.co/google/umt5-small/resolve/main/config.json""", # See all umt5 models at https://huggingface.co/models?filter=umt5 } class _UpperCAmelCase ( __UpperCAmelCase): __lowercase : List[str] = """umt5""" __lowercase : Union[str, Any] = ["""past_key_values"""] def __init__( self , snake_case_=25_01_12 , snake_case_=5_12 , snake_case_=64 , snake_case_=10_24 , snake_case_=8 , snake_case_=None , snake_case_=6 , snake_case_=32 , snake_case_=1_28 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=1.0 , snake_case_="gated-gelu" , snake_case_=True , snake_case_=True , snake_case_="T5Tokenizer" , snake_case_=True , snake_case_=0 , snake_case_=1 , snake_case_=0 , **snake_case_ , ): super().__init__( is_encoder_decoder=snake_case_ , tokenizer_class=snake_case_ , tie_word_embeddings=snake_case_ , pad_token_id=snake_case_ , eos_token_id=snake_case_ , decoder_start_token_id=snake_case_ , **snake_case_ , ) _snake_case : str = vocab_size _snake_case : int = d_model _snake_case : str = d_kv _snake_case : Optional[Any] = d_ff _snake_case : List[str] = num_layers _snake_case : Optional[Any] = ( num_decoder_layers if num_decoder_layers is not None else self.num_layers ) # default = symmetry _snake_case : Optional[Any] = num_heads _snake_case : List[Any] = relative_attention_num_buckets _snake_case : str = relative_attention_max_distance _snake_case : Tuple = dropout_rate _snake_case : List[str] = layer_norm_epsilon _snake_case : Dict = initializer_factor _snake_case : str = feed_forward_proj _snake_case : Dict = use_cache _snake_case : int = self.feed_forward_proj.split("-" ) _snake_case : Any = act_info[-1] _snake_case : Any = act_info[0] == "gated" if len(snake_case_ ) > 1 and act_info[0] != "gated" or len(snake_case_ ) > 2: raise ValueError( F'`feed_forward_proj`: {feed_forward_proj} is not a valid activation function of the dense layer.' "Please make sure `feed_forward_proj` is of the format `gated-{ACT_FN}` or `{ACT_FN}`, e.g. " "\'gated-gelu\' or \'relu\'" ) if feed_forward_proj == "gated-gelu": _snake_case : Dict = "gelu_new" @property def lowerCamelCase__ ( self ): return self.d_model @property def lowerCamelCase__ ( self ): return self.num_heads @property def lowerCamelCase__ ( self ): return self.num_layers class _UpperCAmelCase ( __UpperCAmelCase): @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.inputs def lowerCamelCase__ ( self ): _snake_case : Optional[int] = { "input_ids": {0: "batch", 1: "encoder_sequence"}, "attention_mask": {0: "batch", 1: "encoder_sequence"}, } if self.use_past: _snake_case : int = "past_encoder_sequence + sequence" _snake_case : List[str] = {0: "batch"} _snake_case : Tuple = {0: "batch", 1: "past_decoder_sequence + sequence"} else: _snake_case : List[Any] = {0: "batch", 1: "decoder_sequence"} _snake_case : Optional[Any] = {0: "batch", 1: "decoder_sequence"} if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction="inputs" ) return common_inputs @property # Copied from transformers.models.t5.configuration_t5.T5OnnxConfig.default_onnx_opset def lowerCamelCase__ ( self ): return 13 @property def lowerCamelCase__ ( self ): return 5E-4
716
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, UNetaDConditionModel, VideoToVideoSDPipeline, ) from diffusers.utils import floats_tensor, is_xformers_available, skip_mps from diffusers.utils.testing_utils import enable_full_determinism, slow, torch_device from ..pipeline_params import ( TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( __lowerCAmelCase , unittest.TestCase): __lowercase : str = VideoToVideoSDPipeline __lowercase : int = TEXT_GUIDED_IMAGE_VARIATION_PARAMS.union({"""video"""}) - {"image", "width", "height"} __lowercase : str = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({"""video"""}) - {"image"} __lowercase : Any = PipelineTesterMixin.required_optional_params - {"latents"} __lowercase : List[Any] = False # No `output_type`. __lowercase : List[Any] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : int = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : Optional[int] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=lowerCamelCase__ , set_alpha_to_one=lowerCamelCase__ , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Dict = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : int = CLIPTextModel(lowerCamelCase__ ) _snake_case : Union[str, Any] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Optional[int] = { '''unet''': unet, '''scheduler''': scheduler, '''vae''': vae, '''text_encoder''': text_encoder, '''tokenizer''': tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): _snake_case : str = floats_tensor((1, 3, 3, 32, 32) , rng=random.Random(lowerCamelCase__ ) ).to(lowerCamelCase__ ) if str(lowerCamelCase__ ).startswith("mps" ): _snake_case : List[str] = torch.manual_seed(lowerCamelCase__ ) else: _snake_case : List[str] = torch.Generator(device=lowerCamelCase__ ).manual_seed(lowerCamelCase__ ) _snake_case : Tuple = { '''prompt''': '''A painting of a squirrel eating a burger''', '''video''': video, '''generator''': generator, '''num_inference_steps''': 2, '''guidance_scale''': 6.0, '''output_type''': '''pt''', } return inputs def lowerCamelCase__ ( self ): _snake_case : List[str] = '''cpu''' # ensure determinism for the device-dependent torch.Generator _snake_case : List[Any] = self.get_dummy_components() _snake_case : Any = VideoToVideoSDPipeline(**lowerCamelCase__ ) _snake_case : Optional[int] = sd_pipe.to(lowerCamelCase__ ) sd_pipe.set_progress_bar_config(disable=lowerCamelCase__ ) _snake_case : Any = self.get_dummy_inputs(lowerCamelCase__ ) _snake_case : str = '''np''' _snake_case : Any = sd_pipe(**lowerCamelCase__ ).frames _snake_case : Dict = frames[0][-3:, -3:, -1] assert frames[0].shape == (32, 32, 3) _snake_case : Any = np.array([1_06, 1_17, 1_13, 1_74, 1_37, 1_12, 1_48, 1_51, 1_31] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=lowerCamelCase__ , expected_max_diff=5E-3 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : Any = VideoToVideoSDPipeline.from_pretrained("cerspense/zeroscope_v2_XL" , torch_dtype=torch.floataa ) pipe.enable_model_cpu_offload() # 10 frames _snake_case : Dict = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Optional[int] = torch.randn((1, 10, 3, 10_24, 5_76) , generator=lowerCamelCase__ ) _snake_case : List[Any] = video.to("cuda" ) _snake_case : Optional[int] = '''Spiderman is surfing''' _snake_case : Dict = pipe(lowerCamelCase__ , video=lowerCamelCase__ , generator=lowerCamelCase__ , num_inference_steps=3 , output_type="pt" ).frames _snake_case : Any = np.array([-1.0458984, -1.1279297, -0.9663086, -0.91503906, -0.75097656] ) assert np.abs(video_frames.cpu().numpy()[0, 0, 0, 0, -5:] - expected_array ).sum() < 1E-2
717
"""simple docstring""" def a__ ( a : list , a : int , a : int = 0 , a : int = 0 ): """simple docstring""" _snake_case : Optional[int] = right or len(a ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(a , a , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import unittest from transformers import load_tool from .test_tools_common import ToolTesterMixin class _UpperCAmelCase ( unittest.TestCase , UpperCamelCase_): def lowerCamelCase__ ( self ): _snake_case : int = load_tool("text-classification" ) self.tool.setup() _snake_case : Union[str, Any] = load_tool("text-classification" , remote=snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : int = self.tool("That\'s quite cool" , ["positive", "negative"] ) self.assertEqual(snake_case_ , "positive" ) def lowerCamelCase__ ( self ): _snake_case : int = self.remote_tool("That\'s quite cool" , ["positive", "negative"] ) self.assertEqual(snake_case_ , "positive" ) def lowerCamelCase__ ( self ): _snake_case : str = self.tool(text="That\'s quite cool" , labels=["positive", "negative"] ) self.assertEqual(snake_case_ , "positive" ) def lowerCamelCase__ ( self ): _snake_case : str = self.remote_tool(text="That\'s quite cool" , labels=["positive", "negative"] ) self.assertEqual(snake_case_ , "positive" )
718
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ ): _snake_case , _snake_case : Dict = text, pattern _snake_case , _snake_case : int = len(snake_case_ ), len(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCamelCase__ ( self ): # searches pattern in text and returns index positions _snake_case : List[str] = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case : Union[str, Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: _snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) _snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _a : List[Any] = """ABAABA""" _a : str = """AB""" _a : List[Any] = BoyerMooreSearch(text, pattern) _a : Any = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
87
0
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from transformers import ( VisualBertConfig, VisualBertForMultipleChoice, VisualBertForPreTraining, VisualBertForQuestionAnswering, VisualBertForVisualReasoning, ) from transformers.utils import logging logging.set_verbosity_info() _a : List[str] = logging.get_logger(__name__) _a : Tuple = [ ("""bert.bert""", """visual_bert"""), ("""bert.cls""", """cls"""), ("""bert.classifier""", """cls"""), ("""token_type_embeddings_visual""", """visual_token_type_embeddings"""), ("""position_embeddings_visual""", """visual_position_embeddings"""), ("""projection""", """visual_projection"""), ] _a : Optional[int] = [ """nlvr2_coco_pre_trained.th""", """nlvr2_fine_tuned.th""", """nlvr2_pre_trained.th""", """vcr_coco_pre_train.th""", """vcr_fine_tune.th""", """vcr_pre_train.th""", """vqa_coco_pre_trained.th""", """vqa_fine_tuned.th""", """vqa_pre_trained.th""", ] def a__ ( a : Optional[int] ): """simple docstring""" _snake_case : Any = torch.load(UpperCamelCase__ , map_location="cpu" ) return sd def a__ ( a : Optional[int] , a : Union[str, Any] , a : Dict=rename_keys_prefix ): """simple docstring""" _snake_case : Tuple = OrderedDict() _snake_case : List[Any] = torch.arange(config.max_position_embeddings ).expand((1, -1) ) # detector_d = OrderedDict() for key in d: if "detector" in key: # detector_d[key.replace('detector.','')] = d[key] continue _snake_case : Union[str, Any] = key for name_pair in rename_keys_prefix: _snake_case : Optional[int] = new_key.replace(name_pair[0] , name_pair[1] ) _snake_case : List[str] = d[key] if key == "bert.cls.predictions.decoder.weight": # Old bert code didn't have `decoder.bias`, but was added separately _snake_case : Tuple = new_d["cls.predictions.bias"] return new_d @torch.no_grad() def a__ ( a : Any , a : Dict ): """simple docstring""" assert ( checkpoint_path.split("/" )[-1] in ACCEPTABLE_CHECKPOINTS ), f'The checkpoint provided must be in {ACCEPTABLE_CHECKPOINTS}.' # Get Config if "pre" in checkpoint_path: _snake_case : Any = "pretraining" if "vcr" in checkpoint_path: _snake_case : List[Any] = {"visual_embedding_dim": 512} elif "vqa_advanced" in checkpoint_path: _snake_case : Any = {"visual_embedding_dim": 2_048} elif "vqa" in checkpoint_path: _snake_case : Any = {"visual_embedding_dim": 2_048} elif "nlvr" in checkpoint_path: _snake_case : Optional[Any] = {"visual_embedding_dim": 1_024} else: raise NotImplementedError(f'No implementation found for `{checkpoint_path}`.' ) else: if "vcr" in checkpoint_path: _snake_case : Union[str, Any] = {"visual_embedding_dim": 512} _snake_case : Tuple = "multichoice" elif "vqa_advanced" in checkpoint_path: _snake_case : List[str] = {"visual_embedding_dim": 2_048} _snake_case : List[str] = "vqa_advanced" elif "vqa" in checkpoint_path: _snake_case : List[Any] = {"visual_embedding_dim": 2_048, "num_labels": 3_129} _snake_case : List[str] = "vqa" elif "nlvr" in checkpoint_path: _snake_case : List[Any] = { "visual_embedding_dim": 1_024, "num_labels": 2, } _snake_case : Any = "nlvr" _snake_case : Tuple = VisualBertConfig(**UpperCamelCase__ ) # Load State Dict _snake_case : Dict = load_state_dict(UpperCamelCase__ ) _snake_case : Any = get_new_dict(UpperCamelCase__ , UpperCamelCase__ ) if model_type == "pretraining": _snake_case : List[Any] = VisualBertForPreTraining(UpperCamelCase__ ) elif model_type == "vqa": _snake_case : Optional[Any] = VisualBertForQuestionAnswering(UpperCamelCase__ ) elif model_type == "nlvr": _snake_case : Dict = VisualBertForVisualReasoning(UpperCamelCase__ ) elif model_type == "multichoice": _snake_case : Tuple = VisualBertForMultipleChoice(UpperCamelCase__ ) model.load_state_dict(UpperCamelCase__ ) # Save Checkpoints Path(UpperCamelCase__ ).mkdir(exist_ok=UpperCamelCase__ ) model.save_pretrained(UpperCamelCase__ ) if __name__ == "__main__": _a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument("""orig_checkpoint_path""", type=str, help="""A path to .th on local filesystem.""") parser.add_argument("""pytorch_dump_folder_path""", type=str, help="""Path to the output PyTorch model.""") _a : List[Any] = parser.parse_args() convert_visual_bert_checkpoint(args.orig_checkpoint_path, args.pytorch_dump_folder_path)
719
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
87
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import is_tf_available from transformers.testing_utils import require_sentencepiece, require_tf, require_tokenizers, slow if is_tf_available(): import numpy as np import tensorflow as tf from transformers import TFXLMRobertaModel @require_tf @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Dict = TFXLMRobertaModel.from_pretrained("jplu/tf-xlm-roberta-base" ) _snake_case : Optional[Any] = { "input_ids": tf.convert_to_tensor([[0, 26_46, 1_02_69, 83, 9_99_42, 2]] , dtype=tf.intaa ), # "My dog is cute" "attention_mask": tf.convert_to_tensor([[1, 1, 1, 1, 1, 1]] , dtype=tf.intaa ), } _snake_case : Tuple = model(snake_case_ )["last_hidden_state"] _snake_case : Tuple = tf.TensorShape((1, 6, 7_68) ) self.assertEqual(output.shape , snake_case_ ) # compare the actual values for a slice. _snake_case : str = tf.convert_to_tensor( [ [ [0.0681762, 0.10894451, 0.06772504], [-0.06423668, 0.02366615, 0.04329344], [-0.06057295, 0.09974135, -0.00070584], ] ] , dtype=tf.floataa , ) self.assertTrue(np.allclose(output[:, :3, :3].numpy() , expected_slice.numpy() , atol=1E-4 ) )
720
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" def a__ ( ): """simple docstring""" _snake_case : int = 0 for i in range(1 , 1_001 ): total += i**i return str(a )[-10:] if __name__ == "__main__": print(solution())
721
"""simple docstring""" import argparse import json import subprocess def a__ ( a : Optional[Any] , a : Optional[int] ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[Any] = ( f'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _snake_case : Dict = subprocess.run(a , shell=a , stdout=subprocess.PIPE ) _snake_case : Tuple = output.stdout.decode("utf-8" ) _snake_case : List[str] = json.loads(a ) _snake_case : Any = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(a ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(a ) ) if len(a ) > 0: _snake_case : Any = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def a__ ( a : Optional[int] ): """simple docstring""" return values.split("," ) _a : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _a : List[str] = parser.parse_args() get_runner_status(args.target_runners, args.token)
87
0
from collections import OrderedDict from typing import Any, List, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast, PatchingSpec from ...utils import logging _a : List[str] = logging.get_logger(__name__) _a : Tuple = { """EleutherAI/gpt-j-6B""": """https://huggingface.co/EleutherAI/gpt-j-6B/resolve/main/config.json""", # See all GPT-J models at https://huggingface.co/models?filter=gpt_j } class _UpperCAmelCase ( _snake_case): __lowercase : List[str] = """gptj""" __lowercase : Dict = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , snake_case_=5_04_00 , snake_case_=20_48 , snake_case_=40_96 , snake_case_=28 , snake_case_=16 , snake_case_=64 , snake_case_=None , snake_case_="gelu_new" , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_=True , snake_case_=5_02_56 , snake_case_=5_02_56 , snake_case_=False , **snake_case_ , ): _snake_case : int = vocab_size _snake_case : Union[str, Any] = n_positions _snake_case : int = n_embd _snake_case : int = n_layer _snake_case : str = n_head _snake_case : Tuple = n_inner _snake_case : Union[str, Any] = rotary_dim _snake_case : List[Any] = activation_function _snake_case : Dict = resid_pdrop _snake_case : Optional[Any] = embd_pdrop _snake_case : int = attn_pdrop _snake_case : Any = layer_norm_epsilon _snake_case : Optional[int] = initializer_range _snake_case : int = use_cache _snake_case : Any = bos_token_id _snake_case : List[Any] = eos_token_id super().__init__( bos_token_id=snake_case_ , eos_token_id=snake_case_ , tie_word_embeddings=snake_case_ , **snake_case_ ) class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ = "default" , snake_case_ = None , snake_case_ = False , ): super().__init__(snake_case_ , task=snake_case_ , patching_specs=snake_case_ , use_past=snake_case_ ) if not getattr(self._config , "pad_token_id" , snake_case_ ): # TODO: how to do that better? _snake_case : List[Any] = 0 @property def lowerCamelCase__ ( self ): _snake_case : List[Any] = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(snake_case_ , direction="inputs" ) _snake_case : Union[str, Any] = {0: "batch", 1: "past_sequence + sequence"} else: _snake_case : Optional[Any] = {0: "batch", 1: "sequence"} return common_inputs @property def lowerCamelCase__ ( self ): return self._config.n_layer @property def lowerCamelCase__ ( self ): return self._config.n_head def lowerCamelCase__ ( self , snake_case_ , snake_case_ = -1 , snake_case_ = -1 , snake_case_ = False , snake_case_ = None , ): _snake_case : List[Any] = super(snake_case_ , self ).generate_dummy_inputs( snake_case_ , batch_size=snake_case_ , seq_length=snake_case_ , is_pair=snake_case_ , framework=snake_case_ ) # We need to order the input in the way they appears in the forward() _snake_case : Optional[Any] = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch _snake_case : List[str] = common_inputs["input_ids"].shape # Not using the same length for past_key_values _snake_case : Tuple = seqlen + 2 _snake_case : Tuple = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) _snake_case : str = [ (torch.zeros(snake_case_ ), torch.zeros(snake_case_ )) for _ in range(self.num_layers ) ] _snake_case : List[Any] = common_inputs["attention_mask"] if self.use_past: _snake_case : int = ordered_inputs["attention_mask"].dtype _snake_case : Optional[Any] = torch.cat( [ordered_inputs["attention_mask"], torch.ones(snake_case_ , snake_case_ , dtype=snake_case_ )] , dim=1 ) return ordered_inputs @property def lowerCamelCase__ ( self ): return 13
700
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
87
0
import importlib import math import os from dataclasses import dataclass from enum import Enum from typing import Any, Dict, Optional, Tuple, Union import flax import jax.numpy as jnp from ..utils import BaseOutput _a : Dict = """scheduler_config.json""" class _UpperCAmelCase ( _snake_case): __lowercase : Tuple = 1 __lowercase : Any = 2 __lowercase : Optional[int] = 3 __lowercase : Optional[int] = 4 __lowercase : int = 5 @dataclass class _UpperCAmelCase ( _snake_case): __lowercase : jnp.ndarray class _UpperCAmelCase : __lowercase : Dict = SCHEDULER_CONFIG_NAME __lowercase : Optional[int] = ["""dtype"""] __lowercase : List[Any] = [] __lowercase : List[Any] = True @classmethod def lowerCamelCase__ ( cls , snake_case_ = None , snake_case_ = None , snake_case_=False , **snake_case_ , ): _snake_case : Optional[int] = cls.load_config( pretrained_model_name_or_path=snake_case_ , subfolder=snake_case_ , return_unused_kwargs=snake_case_ , **snake_case_ , ) _snake_case : Dict = cls.from_config(snake_case_ , return_unused_kwargs=snake_case_ , **snake_case_ ) if hasattr(snake_case_ , "create_state" ) and getattr(snake_case_ , "has_state" , snake_case_ ): _snake_case : Optional[Any] = scheduler.create_state() if return_unused_kwargs: return scheduler, state, unused_kwargs return scheduler, state def lowerCamelCase__ ( self , snake_case_ , snake_case_ = False , **snake_case_ ): self.save_config(save_directory=snake_case_ , push_to_hub=snake_case_ , **snake_case_ ) @property def lowerCamelCase__ ( self ): return self._get_compatibles() @classmethod def lowerCamelCase__ ( cls ): _snake_case : Union[str, Any] = list(set([cls.__name__] + cls._compatibles ) ) _snake_case : List[Any] = importlib.import_module(__name__.split("." )[0] ) _snake_case : Optional[int] = [ getattr(snake_case_ , snake_case_ ) for c in compatible_classes_str if hasattr(snake_case_ , snake_case_ ) ] return compatible_classes def a__ ( a : jnp.ndarray , a : Tuple[int] ): """simple docstring""" assert len(a ) >= x.ndim return jnp.broadcast_to(x.reshape(x.shape + (1,) * (len(a ) - x.ndim) ) , a ) def a__ ( a : int , a : str=0.999 , a : Optional[Any]=jnp.floataa ): """simple docstring""" def alpha_bar(a : Optional[int] ): return math.cos((time_step + 0.008) / 1.008 * math.pi / 2 ) ** 2 _snake_case : Tuple = [] for i in range(a ): _snake_case : Tuple = i / num_diffusion_timesteps _snake_case : Any = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar(a ) / alpha_bar(a ) , a ) ) return jnp.array(a , dtype=a ) @flax.struct.dataclass class _UpperCAmelCase : __lowercase : jnp.ndarray __lowercase : jnp.ndarray __lowercase : jnp.ndarray @classmethod def lowerCamelCase__ ( cls , snake_case_ ): _snake_case : int = scheduler.config if config.trained_betas is not None: _snake_case : str = jnp.asarray(config.trained_betas , dtype=scheduler.dtype ) elif config.beta_schedule == "linear": _snake_case : Optional[Any] = jnp.linspace(config.beta_start , config.beta_end , config.num_train_timesteps , dtype=scheduler.dtype ) elif config.beta_schedule == "scaled_linear": # this schedule is very specific to the latent diffusion model. _snake_case : Optional[Any] = ( jnp.linspace( config.beta_start**0.5 , config.beta_end**0.5 , config.num_train_timesteps , dtype=scheduler.dtype ) ** 2 ) elif config.beta_schedule == "squaredcos_cap_v2": # Glide cosine schedule _snake_case : Optional[Any] = betas_for_alpha_bar(config.num_train_timesteps , dtype=scheduler.dtype ) else: raise NotImplementedError( F'beta_schedule {config.beta_schedule} is not implemented for scheduler {scheduler.__class__.__name__}' ) _snake_case : List[str] = 1.0 - betas _snake_case : List[Any] = jnp.cumprod(snake_case_ , axis=0 ) return cls( alphas=snake_case_ , betas=snake_case_ , alphas_cumprod=snake_case_ , ) def a__ ( a : CommonSchedulerState , a : jnp.ndarray , a : jnp.ndarray , a : jnp.ndarray ): """simple docstring""" _snake_case : Optional[int] = state.alphas_cumprod _snake_case : Union[str, Any] = alphas_cumprod[timesteps] ** 0.5 _snake_case : Tuple = sqrt_alpha_prod.flatten() _snake_case : Any = broadcast_to_shape_from_left(a , original_samples.shape ) _snake_case : Any = (1 - alphas_cumprod[timesteps]) ** 0.5 _snake_case : List[Any] = sqrt_one_minus_alpha_prod.flatten() _snake_case : List[str] = broadcast_to_shape_from_left(a , original_samples.shape ) return sqrt_alpha_prod, sqrt_one_minus_alpha_prod def a__ ( a : CommonSchedulerState , a : jnp.ndarray , a : jnp.ndarray , a : jnp.ndarray ): """simple docstring""" _snake_case : Any = get_sqrt_alpha_prod(a , a , a , a ) _snake_case : List[Any] = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples def a__ ( a : CommonSchedulerState , a : jnp.ndarray , a : jnp.ndarray , a : jnp.ndarray ): """simple docstring""" _snake_case : int = get_sqrt_alpha_prod(a , a , a , a ) _snake_case : Any = sqrt_alpha_prod * noise - sqrt_one_minus_alpha_prod * sample return velocity
701
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import numpy as np from cva import destroyAllWindows, imread, imshow, waitKey class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ , snake_case_ ): if dst_width < 0 or dst_height < 0: raise ValueError("Destination width/height should be > 0" ) _snake_case : Union[str, Any] = img _snake_case : List[str] = img.shape[1] _snake_case : List[str] = img.shape[0] _snake_case : str = dst_width _snake_case : Optional[int] = dst_height _snake_case : Optional[Any] = self.src_w / self.dst_w _snake_case : Any = self.src_h / self.dst_h _snake_case : List[str] = ( np.ones((self.dst_h, self.dst_w, 3) , np.uinta ) * 2_55 ) def lowerCamelCase__ ( self ): for i in range(self.dst_h ): for j in range(self.dst_w ): _snake_case : List[str] = self.img[self.get_y(snake_case_ )][self.get_x(snake_case_ )] def lowerCamelCase__ ( self , snake_case_ ): return int(self.ratio_x * x ) def lowerCamelCase__ ( self , snake_case_ ): return int(self.ratio_y * y ) if __name__ == "__main__": _a : Dict = 800, 600 _a : Any = imread("""image_data/lena.jpg""", 1) _a : Dict = NearestNeighbour(im, dst_w, dst_h) n.process() imshow( f'Image resized from: {im.shape[1]}x{im.shape[0]} to {dst_w}x{dst_h}', n.output ) waitKey(0) destroyAllWindows()
702
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Any = TextToVideoSDPipeline __lowercase : str = TEXT_TO_IMAGE_PARAMS __lowercase : int = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __lowercase : Optional[int] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) _snake_case : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : Tuple = CLIPTextModel(snake_case_ ) _snake_case : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : str = torch.manual_seed(snake_case_ ) else: _snake_case : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def lowerCamelCase__ ( self ): _snake_case : int = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : Optional[Any] = self.get_dummy_components() _snake_case : Tuple = TextToVideoSDPipeline(**snake_case_ ) _snake_case : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ ) _snake_case : Union[str, Any] = "np" _snake_case : Dict = sd_pipe(**snake_case_ ).frames _snake_case : Any = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _snake_case : Dict = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=1E-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) _snake_case : int = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _snake_case : Tuple = pipe.to("cuda" ) _snake_case : List[Any] = "Spiderman is surfing" _snake_case : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : int = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="pt" ).frames _snake_case : int = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def lowerCamelCase__ ( self ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) _snake_case : str = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : int = pipe.to("cuda" ) _snake_case : Any = "Spiderman is surfing" _snake_case : str = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Any = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type="pt" ).frames _snake_case : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : str = {"""configuration_yolos""": ["""YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """YolosConfig""", """YolosOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""YolosFeatureExtractor"""] _a : List[Any] = ["""YolosImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST""", """YolosForObjectDetection""", """YolosModel""", """YolosPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
703
"""simple docstring""" from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _UpperCAmelCase ( _snake_case): __lowercase : int = """EncodecFeatureExtractor""" __lowercase : str = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) _snake_case : Dict = self.feature_extractor _snake_case : Any = False def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_=True ): return self.tokenizer.get_decoder_prompt_ids(task=snake_case_ , language=snake_case_ , no_timestamps=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*snake_case_ , **snake_case_ ) _snake_case : str = kwargs.pop("audio" , snake_case_ ) _snake_case : Optional[int] = kwargs.pop("sampling_rate" , snake_case_ ) _snake_case : Optional[Any] = kwargs.pop("text" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Any = args[0] _snake_case : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: _snake_case : Any = self.tokenizer(snake_case_ , **snake_case_ ) if audio is not None: _snake_case : Any = self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) if audio is None: return inputs elif text is None: return audio_inputs else: _snake_case : str = audio_inputs["input_values"] if "padding_mask" in audio_inputs: _snake_case : List[str] = audio_inputs["padding_mask"] return inputs def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): _snake_case : Tuple = kwargs.pop("audio" , snake_case_ ) _snake_case : List[str] = kwargs.pop("padding_mask" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Dict = args[1:] if audio_values is not None: return self._decode_audio(snake_case_ , padding_mask=snake_case_ ) else: return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Optional[int] = to_numpy(snake_case_ ) _snake_case , _snake_case , _snake_case : Tuple = audio_values.shape if padding_mask is None: return list(snake_case_ ) _snake_case : Optional[int] = to_numpy(snake_case_ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _snake_case : Any = seq_len - padding_mask.shape[-1] _snake_case : Optional[Any] = 1 - self.feature_extractor.padding_value _snake_case : Optional[int] = np.pad(snake_case_ , ((0, 0), (0, difference)) , "constant" , constant_values=snake_case_ ) _snake_case : Any = audio_values.tolist() for i in range(snake_case_ ): _snake_case : Tuple = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _snake_case : Tuple = sliced_audio.reshape(snake_case_ , -1 ) return audio_values
87
0
def a__ ( a : list[list[float]] ): """simple docstring""" _snake_case : list[list[float]] = [] for data in source_data: for i, el in enumerate(a ): if len(a ) < i + 1: data_lists.append([] ) data_lists[i].append(float(a ) ) return data_lists def a__ ( a : list[list[float]] , a : list[int] ): """simple docstring""" _snake_case : list[list[float]] = [] for dlist, weight in zip(a , a ): _snake_case : Optional[Any] = min(a ) _snake_case : Tuple = max(a ) _snake_case : list[float] = [] # for weight 0 score is 1 - actual score if weight == 0: for item in dlist: try: score.append(1 - ((item - mind) / (maxd - mind)) ) except ZeroDivisionError: score.append(1 ) elif weight == 1: for item in dlist: try: score.append((item - mind) / (maxd - mind) ) except ZeroDivisionError: score.append(0 ) # weight not 0 or 1 else: _snake_case : Optional[int] = f'Invalid weight of {weight:f} provided' raise ValueError(a ) score_lists.append(a ) return score_lists def a__ ( a : list[list[float]] ): """simple docstring""" _snake_case : list[float] = [0 for i in range(len(score_lists[0] ) )] for slist in score_lists: for j, ele in enumerate(a ): _snake_case : Union[str, Any] = final_scores[j] + ele return final_scores def a__ ( a : list[list[float]] , a : list[int] ): """simple docstring""" _snake_case : List[str] = get_data(a ) _snake_case : List[Any] = calculate_each_score(a , a ) _snake_case : Optional[Any] = generate_final_scores(a ) # append scores to source data for i, ele in enumerate(a ): source_data[i].append(a ) return source_data
704
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : str = {"""configuration_yolos""": ["""YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """YolosConfig""", """YolosOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""YolosFeatureExtractor"""] _a : List[Any] = ["""YolosImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST""", """YolosForObjectDetection""", """YolosModel""", """YolosPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _a : List[str] = { """configuration_convnext""": ["""CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """ConvNextConfig""", """ConvNextOnnxConfig"""] } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = ["""ConvNextFeatureExtractor"""] _a : Optional[Any] = ["""ConvNextImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : str = [ """CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST""", """ConvNextForImageClassification""", """ConvNextModel""", """ConvNextPreTrainedModel""", """ConvNextBackbone""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ """TFConvNextForImageClassification""", """TFConvNextModel""", """TFConvNextPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_convnext import CONVNEXT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvNextConfig, ConvNextOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_convnext import ConvNextFeatureExtractor from .image_processing_convnext import ConvNextImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convnext import ( CONVNEXT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvNextBackbone, ConvNextForImageClassification, ConvNextModel, ConvNextPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convnext import TFConvNextForImageClassification, TFConvNextModel, TFConvNextPreTrainedModel else: import sys _a : Dict = _LazyModule(__name__, globals()["""__file__"""], _import_structure)
705
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = dataset _snake_case : str = process _snake_case : int = params def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): _snake_case : Union[str, Any] = self.dataset[i] _snake_case : Optional[Any] = self.process(snake_case_ , **self.params ) return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): _snake_case : Union[str, Any] = loader _snake_case : Tuple = infer _snake_case : List[Any] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _snake_case : int = None _snake_case : int = loader_batch_size # Internal bookkeeping _snake_case : Any = None _snake_case : Dict = None def __len__( self ): return len(self.loader ) def __iter__( self ): _snake_case : int = iter(self.loader ) return self def lowerCamelCase__ ( self ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _snake_case : List[Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _snake_case : int = {} for k, element in self._loader_batch_data.items(): if isinstance(snake_case_ , snake_case_ ): # Convert ModelOutput to tuple first _snake_case : Tuple = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _snake_case : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(snake_case_ , snake_case_ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _snake_case : Tuple = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _snake_case : Tuple = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : List[Any] = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : Union[str, Any] = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _snake_case : List[Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _snake_case : int = self._loader_batch_data.__class__(snake_case_ ) self._loader_batch_index += 1 return result def lowerCamelCase__ ( self ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _snake_case : Tuple = next(self.iterator ) _snake_case : Any = self.infer(snake_case_ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Optional[int] = list(processed.keys() )[0] _snake_case : List[str] = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = len(snake_case_ ) else: _snake_case : Optional[int] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Union[str, Any] = observed_batch_size # Setting internal index to unwrap the batch _snake_case : str = processed _snake_case : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): super().__init__(snake_case_ , snake_case_ , snake_case_ ) def __iter__( self ): _snake_case : Tuple = iter(self.loader ) _snake_case : List[Any] = None return self def lowerCamelCase__ ( self ): if self.subiterator is None: _snake_case : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _snake_case : Union[str, Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _snake_case : str = self.infer(next(self.iterator ) , **self.params ) _snake_case : Tuple = next(self.subiterator ) return processed class _UpperCAmelCase ( _snake_case): def __iter__( self ): _snake_case : Optional[Any] = iter(self.loader ) return self def lowerCamelCase__ ( self ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _snake_case : Optional[Any] = False _snake_case : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : str = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator while not is_last: _snake_case : List[str] = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Tuple = list(processed.keys() )[0] _snake_case : Tuple = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Any = len(snake_case_ ) else: _snake_case : List[Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Dict = observed_batch_size _snake_case : List[Any] = processed _snake_case : List[str] = 0 while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : int = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator else: _snake_case : Dict = processed _snake_case : Dict = item.pop("is_last" ) accumulator.append(snake_case_ ) return accumulator class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ ): _snake_case : str = dataset _snake_case : Any = key def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return self.dataset[i][self.key] class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = dataset _snake_case : Any = keya _snake_case : int = keya def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
87
0
"""simple docstring""" import re import subprocess import sys lowerCamelCase_ : int = subprocess.check_output("""git merge-base main HEAD""".split()).decode("""utf-8""") lowerCamelCase_ : List[Any] = subprocess.check_output(f'git diff --name-only {fork_point_sha}'.split()).decode("""utf-8""").split() lowerCamelCase_ : Union[str, Any] = """|""".join(sys.argv[1:]) lowerCamelCase_ : Dict = re.compile(rf'^({joined_dirs}).*?\.py$') lowerCamelCase_ : int = [x for x in modified_files if regex.match(x)] print(""" """.join(relevant_modified_files), end="""""")
706
"""simple docstring""" def a__ ( a : int ): """simple docstring""" if not isinstance(a , a ): raise TypeError("Input value must be an 'int' type" ) _snake_case : Union[str, Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import importlib.metadata import warnings from copy import deepcopy from packaging import version from ..utils import logging from .import_utils import is_accelerate_available, is_bitsandbytes_available if is_bitsandbytes_available(): import bitsandbytes as bnb import torch import torch.nn as nn from ..pytorch_utils import ConvaD if is_accelerate_available(): from accelerate import init_empty_weights from accelerate.utils import find_tied_parameters _a : Tuple = logging.get_logger(__name__) def a__ ( a : Dict , a : str , a : Optional[int] , a : Optional[Any]=None , a : Optional[Any]=None ): """simple docstring""" if "." in tensor_name: _snake_case : Any = tensor_name.split("." ) for split in splits[:-1]: _snake_case : int = getattr(a , a ) if new_module is None: raise ValueError(f'{module} has no attribute {split}.' ) _snake_case : List[Any] = new_module _snake_case : Dict = splits[-1] if tensor_name not in module._parameters and tensor_name not in module._buffers: raise ValueError(f'{module} does not have a parameter or a buffer named {tensor_name}.' ) _snake_case : str = tensor_name in module._buffers _snake_case : int = getattr(a , a ) if old_value.device == torch.device("meta" ) and device not in ["meta", torch.device("meta" )] and value is None: raise ValueError(f'{tensor_name} is on the meta device, we need a `value` to put in on {device}.' ) _snake_case : Union[str, Any] = False _snake_case : Any = False if is_buffer or not is_bitsandbytes_available(): _snake_case : Union[str, Any] = False _snake_case : Optional[int] = False else: _snake_case : List[Any] = hasattr(bnb.nn , "Params4bit" ) and isinstance(module._parameters[tensor_name] , bnb.nn.Paramsabit ) _snake_case : Tuple = isinstance(module._parameters[tensor_name] , bnb.nn.IntaParams ) if is_abit or is_abit: _snake_case : Optional[int] = module._parameters[tensor_name] if param.device.type != "cuda": if value is None: _snake_case : Any = old_value.to(a ) elif isinstance(a , torch.Tensor ): _snake_case : Optional[int] = value.to("cpu" ) if value.dtype == torch.inta: _snake_case : Tuple = version.parse(importlib.metadata.version("bitsandbytes" ) ) > version.parse( "0.37.2" ) if not is_abit_serializable: raise ValueError( "Detected int8 weights but the version of bitsandbytes is not compatible with int8 serialization. " "Make sure to download the latest `bitsandbytes` version. `pip install --upgrade bitsandbytes`." ) else: _snake_case : Optional[Any] = torch.tensor(a , device="cpu" ) # Support models using `Conv1D` in place of `nn.Linear` (e.g. gpt2) by transposing the weight matrix prior to quantization. # Since weights are saved in the correct "orientation", we skip transposing when loading. if issubclass(module.source_cls , a ) and fpaa_statistics is None: _snake_case : Tuple = new_value.T _snake_case : int = old_value.__dict__ if is_abit: _snake_case : Dict = bnb.nn.IntaParams(a , requires_grad=a , **a ).to(a ) elif is_abit: _snake_case : Optional[Any] = bnb.nn.Paramsabit(a , requires_grad=a , **a ).to(a ) _snake_case : Optional[Any] = new_value if fpaa_statistics is not None: setattr(module.weight , "SCB" , fpaa_statistics.to(a ) ) else: if value is None: _snake_case : Tuple = old_value.to(a ) elif isinstance(a , torch.Tensor ): _snake_case : List[str] = value.to(a ) else: _snake_case : List[Any] = torch.tensor(a , device=a ) if is_buffer: _snake_case : Tuple = new_value else: _snake_case : Optional[int] = nn.Parameter(a , requires_grad=old_value.requires_grad ) _snake_case : int = new_value def a__ ( a : List[str] , a : List[str]=None , a : Tuple=None , a : Optional[int]=None , a : List[Any]=False ): """simple docstring""" for name, module in model.named_children(): if current_key_name is None: _snake_case : Any = [] current_key_name.append(a ) if (isinstance(a , nn.Linear ) or isinstance(a , a )) and name not in modules_to_not_convert: # Check if the current key is not in the `modules_to_not_convert` if not any(key in ".".join(a ) for key in modules_to_not_convert ): with init_empty_weights(): if isinstance(a , a ): _snake_case : List[str] = module.weight.shape else: _snake_case : List[str] = module.in_features _snake_case : Tuple = module.out_features if quantization_config.quantization_method() == "llm_int8": _snake_case : Optional[int] = bnb.nn.LinearabitLt( a , a , module.bias is not None , has_fpaa_weights=quantization_config.llm_inta_has_fpaa_weight , threshold=quantization_config.llm_inta_threshold , ) _snake_case : List[str] = True else: if ( quantization_config.llm_inta_skip_modules is not None and name in quantization_config.llm_inta_skip_modules ): pass else: _snake_case : str = bnb.nn.Linearabit( a , a , module.bias is not None , quantization_config.bnb_abit_compute_dtype , compress_statistics=quantization_config.bnb_abit_use_double_quant , quant_type=quantization_config.bnb_abit_quant_type , ) _snake_case : Dict = True # Store the module class in case we need to transpose the weight later _snake_case : Optional[int] = type(a ) # Force requires grad to False to avoid unexpected errors model._modules[name].requires_grad_(a ) if len(list(module.children() ) ) > 0: _snake_case : Dict = _replace_with_bnb_linear( a , a , a , a , has_been_replaced=a , ) # Remove the last key for recursion current_key_name.pop(-1 ) return model, has_been_replaced def a__ ( a : Dict , a : str=None , a : str=None , a : int=None ): """simple docstring""" _snake_case : Any = ["lm_head"] if modules_to_not_convert is None else modules_to_not_convert _snake_case : List[str] = _replace_with_bnb_linear( a , a , a , a ) if not has_been_replaced: logger.warning( "You are loading your model in 8bit or 4bit but no linear modules were found in your model." " Please double check your model architecture, or submit an issue on github if you think this is" " a bug." ) return model def a__ ( *a : Dict , **a : Dict ): """simple docstring""" warnings.warn( "`replace_8bit_linear` will be deprecated in a future version, please use `replace_with_bnb_linear` instead" , a , ) return replace_with_bnb_linear(*a , **a ) def a__ ( *a : Optional[Any] , **a : Optional[Any] ): """simple docstring""" warnings.warn( "`set_module_8bit_tensor_to_device` will be deprecated in a future version, please use `set_module_quantized_tensor_to_device` instead" , a , ) return set_module_quantized_tensor_to_device(*a , **a ) def a__ ( a : Union[str, Any] ): """simple docstring""" _snake_case : Optional[int] = deepcopy(a ) # this has 0 cost since it is done inside `init_empty_weights` context manager` tied_model.tie_weights() _snake_case : Optional[int] = find_tied_parameters(a ) # For compatibility with Accelerate < 0.18 if isinstance(a , a ): _snake_case : Dict = sum(list(tied_params.values() ) , [] ) + list(tied_params.keys() ) else: _snake_case : Dict = sum(a , [] ) _snake_case : Optional[Any] = len(a ) > 0 # Check if it is a base model _snake_case : List[Any] = not hasattr(a , model.base_model_prefix ) # Ignore this for base models (BertModel, GPT2Model, etc.) if (not has_tied_params) and is_base_model: return [] # otherwise they have an attached head _snake_case : List[str] = list(model.named_children() ) _snake_case : Union[str, Any] = [list_modules[-1][0]] # add last module together with tied weights _snake_case : Optional[int] = set(a ) - set(a ) _snake_case : Tuple = list(set(a ) ) + list(a ) # remove ".weight" from the keys _snake_case : Optional[int] = [".weight", ".bias"] _snake_case : List[str] = [] for name in list_untouched: for name_to_remove in names_to_remove: if name_to_remove in name: _snake_case : Tuple = name.replace(a , "" ) filtered_module_names.append(a ) return filtered_module_names
707
"""simple docstring""" from __future__ import annotations import requests _a : List[str] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def a__ ( a : str , a : int = 1 , a : str = "new" , a : list | None = None ): """simple docstring""" _snake_case : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(a ) - valid_terms ) ): _snake_case : Optional[int] = f'Invalid search term: {invalid_search_terms}' raise ValueError(a ) _snake_case : int = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"User-agent": "A random string"} , ) if response.status_code == 429: raise requests.HTTPError _snake_case : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(a )} _snake_case : Tuple = {} for id_ in range(a ): _snake_case : List[str] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
87
0
"""simple docstring""" import argparse import os import gluonnlp as nlp import mxnet as mx import numpy as np import torch from gluonnlp.base import get_home_dir from gluonnlp.model.bert import BERTEncoder from gluonnlp.model.utils import _load_vocab from gluonnlp.vocab import Vocab from packaging import version from torch import nn from transformers import BertConfig, BertForMaskedLM, BertModel, RobertaTokenizer from transformers.models.bert.modeling_bert import ( BertIntermediate, BertLayer, BertOutput, BertSelfAttention, BertSelfOutput, ) from transformers.utils import logging if version.parse(nlp.__version__) != version.parse("""0.8.3"""): raise Exception("""requires gluonnlp == 0.8.3""") if version.parse(mx.__version__) != version.parse("""1.5.0"""): raise Exception("""requires mxnet == 1.5.0""") logging.set_verbosity_info() _a : Optional[int] = logging.get_logger(__name__) _a : int = """The Nymphenburg Palace is a beautiful palace in Munich!""" def a__ ( a : str , a : str ): """simple docstring""" _snake_case : List[str] = { "attention_cell": "multi_head", "num_layers": 4, "units": 1_024, "hidden_size": 768, "max_length": 512, "num_heads": 8, "scaled": True, "dropout": 0.1, "use_residual": True, "embed_size": 1_024, "embed_dropout": 0.1, "word_embed": None, "layer_norm_eps": 1e-5, "token_type_vocab_size": 2, } _snake_case : Optional[int] = bort_4_8_768_1024_hparams # Let's construct the original Bort model here # Taken from official BERT implementation, see: # https://github.com/alexa/bort/blob/master/bort/bort.py _snake_case : int = BERTEncoder( attention_cell=predefined_args["attention_cell"] , num_layers=predefined_args["num_layers"] , units=predefined_args["units"] , hidden_size=predefined_args["hidden_size"] , max_length=predefined_args["max_length"] , num_heads=predefined_args["num_heads"] , scaled=predefined_args["scaled"] , dropout=predefined_args["dropout"] , output_attention=a , output_all_encodings=a , use_residual=predefined_args["use_residual"] , activation=predefined_args.get("activation" , "gelu" ) , layer_norm_eps=predefined_args.get("layer_norm_eps" , a ) , ) # Vocab information needs to be fetched first # It's the same as RoBERTa, so RobertaTokenizer can be used later _snake_case : str = "openwebtext_ccnews_stories_books_cased" # Specify download folder to Gluonnlp's vocab _snake_case : Tuple = os.path.join(get_home_dir() , "models" ) _snake_case : str = _load_vocab(a , a , a , cls=a ) _snake_case : Tuple = nlp.model.BERTModel( a , len(a ) , units=predefined_args["units"] , embed_size=predefined_args["embed_size"] , embed_dropout=predefined_args["embed_dropout"] , word_embed=predefined_args["word_embed"] , use_pooler=a , use_token_type_embed=a , token_type_vocab_size=predefined_args["token_type_vocab_size"] , use_classifier=a , use_decoder=a , ) original_bort.load_parameters(a , cast_dtype=a , ignore_extra=a ) _snake_case : Optional[Any] = original_bort._collect_params_with_prefix() # Build our config 🤗 _snake_case : str = { "architectures": ["BertForMaskedLM"], "attention_probs_dropout_prob": predefined_args["dropout"], "hidden_act": "gelu", "hidden_dropout_prob": predefined_args["dropout"], "hidden_size": predefined_args["embed_size"], "initializer_range": 0.02, "intermediate_size": predefined_args["hidden_size"], "layer_norm_eps": predefined_args["layer_norm_eps"], "max_position_embeddings": predefined_args["max_length"], "model_type": "bort", "num_attention_heads": predefined_args["num_heads"], "num_hidden_layers": predefined_args["num_layers"], "pad_token_id": 1, # 2 = BERT, 1 = RoBERTa "type_vocab_size": 1, # 2 = BERT, 1 = RoBERTa "vocab_size": len(a ), } _snake_case : Any = BertConfig.from_dict(a ) _snake_case : Dict = BertForMaskedLM(a ) hf_bort_model.eval() # Parameter mapping table (Gluonnlp to Transformers) # * denotes layer index # # | Gluon Parameter | Transformers Parameter # | -------------------------------------------------------------- | ---------------------- # | `encoder.layer_norm.beta` | `bert.embeddings.LayerNorm.bias` # | `encoder.layer_norm.gamma` | `bert.embeddings.LayerNorm.weight` # | `encoder.position_weight` | `bert.embeddings.position_embeddings.weight` # | `word_embed.0.weight` | `bert.embeddings.word_embeddings.weight` # | `encoder.transformer_cells.*.attention_cell.proj_key.bias` | `bert.encoder.layer.*.attention.self.key.bias` # | `encoder.transformer_cells.*.attention_cell.proj_key.weight` | `bert.encoder.layer.*.attention.self.key.weight` # | `encoder.transformer_cells.*.attention_cell.proj_query.bias` | `bert.encoder.layer.*.attention.self.query.bias` # | `encoder.transformer_cells.*.attention_cell.proj_query.weight` | `bert.encoder.layer.*.attention.self.query.weight` # | `encoder.transformer_cells.*.attention_cell.proj_value.bias` | `bert.encoder.layer.*.attention.self.value.bias` # | `encoder.transformer_cells.*.attention_cell.proj_value.weight` | `bert.encoder.layer.*.attention.self.value.weight` # | `encoder.transformer_cells.*.ffn.ffn_2.bias` | `bert.encoder.layer.*.attention.output.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_2.weight` | `bert.encoder.layer.*.attention.output.dense.weight` # | `encoder.transformer_cells.*.layer_norm.beta` | `bert.encoder.layer.*.attention.output.LayerNorm.bias` # | `encoder.transformer_cells.*.layer_norm.gamma` | `bert.encoder.layer.*.attention.output.LayerNorm.weight` # | `encoder.transformer_cells.*.ffn.ffn_1.bias` | `bert.encoder.layer.*.intermediate.dense.bias` # | `encoder.transformer_cells.*.ffn.ffn_1.weight` | `bert.encoder.layer.*.intermediate.dense.weight` # | `encoder.transformer_cells.*.ffn.layer_norm.beta` | `bert.encoder.layer.*.output.LayerNorm.bias` # | `encoder.transformer_cells.*.ffn.layer_norm.gamma` | `bert.encoder.layer.*.output.LayerNorm.weight` # | `encoder.transformer_cells.*.proj.bias` | `bert.encoder.layer.*.output.dense.bias` # | `encoder.transformer_cells.*.proj.weight` | `bert.encoder.layer.*.output.dense.weight` # Helper function to convert MXNET Arrays to PyTorch def to_torch(a : Any ) -> nn.Parameter: return nn.Parameter(torch.FloatTensor(mx_array.data().asnumpy() ) ) # Check param shapes and map new HF param back def check_and_map_params(a : List[Any] , a : List[str] ): _snake_case : Optional[Any] = hf_param.shape _snake_case : str = to_torch(params[gluon_param] ) _snake_case : Union[str, Any] = gluon_param.shape assert ( shape_hf == shape_gluon ), f'The gluon parameter {gluon_param} has shape {shape_gluon}, but expects shape {shape_hf} for Transformers' return gluon_param _snake_case : Optional[Any] = check_and_map_params( hf_bort_model.bert.embeddings.word_embeddings.weight , "word_embed.0.weight" ) _snake_case : Optional[Any] = check_and_map_params( hf_bort_model.bert.embeddings.position_embeddings.weight , "encoder.position_weight" ) _snake_case : List[str] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.bias , "encoder.layer_norm.beta" ) _snake_case : Union[str, Any] = check_and_map_params( hf_bort_model.bert.embeddings.LayerNorm.weight , "encoder.layer_norm.gamma" ) # Inspired by RoBERTa conversion script, we just zero them out (Bort does not use them) _snake_case : Dict = torch.zeros_like( hf_bort_model.bert.embeddings.token_type_embeddings.weight.data ) for i in range(hf_bort_config.num_hidden_layers ): _snake_case : BertLayer = hf_bort_model.bert.encoder.layer[i] # self attention _snake_case : BertSelfAttention = layer.attention.self _snake_case : int = check_and_map_params( self_attn.key.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_key.bias' ) _snake_case : Dict = check_and_map_params( self_attn.key.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_key.weight' ) _snake_case : Dict = check_and_map_params( self_attn.query.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_query.bias' ) _snake_case : str = check_and_map_params( self_attn.query.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_query.weight' ) _snake_case : Optional[int] = check_and_map_params( self_attn.value.bias.data , f'encoder.transformer_cells.{i}.attention_cell.proj_value.bias' ) _snake_case : List[str] = check_and_map_params( self_attn.value.weight.data , f'encoder.transformer_cells.{i}.attention_cell.proj_value.weight' ) # self attention output _snake_case : BertSelfOutput = layer.attention.output _snake_case : Any = check_and_map_params( self_output.dense.bias , f'encoder.transformer_cells.{i}.proj.bias' ) _snake_case : Dict = check_and_map_params( self_output.dense.weight , f'encoder.transformer_cells.{i}.proj.weight' ) _snake_case : int = check_and_map_params( self_output.LayerNorm.bias , f'encoder.transformer_cells.{i}.layer_norm.beta' ) _snake_case : Any = check_and_map_params( self_output.LayerNorm.weight , f'encoder.transformer_cells.{i}.layer_norm.gamma' ) # intermediate _snake_case : BertIntermediate = layer.intermediate _snake_case : int = check_and_map_params( intermediate.dense.bias , f'encoder.transformer_cells.{i}.ffn.ffn_1.bias' ) _snake_case : Union[str, Any] = check_and_map_params( intermediate.dense.weight , f'encoder.transformer_cells.{i}.ffn.ffn_1.weight' ) # output _snake_case : BertOutput = layer.output _snake_case : int = check_and_map_params( bert_output.dense.bias , f'encoder.transformer_cells.{i}.ffn.ffn_2.bias' ) _snake_case : Dict = check_and_map_params( bert_output.dense.weight , f'encoder.transformer_cells.{i}.ffn.ffn_2.weight' ) _snake_case : Any = check_and_map_params( bert_output.LayerNorm.bias , f'encoder.transformer_cells.{i}.ffn.layer_norm.beta' ) _snake_case : Union[str, Any] = check_and_map_params( bert_output.LayerNorm.weight , f'encoder.transformer_cells.{i}.ffn.layer_norm.gamma' ) # Save space and energy 🎄 hf_bort_model.half() # Compare output of both models _snake_case : Optional[Any] = RobertaTokenizer.from_pretrained("roberta-base" ) _snake_case : Any = tokenizer.encode_plus(a )["input_ids"] # Get gluon output _snake_case : Union[str, Any] = mx.nd.array([input_ids] ) _snake_case : Optional[int] = original_bort(inputs=a , token_types=[] ) # Get Transformer output (save and reload model again) hf_bort_model.save_pretrained(a ) _snake_case : str = BertModel.from_pretrained(a ) hf_bort_model.eval() _snake_case : str = tokenizer.encode_plus(a , return_tensors="pt" ) _snake_case : Tuple = hf_bort_model(**a )[0] _snake_case : Any = output_gluon[0].asnumpy() _snake_case : Tuple = output_hf[0].detach().numpy() _snake_case : Optional[int] = np.max(np.abs(hf_layer - gluon_layer ) ).item() _snake_case : List[Any] = np.allclose(a , a , atol=1e-3 ) if success: print("✔️ Both model do output the same tensors" ) else: print("❌ Both model do **NOT** output the same tensors" ) print("Absolute difference is:" , a ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() # Required parameters parser.add_argument( """--bort_checkpoint_path""", default=None, type=str, required=True, help="""Path the official Bort params file.""" ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) _a : Tuple = parser.parse_args() convert_bort_checkpoint_to_pytorch(args.bort_checkpoint_path, args.pytorch_dump_folder_path)
708
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def a__ ( a : float , a : float , a : bool = False ): """simple docstring""" if radian_mode: return [magnitude * cos(a ), magnitude * sin(a )] return [magnitude * cos(radians(a ) ), magnitude * sin(radians(a ) )] def a__ ( a : NDArray[floataa] , a : NDArray[floataa] , a : float = 10**-1 ): """simple docstring""" _snake_case : NDArray[floataa] = cross(a , a ) _snake_case : float = sum(a ) return abs(a ) < eps if __name__ == "__main__": # Test to check if it works _a : Tuple = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) _a : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg _a : List[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) _a : List[Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg _a : List[str] = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) _a : Optional[int] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
87
0
"""simple docstring""" import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) _a : Optional[int] = """https://openaipublic.azureedge.net/jukebox/models/""" _a : int = { """jukebox-1b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """1b_lyrics/prior_level_2.pth.tar""", ], """jukebox-5b-lyrics""": [ """5b/vqvae.pth.tar""", """5b/prior_level_0.pth.tar""", """5b/prior_level_1.pth.tar""", """5b_lyrics/prior_level_2.pth.tar""", ], } def a__ ( a : Tuple ): """simple docstring""" if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 10: _snake_case : List[Any] = key.replace(".model.1.bias" , ".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 10: _snake_case : Optional[Any] = key.replace(".model.1.weight" , ".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 10: _snake_case : Union[str, Any] = key.replace(".model.3.bias" , ".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 10: _snake_case : Any = key.replace(".model.3.weight" , ".conv1d_2.weight" ) if "conditioner_blocks.0." in key: _snake_case : Optional[int] = key.replace("conditioner_blocks.0" , "conditioner_blocks" ) if "prime_prior" in key: _snake_case : Optional[int] = key.replace("prime_prior" , "encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: _snake_case : List[str] = key.replace(".emb." , "." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k" , ".codebook" ) if "y_emb." in key: return key.replace("y_emb." , "metadata_embedding." ) if "x_emb.emb." in key: _snake_case : Optional[Any] = key.replace("0.x_emb.emb" , "embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln" , "encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln" , ".layer_norm" ) if "_ln" in key: return key.replace("_ln" , "_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj" , "encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out" , "encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out" , "fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb" , "embed_tokens" ) return key def a__ ( a : int , a : int , a : List[str] , a : List[str] ): """simple docstring""" _snake_case : int = {} import re _snake_case : Union[str, Any] = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) _snake_case : Tuple = re.compile( R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) _snake_case : str = re.compile(R"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) _snake_case : Optional[Any] = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) _snake_case : Any = re.compile( R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) _snake_case : Any = re.compile(R"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) _snake_case : List[Any] = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) _snake_case : Any = re.compile( R"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) _snake_case : Union[str, Any] = re.compile(R"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(a ): _snake_case : List[Any] = re_encoder_block_conv_in.match(a ) _snake_case : Optional[int] = regex_match.groups() _snake_case : Any = int(groups[2] ) * 2 + int(groups[3] ) _snake_case : str = f'encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}' _snake_case : Tuple = re_encoder_block_conv_in.sub(a , a ) elif re_encoder_block_resnet.fullmatch(a ): _snake_case : Any = re_encoder_block_resnet.match(a ) _snake_case : Optional[Any] = regex_match.groups() _snake_case : Optional[int] = int(groups[2] ) * 2 + int(groups[3] ) _snake_case : Optional[int] = {"1": 1, "3": 2}[groups[-2]] _snake_case : List[str] = f'encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.' _snake_case : Optional[Any] = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' _snake_case : Tuple = prefix + resnet_block _snake_case : Dict = re_encoder_block_resnet.sub(a , a ) elif re_encoder_block_proj_out.fullmatch(a ): _snake_case : List[str] = re_encoder_block_proj_out.match(a ) _snake_case : int = regex_match.groups() _snake_case : Tuple = f'encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}' _snake_case : str = re_encoder_block_proj_out.sub(a , a ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(a ): _snake_case : Union[str, Any] = re_decoder_block_conv_out.match(a ) _snake_case : Union[str, Any] = regex_match.groups() _snake_case : Optional[Any] = int(groups[2] ) * 2 + int(groups[3] ) - 2 _snake_case : Tuple = f'decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}' _snake_case : Tuple = re_decoder_block_conv_out.sub(a , a ) elif re_decoder_block_resnet.fullmatch(a ): _snake_case : Dict = re_decoder_block_resnet.match(a ) _snake_case : Optional[Any] = regex_match.groups() _snake_case : Dict = int(groups[2] ) * 2 + int(groups[3] ) - 2 _snake_case : Tuple = {"1": 1, "3": 2}[groups[-2]] _snake_case : Optional[int] = f'decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.' _snake_case : int = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' _snake_case : str = prefix + resnet_block _snake_case : List[Any] = re_decoder_block_resnet.sub(a , a ) elif re_decoder_block_proj_in.fullmatch(a ): _snake_case : int = re_decoder_block_proj_in.match(a ) _snake_case : Dict = regex_match.groups() _snake_case : str = f'decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}' _snake_case : Dict = re_decoder_block_proj_in.sub(a , a ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(a ): _snake_case : Optional[Any] = re_prior_cond_conv_out.match(a ) _snake_case : Any = regex_match.groups() _snake_case : List[str] = int(groups[1] ) * 2 + int(groups[2] ) - 2 _snake_case : Any = f'conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}' _snake_case : Union[str, Any] = re_prior_cond_conv_out.sub(a , a ) elif re_prior_cond_resnet.fullmatch(a ): _snake_case : List[Any] = re_prior_cond_resnet.match(a ) _snake_case : Tuple = regex_match.groups() _snake_case : str = int(groups[1] ) * 2 + int(groups[2] ) - 2 _snake_case : Optional[int] = {"1": 1, "3": 2}[groups[-2]] _snake_case : List[Any] = f'conditioner_blocks.upsampler.upsample_block.{block_index}.' _snake_case : str = f'resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}' _snake_case : List[Any] = prefix + resnet_block _snake_case : Tuple = re_prior_cond_resnet.sub(a , a ) elif re_prior_cond_proj_in.fullmatch(a ): _snake_case : Tuple = re_prior_cond_proj_in.match(a ) _snake_case : List[Any] = regex_match.groups() _snake_case : Any = f'conditioner_blocks.upsampler.proj_in.{groups[-1]}' _snake_case : Tuple = re_prior_cond_proj_in.sub(a , a ) # keep original key else: _snake_case : Optional[int] = original_key _snake_case : Union[str, Any] = replace_key(a ) if f'{key_prefix}.{key}' not in model_state_dict or key is None: print(f'failed converting {original_key} to {key}, does not match' ) # handle missmatched shape elif value.shape != model_state_dict[f'{key_prefix}.{key}'].shape: _snake_case : int = model_state_dict[f'{key_prefix}.{key}'] print(f'{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match' ) _snake_case : int = original_key _snake_case : Union[str, Any] = original_key _snake_case : Union[str, Any] = value return new_dict @torch.no_grad() def a__ ( a : int=None , a : Tuple=None ): """simple docstring""" for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f'{pytorch_dump_folder_path}/{file.split("/" )[-1]}' ): _snake_case : Tuple = requests.get(f'{PREFIX}{file}' , allow_redirects=a ) os.makedirs(f'{pytorch_dump_folder_path}/' , exist_ok=a ) open(f'{pytorch_dump_folder_path}/{file.split("/" )[-1]}' , "wb" ).write(r.content ) _snake_case : Any = MODEL_MAPPING[model_name.split("/" )[-1]] _snake_case : str = JukeboxConfig.from_pretrained(a ) _snake_case : Any = JukeboxModel(a ) _snake_case : List[Any] = [] _snake_case : int = {} for i, dict_name in enumerate(a ): _snake_case : Optional[Any] = torch.load(f'{pytorch_dump_folder_path}/{dict_name.split("/" )[-1]}' )["model"] _snake_case : Dict = {} for k in old_dic.keys(): if k.endswith(".b" ): _snake_case : int = old_dic[k] elif k.endswith(".w" ): _snake_case : List[str] = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: _snake_case : str = old_dic[k] else: _snake_case : Any = old_dic[k] _snake_case : int = "vqvae" if i == 0 else f'priors.{3 - i}' _snake_case : int = fix_jukebox_keys(a , model.state_dict() , a , a ) weight_dict.append(a ) _snake_case : List[Any] = weight_dict.pop(0 ) model.vqvae.load_state_dict(a ) for i in range(len(a ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(a ).mkdir(exist_ok=a ) with open(f'{pytorch_dump_folder_path}/mapping.json' , "w" ) as txtfile: json.dump(a , a ) print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a ) return weight_dict if __name__ == "__main__": _a : Union[str, Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""jukebox-5b-lyrics""", type=str, help="""Name of the model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default="""jukebox-5b-lyrics-converted""", type=str, help="""Path to the output PyTorch model directory.""", ) _a : Optional[Any] = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
709
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[int] = logging.get_logger(__name__) _a : str = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class _UpperCAmelCase ( _snake_case): __lowercase : Optional[Any] = """openai-gpt""" __lowercase : Dict = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , snake_case_=4_04_78 , snake_case_=5_12 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_="cls_index" , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=0.1 , **snake_case_ , ): _snake_case : Tuple = vocab_size _snake_case : Dict = n_positions _snake_case : Any = n_embd _snake_case : Any = n_layer _snake_case : Optional[int] = n_head _snake_case : Union[str, Any] = afn _snake_case : Dict = resid_pdrop _snake_case : str = embd_pdrop _snake_case : Union[str, Any] = attn_pdrop _snake_case : str = layer_norm_epsilon _snake_case : Union[str, Any] = initializer_range _snake_case : Any = summary_type _snake_case : List[str] = summary_use_proj _snake_case : Optional[int] = summary_activation _snake_case : Union[str, Any] = summary_first_dropout _snake_case : Optional[int] = summary_proj_to_labels super().__init__(**snake_case_ )
87
0
"""simple docstring""" import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import is_accelerate_available, is_torch_available, is_transformers_available, is_xformers_available from . import BaseDiffusersCLICommand def a__ ( a : int ): """simple docstring""" return EnvironmentCommand() class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : List[str] = parser.add_parser("env" ) download_parser.set_defaults(func=snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = huggingface_hub.__version__ _snake_case : int = "not installed" _snake_case : Tuple = "NA" if is_torch_available(): import torch _snake_case : int = torch.__version__ _snake_case : int = torch.cuda.is_available() _snake_case : List[str] = "not installed" if is_transformers_available(): import transformers _snake_case : Optional[Any] = transformers.__version__ _snake_case : Tuple = "not installed" if is_accelerate_available(): import accelerate _snake_case : Optional[int] = accelerate.__version__ _snake_case : List[str] = "not installed" if is_xformers_available(): import xformers _snake_case : Any = xformers.__version__ _snake_case : int = { "`diffusers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "PyTorch version (GPU?)": F'{pt_version} ({pt_cuda_available})', "Huggingface_hub version": hub_version, "Transformers version": transformers_version, "Accelerate version": accelerate_version, "xFormers version": xformers_version, "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n" ) print(self.format_dict(snake_case_ ) ) return info @staticmethod def lowerCamelCase__ ( snake_case_ ): return "\n".join([F'- {prop}: {val}' for prop, val in d.items()] ) + "\n"
710
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _a : Tuple = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _a : int = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def a__ ( a : List[str] , a : int , a : int ): """simple docstring""" _snake_case : Union[str, Any] = state_dict.pop(a ) _snake_case : Union[str, Any] = val def a__ ( a : Tuple ): """simple docstring""" _snake_case : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case : Dict = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) _snake_case : Tuple = value else: _snake_case : Dict = value return new_state_dict def a__ ( a : int ): """simple docstring""" _snake_case : Any = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[:256, :] _snake_case : List[str] = in_proj_bias[:256] _snake_case : Optional[Any] = in_proj_weight[256:512, :] _snake_case : List[str] = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : Dict = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case : List[Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : Union[str, Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : Union[str, Any] = in_proj_weight[:256, :] _snake_case : Tuple = in_proj_bias[:256] _snake_case : int = in_proj_weight[256:512, :] _snake_case : int = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : str = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _snake_case : Dict = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _snake_case : Optional[int] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case : Dict = in_proj_weight_cross_attn[:256, :] _snake_case : Any = in_proj_bias_cross_attn[:256] _snake_case : Union[str, Any] = in_proj_weight_cross_attn[256:512, :] _snake_case : Optional[int] = in_proj_bias_cross_attn[256:512] _snake_case : Any = in_proj_weight_cross_attn[-256:, :] _snake_case : str = in_proj_bias_cross_attn[-256:] def a__ ( a : str , a : int ): """simple docstring""" _snake_case , _snake_case : List[str] = image.size _snake_case : Dict = max(a , a ) _snake_case : Union[str, Any] = 800 if "detection" in checkpoint_url else 1_000 _snake_case : Any = target_max_size / current_max_size _snake_case : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def a__ ( a : str ): """simple docstring""" _snake_case : str = F.to_tensor(a ) _snake_case : Union[str, Any] = F.normalize(a , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def a__ ( a : Optional[Any] , a : Any , a : Union[str, Any] ): """simple docstring""" logger.info("Converting model..." ) # load original state dict _snake_case : Tuple = torch.hub.load_state_dict_from_url(a , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(a , a , a ) _snake_case : Union[str, Any] = rename_backbone_keys(a ) # query, key and value matrices need special treatment read_in_q_k_v(a ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case : int = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): _snake_case : Optional[int] = state_dict.pop(a ) _snake_case : Any = val # create HuggingFace model and load state dict _snake_case : Tuple = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case : Any = 15 _snake_case : int = 2 _snake_case : Optional[Any] = {0: "table", 1: "table rotated"} _snake_case : Union[str, Any] = idalabel _snake_case : Union[str, Any] = {v: k for k, v in idalabel.items()} else: _snake_case : Any = 125 _snake_case : Union[str, Any] = 6 _snake_case : List[str] = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } _snake_case : Any = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} _snake_case : Union[str, Any] = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1_000 ) _snake_case : str = TableTransformerForObjectDetection(a ) model.load_state_dict(a ) model.eval() # verify our conversion _snake_case : Optional[int] = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" _snake_case : Optional[Any] = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=a ) _snake_case : Dict = Image.open(a ).convert("RGB" ) _snake_case : Union[str, Any] = normalize(resize(a , a ) ).unsqueeze(0 ) _snake_case : str = model(a ) if "detection" in checkpoint_url: _snake_case : int = (1, 15, 3) _snake_case : List[str] = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) _snake_case : List[str] = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: _snake_case : Union[str, Any] = (1, 125, 7) _snake_case : str = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) _snake_case : Optional[Any] = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , a , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , a , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) _snake_case : int = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(a ) image_processor.push_to_hub(a ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _a : Any = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : List[Any] = {"""configuration_xlnet""": ["""XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP""", """XLNetConfig"""]} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = ["""XLNetTokenizer"""] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""XLNetTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ """XLNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """XLNetForMultipleChoice""", """XLNetForQuestionAnswering""", """XLNetForQuestionAnsweringSimple""", """XLNetForSequenceClassification""", """XLNetForTokenClassification""", """XLNetLMHeadModel""", """XLNetModel""", """XLNetPreTrainedModel""", """load_tf_weights_in_xlnet""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFXLNetForMultipleChoice""", """TFXLNetForQuestionAnsweringSimple""", """TFXLNetForSequenceClassification""", """TFXLNetForTokenClassification""", """TFXLNetLMHeadModel""", """TFXLNetMainLayer""", """TFXLNetModel""", """TFXLNetPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_xlnet import XLNET_PRETRAINED_CONFIG_ARCHIVE_MAP, XLNetConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet import XLNetTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_xlnet_fast import XLNetTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_xlnet import ( XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, XLNetForMultipleChoice, XLNetForQuestionAnswering, XLNetForQuestionAnsweringSimple, XLNetForSequenceClassification, XLNetForTokenClassification, XLNetLMHeadModel, XLNetModel, XLNetPreTrainedModel, load_tf_weights_in_xlnet, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_xlnet import ( TF_XLNET_PRETRAINED_MODEL_ARCHIVE_LIST, TFXLNetForMultipleChoice, TFXLNetForQuestionAnsweringSimple, TFXLNetForSequenceClassification, TFXLNetForTokenClassification, TFXLNetLMHeadModel, TFXLNetMainLayer, TFXLNetModel, TFXLNetPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
711
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
87
0
import colorsys from PIL import Image # type: ignore def a__ ( a : float , a : float , a : int ): """simple docstring""" _snake_case : Any = x _snake_case : int = y for step in range(a ): # noqa: B007 _snake_case : str = a * a - b * b + x _snake_case : Tuple = 2 * a * b + y _snake_case : List[Any] = a_new # divergence happens for all complex number with an absolute value # greater than 4 if a * a + b * b > 4: break return step / (max_step - 1) def a__ ( a : float ): """simple docstring""" if distance == 1: return (0, 0, 0) else: return (255, 255, 255) def a__ ( a : float ): """simple docstring""" if distance == 1: return (0, 0, 0) else: return tuple(round(i * 255 ) for i in colorsys.hsv_to_rgb(a , 1 , 1 ) ) def a__ ( a : int = 800 , a : int = 600 , a : float = -0.6 , a : float = 0 , a : float = 3.2 , a : int = 50 , a : bool = True , ): """simple docstring""" _snake_case : Any = Image.new("RGB" , (image_width, image_height) ) _snake_case : str = img.load() # loop through the image-coordinates for image_x in range(a ): for image_y in range(a ): # determine the figure-coordinates based on the image-coordinates _snake_case : str = figure_width / image_width * image_height _snake_case : Dict = figure_center_x + (image_x / image_width - 0.5) * figure_width _snake_case : Union[str, Any] = figure_center_y + (image_y / image_height - 0.5) * figure_height _snake_case : List[str] = get_distance(a , a , a ) # color the corresponding pixel based on the selected coloring-function if use_distance_color_coding: _snake_case : Tuple = get_color_coded_rgb(a ) else: _snake_case : List[str] = get_black_and_white_rgb(a ) return img if __name__ == "__main__": import doctest doctest.testmod() # colored version, full figure _a : Optional[Any] = get_image() # uncomment for colored version, different section, zoomed in # img = get_image(figure_center_x = -0.6, figure_center_y = -0.4, # figure_width = 0.8) # uncomment for black and white version, full figure # img = get_image(use_distance_color_coding = False) # uncomment to save the image # img.save("mandelbrot.png") img.show()
712
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _UpperCAmelCase ( _snake_case , _snake_case): __lowercase : List[Any] = """convnextv2""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=4 , snake_case_=None , snake_case_=None , snake_case_="gelu" , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=0.0 , snake_case_=2_24 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Tuple = num_channels _snake_case : Optional[int] = patch_size _snake_case : Tuple = num_stages _snake_case : int = [96, 1_92, 3_84, 7_68] if hidden_sizes is None else hidden_sizes _snake_case : str = [3, 3, 9, 3] if depths is None else depths _snake_case : int = hidden_act _snake_case : Tuple = initializer_range _snake_case : Union[str, Any] = layer_norm_eps _snake_case : Optional[int] = drop_path_rate _snake_case : Union[str, Any] = image_size _snake_case : List[Any] = ["stem"] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] _snake_case , _snake_case : Dict = get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names )
87
0
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
713
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def a__ ( a : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) _a : int = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : Dict = parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=snake_case_ , required=snake_case_ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=snake_case_ , required=snake_case_ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=snake_case_ , required=snake_case_ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=snake_case_ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=snake_case_ , default=snake_case_ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : str = logging.get_logger("transformers-cli/converting" ) self._logger.info(F'Loading model {model_type}' ) _snake_case : Optional[int] = model_type _snake_case : Any = tf_checkpoint _snake_case : Optional[int] = pytorch_dump_output _snake_case : Tuple = config _snake_case : Tuple = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) if "ckpt" in self._tf_checkpoint.lower(): _snake_case : int = self._tf_checkpoint _snake_case : Optional[Any] = "" else: _snake_case : Optional[int] = self._tf_checkpoint _snake_case : List[str] = "" convert_transfo_xl_checkpoint_to_pytorch( snake_case_ , self._config , self._pytorch_dump_output , snake_case_ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
87
0
import gc import unittest import numpy as np import torch from diffusers import AutoencoderKL, DDIMScheduler, DiTPipeline, DPMSolverMultistepScheduler, TransformeraDModel from diffusers.utils import is_xformers_available, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu from ..pipeline_params import ( CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS, CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS, ) from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : str = DiTPipeline __lowercase : Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_PARAMS __lowercase : Any = PipelineTesterMixin.required_optional_params - { """latents""", """num_images_per_prompt""", """callback""", """callback_steps""", } __lowercase : Optional[int] = CLASS_CONDITIONED_IMAGE_GENERATION_BATCH_PARAMS __lowercase : Optional[int] = False def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : Union[str, Any] = TransformeraDModel( sample_size=16 , num_layers=2 , patch_size=4 , attention_head_dim=8 , num_attention_heads=2 , in_channels=4 , out_channels=8 , attention_bias=snake_case_ , activation_fn="gelu-approximate" , num_embeds_ada_norm=10_00 , norm_type="ada_norm_zero" , norm_elementwise_affine=snake_case_ , ) _snake_case : int = AutoencoderKL() _snake_case : List[str] = DDIMScheduler() _snake_case : Dict = {"transformer": transformer.eval(), "vae": vae.eval(), "scheduler": scheduler} return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : Dict = torch.manual_seed(snake_case_ ) else: _snake_case : List[Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : List[Any] = { "class_labels": [1], "generator": generator, "num_inference_steps": 2, "output_type": "numpy", } return inputs def lowerCamelCase__ ( self ): _snake_case : Any = "cpu" _snake_case : str = self.get_dummy_components() _snake_case : Union[str, Any] = self.pipeline_class(**snake_case_ ) pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Tuple = self.get_dummy_inputs(snake_case_ ) _snake_case : List[str] = pipe(**snake_case_ ).images _snake_case : Optional[int] = image[0, -3:, -3:, -1] self.assertEqual(image.shape , (1, 16, 16, 3) ) _snake_case : List[str] = np.array([0.2946, 0.6601, 0.4329, 0.3296, 0.4144, 0.5319, 0.7273, 0.5013, 0.4457] ) _snake_case : Optional[Any] = np.abs(image_slice.flatten() - expected_slice ).max() self.assertLessEqual(snake_case_ , 1E-3 ) def lowerCamelCase__ ( self ): self._test_inference_batch_single_identical(relax_max_difference=snake_case_ , expected_max_diff=1E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(expected_max_diff=1E-3 ) @require_torch_gpu @slow class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self ): _snake_case : List[str] = torch.manual_seed(0 ) _snake_case : Dict = DiTPipeline.from_pretrained("facebook/DiT-XL-2-256" ) pipe.to("cuda" ) _snake_case : Union[str, Any] = ["vase", "umbrella", "white shark", "white wolf"] _snake_case : Optional[Any] = pipe.get_label_ids(snake_case_ ) _snake_case : Tuple = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=40 , output_type="np" ).images for word, image in zip(snake_case_ , snake_case_ ): _snake_case : str = load_numpy( F'https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/dit/{word}.npy' ) assert np.abs((expected_image - image).max() ) < 1E-2 def lowerCamelCase__ ( self ): _snake_case : int = DiTPipeline.from_pretrained("facebook/DiT-XL-2-512" ) _snake_case : Optional[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.to("cuda" ) _snake_case : str = ["vase", "umbrella"] _snake_case : Any = pipe.get_label_ids(snake_case_ ) _snake_case : List[str] = torch.manual_seed(0 ) _snake_case : Tuple = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="np" ).images for word, image in zip(snake_case_ , snake_case_ ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" F'/dit/{word}_512.npy' ) assert np.abs((expected_image - image).max() ) < 1E-1
714
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def a__ ( a : List[str] , a : Any ): """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _snake_case : Any = flax_key_tuple[:-1] + ("weight",) _snake_case : str = torch.permute(a , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(a ): # linear layer _snake_case : Optional[int] = flax_key_tuple[:-1] + ("weight",) _snake_case : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _snake_case : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) return flax_key_tuple, flax_tensor def a__ ( a : List[Any] , a : Union[str, Any] , a : List[str] ): """simple docstring""" if "metadata" in layer: _snake_case : Optional[int] = layer.split("metadata" ) _snake_case : Optional[int] = "".join(split_layer[0] )[:-1] _snake_case : int = [tuple(("metadata" + split_layer[1]).split("/" ) )] elif "kvstore" in layer: _snake_case : Any = layer.split("kvstore" ) _snake_case : str = "".join(split_layer[0] )[:-1] _snake_case : Any = [tuple(("kvstore" + split_layer[1]).split("/" ) )] else: _snake_case : List[Any] = layer.split("/" ) _snake_case : Tuple = "/".join(split_layer[:-1] ) _snake_case : int = (split_layer[-1],) if "kvstore/path" in layer: _snake_case : Optional[Any] = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _snake_case : Tuple = "file" else: _snake_case : Optional[int] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def a__ ( a : List[Any] , a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = rename_keys(a ) _snake_case : int = {} for k, v in current_block.items(): _snake_case : Optional[int] = v _snake_case : Optional[int] = new_current_block torch.save(a , a ) def a__ ( a : Dict , a : Tuple , a : List[str] , a : int , a : str = WEIGHTS_NAME ): """simple docstring""" _snake_case : Any = convert_file_size_to_int(a ) _snake_case : Tuple = [] _snake_case : Optional[int] = {} _snake_case : Tuple = 0 _snake_case : Optional[Any] = 0 os.makedirs(a , exist_ok=a ) with gfile.GFile(switch_checkpoint_path + "/checkpoint" , "rb" ) as fp: _snake_case : Any = serialization.msgpack_restore(fp.read() )["optimizer"]["target"] _snake_case : Optional[Any] = flatten_dict(a , sep="/" ) _snake_case : Optional[Any] = {} for layer in checkpoint_info.keys(): _snake_case , _snake_case , _snake_case : int = get_key_and_tensorstore_dict( a , a , a ) if curr_real_layer_name in all_layers: _snake_case : Dict = content else: _snake_case : Tuple = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _snake_case : List[str] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _snake_case : Dict = torch.tensor(a ) _snake_case : Dict = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _snake_case , _snake_case : Optional[int] = rename_base_flax_keys(tuple(key.split("/" ) ) , a ) _snake_case : Optional[Any] = "/".join(a ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _snake_case : Any = os.path.join( a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) del current_block _snake_case : List[Any] = {} _snake_case : str = 0 _snake_case : List[str] = raw_weights.to(getattr(a , a ) ) current_block_size += weight_size total_size += weight_size # Add the last block _snake_case : int = os.path.join(a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(a ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _snake_case : str = {} _snake_case : Any = {} for idx, shard in enumerate(a ): _snake_case : Optional[int] = weights_name.replace( ".bin" , f'-{idx+1:05d}-of-{len(a ):05d}.bin' ) # len(sharded_state_dicts):05d} _snake_case : Dict = os.path.join(a , weights_name.replace(".bin" , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(a , os.path.join(a , a ) ) _snake_case : Dict = shard for key in shard: _snake_case : int = shard_file # Add the metadata _snake_case : List[Any] = {"total_size": total_size} _snake_case : Any = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(a , a ) , "w" , encoding="utf-8" ) as f: _snake_case : Union[str, Any] = json.dumps(a , indent=2 , sort_keys=a ) + "\n" f.write(a ) return metadata, index if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) _a : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def a__ ( ): """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _snake_case : List[str] = SwitchTransformersConfig.from_pretrained("google/switch-base-8" ) config.save_pretrained("/home/arthur_huggingface_co/transformers/switch_converted" ) _snake_case : str = SwitchTransformersForConditionalGeneration.from_pretrained( "/home/arthur_huggingface_co/transformers/switch_converted" , device_map="auto" ) _snake_case : List[Any] = TaTokenizer.from_pretrained("t5-small" ) _snake_case : Optional[Any] = "A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>." _snake_case : Dict = tokenizer(a , return_tensors="pt" ).input_ids _snake_case : List[Any] = model.generate(a , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
87
0
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
715
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
87
0
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( _snake_case ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
716
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : Optional[int] = { """configuration_whisper""": ["""WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP""", """WhisperConfig""", """WhisperOnnxConfig"""], """feature_extraction_whisper""": ["""WhisperFeatureExtractor"""], """processing_whisper""": ["""WhisperProcessor"""], """tokenization_whisper""": ["""WhisperTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Tuple = ["""WhisperTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = [ """WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """WhisperForConditionalGeneration""", """WhisperModel""", """WhisperPreTrainedModel""", """WhisperForAudioClassification""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ """TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFWhisperForConditionalGeneration""", """TFWhisperModel""", """TFWhisperPreTrainedModel""", ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = [ """FlaxWhisperForConditionalGeneration""", """FlaxWhisperModel""", """FlaxWhisperPreTrainedModel""", """FlaxWhisperForAudioClassification""", ] if TYPE_CHECKING: from .configuration_whisper import WHISPER_PRETRAINED_CONFIG_ARCHIVE_MAP, WhisperConfig, WhisperOnnxConfig from .feature_extraction_whisper import WhisperFeatureExtractor from .processing_whisper import WhisperProcessor from .tokenization_whisper import WhisperTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_whisper_fast import WhisperTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_whisper import ( WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, WhisperForAudioClassification, WhisperForConditionalGeneration, WhisperModel, WhisperPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_whisper import ( TF_WHISPER_PRETRAINED_MODEL_ARCHIVE_LIST, TFWhisperForConditionalGeneration, TFWhisperModel, TFWhisperPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_whisper import ( FlaxWhisperForAudioClassification, FlaxWhisperForConditionalGeneration, FlaxWhisperModel, FlaxWhisperPreTrainedModel, ) else: import sys _a : Tuple = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
717
"""simple docstring""" def a__ ( a : list , a : int , a : int = 0 , a : int = 0 ): """simple docstring""" _snake_case : Optional[int] = right or len(a ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(a , a , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_=3 , snake_case_=32 , snake_case_=3 , snake_case_=10 , snake_case_=[10, 20, 30, 40] , snake_case_=[1, 1, 2, 1] , snake_case_=True , snake_case_=True , snake_case_="relu" , snake_case_=3 , snake_case_=None , ): _snake_case : List[Any] = parent _snake_case : Optional[Any] = batch_size _snake_case : Optional[Any] = image_size _snake_case : str = num_channels _snake_case : str = embeddings_size _snake_case : str = hidden_sizes _snake_case : int = depths _snake_case : Optional[int] = is_training _snake_case : Optional[Any] = use_labels _snake_case : List[Any] = hidden_act _snake_case : Union[str, Any] = num_labels _snake_case : int = scope _snake_case : List[Any] = len(snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case : List[Any] = None if self.use_labels: _snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_labels ) _snake_case : Dict = self.get_config() return config, pixel_values, labels def lowerCamelCase__ ( self ): return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = RegNetModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : str = model(snake_case_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_labels _snake_case : List[str] = RegNetForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : List[Any] = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.prepare_config_and_inputs() _snake_case : int = config_and_inputs _snake_case : str = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () __lowercase : Dict = ( {"""feature-extraction""": RegNetModel, """image-classification""": RegNetForImageClassification} if is_torch_available() else {} ) __lowercase : Optional[Any] = False __lowercase : Dict = False __lowercase : List[Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self ): _snake_case : Optional[int] = RegNetModelTester(self ) _snake_case : int = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ ) def lowerCamelCase__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase__ ( self ): return @unittest.skip(reason="RegNet does not use inputs_embeds" ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="RegNet does not support input and output embeddings" ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Any = model_class(snake_case_ ) _snake_case : Optional[int] = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : List[str] = [*signature.parameters.keys()] _snake_case : Dict = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Union[str, Any] = model_class(config=snake_case_ ) for name, module in model.named_modules(): if isinstance(snake_case_ , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=F'Parameter {name} of model {model_class} seems not properly initialized' , ) def lowerCamelCase__ ( self ): def check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[str] = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): _snake_case : Tuple = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) _snake_case : str = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case : Optional[int] = self.model_tester.num_stages self.assertEqual(len(snake_case_ ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) _snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : str = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: _snake_case : str = layer_type _snake_case : List[str] = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : str = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : Tuple = RegNetModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def a__ ( ): """simple docstring""" _snake_case : str = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase): @cached_property def lowerCamelCase__ ( self ): return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowerCamelCase__ ( self ): _snake_case : Tuple = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(snake_case_ ) _snake_case : Dict = self.default_image_processor _snake_case : List[Any] = prepare_img() _snake_case : Tuple = image_processor(images=snake_case_ , return_tensors="pt" ).to(snake_case_ ) # forward pass with torch.no_grad(): _snake_case : int = model(**snake_case_ ) # verify the logits _snake_case : List[str] = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case_ ) _snake_case : Optional[int] = torch.tensor([-0.4180, -1.5051, -3.4836] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1E-4 ) )
718
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ ): _snake_case , _snake_case : Dict = text, pattern _snake_case , _snake_case : int = len(snake_case_ ), len(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCamelCase__ ( self ): # searches pattern in text and returns index positions _snake_case : List[str] = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case : Union[str, Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: _snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) _snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _a : List[Any] = """ABAABA""" _a : str = """AB""" _a : List[Any] = BoyerMooreSearch(text, pattern) _a : Any = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _a : Union[str, Any] = { """configuration_nezha""": ["""NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP""", """NezhaConfig"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ """NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST""", """NezhaForNextSentencePrediction""", """NezhaForMaskedLM""", """NezhaForPreTraining""", """NezhaForMultipleChoice""", """NezhaForQuestionAnswering""", """NezhaForSequenceClassification""", """NezhaForTokenClassification""", """NezhaModel""", """NezhaPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_nezha import NEZHA_PRETRAINED_CONFIG_ARCHIVE_MAP, NezhaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_nezha import ( NEZHA_PRETRAINED_MODEL_ARCHIVE_LIST, NezhaForMaskedLM, NezhaForMultipleChoice, NezhaForNextSentencePrediction, NezhaForPreTraining, NezhaForQuestionAnswering, NezhaForSequenceClassification, NezhaForTokenClassification, NezhaModel, NezhaPreTrainedModel, ) else: import sys _a : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
719
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
87
0
"""simple docstring""" import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import logging _a : int = logging.get_logger(__name__) _a : int = {"""vocab_file""": """sentencepiece.bpe.model"""} _a : List[Any] = { """vocab_file""": { """moussaKam/mbarthez""": """https://huggingface.co/moussaKam/mbarthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez""": """https://huggingface.co/moussaKam/barthez/resolve/main/sentencepiece.bpe.model""", """moussaKam/barthez-orangesum-title""": ( """https://huggingface.co/moussaKam/barthez-orangesum-title/resolve/main/sentencepiece.bpe.model""" ), }, } _a : Optional[Any] = { """moussaKam/mbarthez""": 1_024, """moussaKam/barthez""": 1_024, """moussaKam/barthez-orangesum-title""": 1_024, } _a : Union[str, Any] = """▁""" class _UpperCAmelCase ( _snake_case): __lowercase : Union[str, Any] = VOCAB_FILES_NAMES __lowercase : Union[str, Any] = PRETRAINED_VOCAB_FILES_MAP __lowercase : Any = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Union[str, Any] = ["""input_ids""", """attention_mask"""] def __init__( self , snake_case_ , snake_case_="<s>" , snake_case_="</s>" , snake_case_="</s>" , snake_case_="<s>" , snake_case_="<unk>" , snake_case_="<pad>" , snake_case_="<mask>" , snake_case_ = None , **snake_case_ , ): # Mask token behave like a normal word, i.e. include the space before it _snake_case : List[str] = AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token _snake_case : Optional[Any] = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , cls_token=snake_case_ , pad_token=snake_case_ , mask_token=snake_case_ , sp_model_kwargs=self.sp_model_kwargs , **snake_case_ , ) _snake_case : Optional[Any] = vocab_file _snake_case : Tuple = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(snake_case_ ) ) _snake_case : Optional[Any] = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} _snake_case : Any = len(self.sp_model ) - 1 _snake_case : int = {v: k for k, v in self.fairseq_tokens_to_ids.items()} def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] _snake_case : List[Any] = [self.cls_token_id] _snake_case : Union[str, Any] = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None , snake_case_ = False ): if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=snake_case_ , token_ids_a=snake_case_ , already_has_special_tokens=snake_case_ ) if token_ids_a is None: return [1] + ([0] * len(snake_case_ )) + [1] return [1] + ([0] * len(snake_case_ )) + [1, 1] + ([0] * len(snake_case_ )) + [1] def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Dict = [self.sep_token_id] _snake_case : List[str] = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] @property def lowerCamelCase__ ( self ): return len(self.sp_model ) def lowerCamelCase__ ( self ): _snake_case : List[str] = {self.convert_ids_to_tokens(snake_case_ ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowerCamelCase__ ( self , snake_case_ ): return self.sp_model.encode(snake_case_ , out_type=snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] _snake_case : Optional[int] = self.sp_model.PieceToId(snake_case_ ) return spm_id if spm_id else self.unk_token_id def lowerCamelCase__ ( self , snake_case_ ): if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Optional[Any] = [] _snake_case : List[Any] = "" _snake_case : Dict = False for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: if not prev_is_special: out_string += " " out_string += self.sp_model.decode(snake_case_ ) + token _snake_case : Tuple = True _snake_case : int = [] else: current_sub_tokens.append(snake_case_ ) _snake_case : Dict = False out_string += self.sp_model.decode(snake_case_ ) return out_string.strip() def __getstate__( self ): _snake_case : Dict = self.__dict__.copy() _snake_case : int = None return state def __setstate__( self , snake_case_ ): _snake_case : str = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): _snake_case : List[str] = {} _snake_case : Any = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _snake_case : Any = os.path.join( snake_case_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , snake_case_ ) elif not os.path.isfile(self.vocab_file ): with open(snake_case_ , "wb" ) as fi: _snake_case : Optional[int] = self.sp_model.serialized_model_proto() fi.write(snake_case_ ) return (out_vocab_file,)
720
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_albert import AlbertTokenizer else: _a : str = None _a : List[str] = logging.get_logger(__name__) _a : Any = {"""vocab_file""": """spiece.model""", """tokenizer_file""": """tokenizer.json"""} _a : Tuple = { """vocab_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/spiece.model""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/spiece.model""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/spiece.model""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/spiece.model""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/spiece.model""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/spiece.model""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/spiece.model""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/spiece.model""", }, """tokenizer_file""": { """albert-base-v1""": """https://huggingface.co/albert-base-v1/resolve/main/tokenizer.json""", """albert-large-v1""": """https://huggingface.co/albert-large-v1/resolve/main/tokenizer.json""", """albert-xlarge-v1""": """https://huggingface.co/albert-xlarge-v1/resolve/main/tokenizer.json""", """albert-xxlarge-v1""": """https://huggingface.co/albert-xxlarge-v1/resolve/main/tokenizer.json""", """albert-base-v2""": """https://huggingface.co/albert-base-v2/resolve/main/tokenizer.json""", """albert-large-v2""": """https://huggingface.co/albert-large-v2/resolve/main/tokenizer.json""", """albert-xlarge-v2""": """https://huggingface.co/albert-xlarge-v2/resolve/main/tokenizer.json""", """albert-xxlarge-v2""": """https://huggingface.co/albert-xxlarge-v2/resolve/main/tokenizer.json""", }, } _a : Optional[Any] = { """albert-base-v1""": 512, """albert-large-v1""": 512, """albert-xlarge-v1""": 512, """albert-xxlarge-v1""": 512, """albert-base-v2""": 512, """albert-large-v2""": 512, """albert-xlarge-v2""": 512, """albert-xxlarge-v2""": 512, } _a : Dict = """▁""" class _UpperCAmelCase ( _snake_case): __lowercase : Dict = VOCAB_FILES_NAMES __lowercase : int = PRETRAINED_VOCAB_FILES_MAP __lowercase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : Tuple = AlbertTokenizer def __init__( self , snake_case_=None , snake_case_=None , snake_case_=True , snake_case_=True , snake_case_=False , snake_case_="[CLS]" , snake_case_="[SEP]" , snake_case_="<unk>" , snake_case_="[SEP]" , snake_case_="<pad>" , snake_case_="[CLS]" , snake_case_="[MASK]" , **snake_case_ , ): # Mask token behave like a normal word, i.e. include the space before it and # is included in the raw text, there should be a match in a non-normalized sentence. _snake_case : str = ( AddedToken(snake_case_ , lstrip=snake_case_ , rstrip=snake_case_ , normalized=snake_case_ ) if isinstance(snake_case_ , snake_case_ ) else mask_token ) super().__init__( snake_case_ , tokenizer_file=snake_case_ , do_lower_case=snake_case_ , remove_space=snake_case_ , keep_accents=snake_case_ , bos_token=snake_case_ , eos_token=snake_case_ , unk_token=snake_case_ , sep_token=snake_case_ , pad_token=snake_case_ , cls_token=snake_case_ , mask_token=snake_case_ , **snake_case_ , ) _snake_case : Dict = do_lower_case _snake_case : Tuple = remove_space _snake_case : Optional[int] = keep_accents _snake_case : Optional[int] = vocab_file _snake_case : Dict = False if not self.vocab_file else True def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Any = [self.sep_token_id] _snake_case : Tuple = [self.cls_token_id] if token_ids_a is None: return cls + token_ids_a + sep return cls + token_ids_a + sep + token_ids_a + sep def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Union[str, Any] = [self.sep_token_id] _snake_case : Any = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(snake_case_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _snake_case : List[str] = os.path.join( snake_case_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(snake_case_ ): copyfile(self.vocab_file , snake_case_ ) return (out_vocab_file,)
721
"""simple docstring""" import argparse import json import subprocess def a__ ( a : Optional[Any] , a : Optional[int] ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[Any] = ( f'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _snake_case : Dict = subprocess.run(a , shell=a , stdout=subprocess.PIPE ) _snake_case : Tuple = output.stdout.decode("utf-8" ) _snake_case : List[str] = json.loads(a ) _snake_case : Any = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(a ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(a ) ) if len(a ) > 0: _snake_case : Any = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def a__ ( a : Optional[int] ): """simple docstring""" return values.split("," ) _a : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _a : List[str] = parser.parse_args() get_runner_status(args.target_runners, args.token)
87
0
import argparse import requests import torch from PIL import Image from transformers import SwinConfig, SwinForMaskedImageModeling, ViTImageProcessor def a__ ( a : List[str] ): """simple docstring""" _snake_case : Dict = SwinConfig(image_size=192 ) if "base" in model_name: _snake_case : List[Any] = 6 _snake_case : Dict = 128 _snake_case : Dict = (2, 2, 18, 2) _snake_case : Tuple = (4, 8, 16, 32) elif "large" in model_name: _snake_case : Dict = 12 _snake_case : Tuple = 192 _snake_case : List[str] = (2, 2, 18, 2) _snake_case : Tuple = (6, 12, 24, 48) else: raise ValueError("Model not supported, only supports base and large variants" ) _snake_case : Tuple = window_size _snake_case : Union[str, Any] = embed_dim _snake_case : int = depths _snake_case : Tuple = num_heads return config def a__ ( a : List[str] ): """simple docstring""" if "encoder.mask_token" in name: _snake_case : Union[str, Any] = name.replace("encoder.mask_token" , "embeddings.mask_token" ) if "encoder.patch_embed.proj" in name: _snake_case : Any = name.replace("encoder.patch_embed.proj" , "embeddings.patch_embeddings.projection" ) if "encoder.patch_embed.norm" in name: _snake_case : Any = name.replace("encoder.patch_embed.norm" , "embeddings.norm" ) if "attn.proj" in name: _snake_case : Optional[int] = name.replace("attn.proj" , "attention.output.dense" ) if "attn" in name: _snake_case : Tuple = name.replace("attn" , "attention.self" ) if "norm1" in name: _snake_case : Any = name.replace("norm1" , "layernorm_before" ) if "norm2" in name: _snake_case : Optional[Any] = name.replace("norm2" , "layernorm_after" ) if "mlp.fc1" in name: _snake_case : int = name.replace("mlp.fc1" , "intermediate.dense" ) if "mlp.fc2" in name: _snake_case : Union[str, Any] = name.replace("mlp.fc2" , "output.dense" ) if name == "encoder.norm.weight": _snake_case : int = "layernorm.weight" if name == "encoder.norm.bias": _snake_case : str = "layernorm.bias" if "decoder" in name: pass else: _snake_case : List[Any] = "swin." + name return name def a__ ( a : int , a : Dict ): """simple docstring""" for key in orig_state_dict.copy().keys(): _snake_case : List[Any] = orig_state_dict.pop(a ) if "attn_mask" in key: pass elif "qkv" in key: _snake_case : Union[str, Any] = key.split("." ) _snake_case : List[str] = int(key_split[2] ) _snake_case : Union[str, Any] = int(key_split[4] ) _snake_case : List[Any] = model.swin.encoder.layers[layer_num].blocks[block_num].attention.self.all_head_size if "weight" in key: _snake_case : List[Any] = val[:dim, :] _snake_case : Optional[int] = val[ dim : dim * 2, : ] _snake_case : Any = val[-dim:, :] else: _snake_case : Optional[int] = val[ :dim ] _snake_case : List[Any] = val[ dim : dim * 2 ] _snake_case : Any = val[ -dim: ] else: _snake_case : str = val return orig_state_dict def a__ ( a : Optional[Any] , a : Optional[Any] , a : Optional[Any] , a : List[str] ): """simple docstring""" _snake_case : str = torch.load(a , map_location="cpu" )["model"] _snake_case : Union[str, Any] = get_swin_config(a ) _snake_case : Optional[Any] = SwinForMaskedImageModeling(a ) model.eval() _snake_case : List[str] = convert_state_dict(a , a ) model.load_state_dict(a ) _snake_case : Tuple = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case : List[str] = ViTImageProcessor(size={"height": 192, "width": 192} ) _snake_case : int = Image.open(requests.get(a , stream=a ).raw ) _snake_case : int = image_processor(images=a , return_tensors="pt" ) with torch.no_grad(): _snake_case : str = model(**a ).logits print(outputs.keys() ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: print(f'Saving model {model_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(a ) if push_to_hub: print(f'Pushing model and image processor for {model_name} to hub' ) model.push_to_hub(f'microsoft/{model_name}' ) image_processor.push_to_hub(f'microsoft/{model_name}' ) if __name__ == "__main__": _a : Optional[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--model_name""", default="""swin-base-simmim-window6-192""", type=str, choices=["""swin-base-simmim-window6-192""", """swin-large-simmim-window12-192"""], help="""Name of the Swin SimMIM model you'd like to convert.""", ) parser.add_argument( """--checkpoint_path""", default="""/Users/nielsrogge/Documents/SwinSimMIM/simmim_pretrain__swin_base__img192_window6__100ep.pth""", type=str, help="""Path to the original PyTorch checkpoint (.pth file).""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _a : Optional[Any] = parser.parse_args() convert_swin_checkpoint(args.model_name, args.checkpoint_path, args.pytorch_dump_folder_path, args.push_to_hub)
700
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
87
0
import unittest from transformers import AutoTokenizer, FalconConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ( FalconForCausalLM, FalconForQuestionAnswering, FalconForSequenceClassification, FalconForTokenClassification, FalconModel, ) class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_=3 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=False , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : List[Any] = parent _snake_case : Dict = batch_size _snake_case : int = seq_length _snake_case : int = is_training _snake_case : Tuple = use_input_mask _snake_case : str = use_token_type_ids _snake_case : Optional[int] = use_labels _snake_case : Tuple = vocab_size _snake_case : Optional[int] = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : Dict = num_attention_heads _snake_case : Dict = intermediate_size _snake_case : Any = hidden_act _snake_case : List[str] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : Optional[int] = max_position_embeddings _snake_case : Union[str, Any] = type_vocab_size _snake_case : List[str] = type_sequence_label_size _snake_case : Union[str, Any] = initializer_range _snake_case : List[str] = num_labels _snake_case : Union[str, Any] = num_choices _snake_case : Union[str, Any] = scope def lowerCamelCase__ ( self ): _snake_case : List[str] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : str = None if self.use_input_mask: _snake_case : str = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : Tuple = None _snake_case : Any = None _snake_case : Any = None _snake_case : int = None if self.use_labels: _snake_case : str = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Tuple = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : int = self.get_config() return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self ): return FalconConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=snake_case_ , initializer_range=self.initializer_range , pad_token_id=1 , new_decoder_architecture=snake_case_ , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = FalconModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Dict = model(snake_case_ , attention_mask=snake_case_ ) _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _snake_case : Optional[int] = True _snake_case : Optional[int] = FalconModel(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Tuple = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , ) _snake_case : Any = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , ) _snake_case : Tuple = model(snake_case_ , attention_mask=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _snake_case : Dict = FalconForCausalLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Dict = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ): _snake_case : Optional[Any] = True _snake_case : Dict = True _snake_case : List[str] = FalconForCausalLM(config=snake_case_ ) model.to(snake_case_ ) model.eval() # first forward pass _snake_case : int = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , use_cache=snake_case_ , ) _snake_case : List[Any] = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _snake_case : Union[str, Any] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : Union[str, Any] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _snake_case : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) _snake_case : List[Any] = torch.cat([input_mask, next_mask] , dim=-1 ) _snake_case : Optional[int] = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , output_hidden_states=snake_case_ , )["hidden_states"][0] _snake_case : Tuple = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , past_key_values=snake_case_ , output_hidden_states=snake_case_ , )["hidden_states"][0] # select random slice _snake_case : List[str] = ids_tensor((1,) , output_from_past.shape[-1] ).item() _snake_case : Union[str, Any] = output_from_no_past[:, -3:, random_slice_idx].detach() _snake_case : Tuple = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case_ , snake_case_ , atol=1E-3 ) ) def lowerCamelCase__ ( self ): _snake_case : List[str] = self.prepare_config_and_inputs() ( _snake_case ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( _snake_case , _snake_case , _snake_case , unittest.TestCase): __lowercase : str = ( ( FalconModel, FalconForCausalLM, FalconForSequenceClassification, FalconForTokenClassification, FalconForQuestionAnswering, ) if is_torch_available() else () ) __lowercase : Dict = (FalconForCausalLM,) if is_torch_available() else () __lowercase : str = ( { """feature-extraction""": FalconModel, """text-classification""": FalconForSequenceClassification, """text-generation""": FalconForCausalLM, """question-answering""": FalconForQuestionAnswering, """token-classification""": FalconForTokenClassification, """zero-shot""": FalconForSequenceClassification, } if is_torch_available() else {} ) __lowercase : Optional[Any] = False __lowercase : List[str] = False def lowerCamelCase__ ( self ): _snake_case : str = FalconModelTester(self ) _snake_case : Optional[int] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() for alibi in [True, False]: _snake_case : Optional[int] = alibi self.model_tester.create_and_check_model(snake_case_ , *snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Dict = 3 _snake_case : List[str] = input_dict["input_ids"] _snake_case : Any = input_ids.ne(1 ).to(snake_case_ ) _snake_case : Dict = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _snake_case : int = FalconForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Any = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Optional[Any] = 3 _snake_case : int = "single_label_classification" _snake_case : Any = input_dict["input_ids"] _snake_case : List[Any] = input_ids.ne(1 ).to(snake_case_ ) _snake_case : Union[str, Any] = ids_tensor([self.model_tester.batch_size] , self.model_tester.type_sequence_label_size ) _snake_case : int = FalconForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Optional[Any] = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Optional[int] = input_dict["input_ids"] _snake_case : Optional[Any] = FalconForCausalLM(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : str = model(snake_case_ , use_cache=snake_case_ ) _snake_case : Any = input_ids.shape[0] _snake_case : Union[str, Any] = model._convert_to_rw_cache(result.past_key_values ) _snake_case : Optional[Any] = model._convert_cache_to_standard_format(snake_case_ , snake_case_ ) for layer in range(len(snake_case_ ) ): for tensor_idx in range(2 ): self.assertTrue(rw_cache[layer][tensor_idx].ndim == 3 ) self.assertTrue(result.past_key_values[layer][tensor_idx].ndim == 4 ) self.assertTrue( torch.all(result.past_key_values[layer][tensor_idx] == standard_cache[layer][tensor_idx] ) ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Tuple = 3 _snake_case : Optional[Any] = "multi_label_classification" _snake_case : Optional[int] = input_dict["input_ids"] _snake_case : Dict = input_ids.ne(1 ).to(snake_case_ ) _snake_case : Optional[int] = ids_tensor( [self.model_tester.batch_size, config.num_labels] , self.model_tester.type_sequence_label_size ).to(torch.float ) _snake_case : Any = FalconForSequenceClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : int = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.assertEqual(result.logits.shape , (self.model_tester.batch_size, self.model_tester.num_labels) ) def lowerCamelCase__ ( self ): # Falcon can have different numbers of KV-heads than the number of query heads, so we need # to override this test to use the right head counts. for model_class in self.all_generative_model_classes: _snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_common() # If it doesn't support cache, pass the test if not hasattr(snake_case_ , "use_cache" ): return _snake_case : Dict = model_class(snake_case_ ).to(snake_case_ ) if "use_cache" not in inputs: _snake_case : str = True _snake_case : Optional[Any] = model(**snake_case_ ) # If "past_key_values" is not returned, pass the test (e.g. RWKV uses a different cache name and format) if "past_key_values" not in outputs: return _snake_case : str = ( getattr(snake_case_ , "decoder_layers" , snake_case_ ) or getattr(snake_case_ , "num_decoder_layers" , snake_case_ ) or config.num_hidden_layers ) _snake_case : Tuple = getattr(snake_case_ , "num_kv_heads" , config.num_attention_heads ) _snake_case : Dict = getattr(snake_case_ , "d_model" , config.hidden_size ) _snake_case : Dict = embed_dim // num_attention_heads _snake_case : str = outputs["past_key_values"] self.assertEqual(len(snake_case_ ) , snake_case_ ) _snake_case : str = inputs["input_ids"].shape for i in range(snake_case_ ): if config.new_decoder_architecture: _snake_case : Any = config.num_attention_heads elif config.multi_query: _snake_case : str = 1 self.assertEqual(len(past_kv[0] ) , 2 ) # K V for the decoder = 2 self.assertEqual( past_kv[i][0].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) self.assertEqual( past_kv[i][1].shape , (batch_size, num_attention_heads, seq_length, per_head_embed_dim) ) @require_torch class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : List[str] = AutoTokenizer.from_pretrained("Rocketknight1/falcon-rw-1b" ) _snake_case : List[str] = FalconForCausalLM.from_pretrained("Rocketknight1/falcon-rw-1b" ) model.eval() model.to(snake_case_ ) _snake_case : str = tokenizer("My favorite food is" , return_tensors="pt" ).to(snake_case_ ) _snake_case : Union[str, Any] = ( "My favorite food is pizza. I love it so much that I have a pizza party every year for my birthday." ) _snake_case : Dict = model.generate(**snake_case_ , do_sample=snake_case_ , max_new_tokens=19 ) _snake_case : List[str] = tokenizer.batch_decode(snake_case_ )[0] self.assertEqual(snake_case_ , snake_case_ ) @slow def lowerCamelCase__ ( self ): # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers for repo in ["Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b"]: _snake_case : Optional[int] = AutoTokenizer.from_pretrained(snake_case_ ) _snake_case : List[Any] = FalconForCausalLM.from_pretrained(snake_case_ ) model.eval() model.to(snake_case_ ) _snake_case : List[str] = tokenizer("My favorite food is" , return_tensors="pt" ).to(snake_case_ ) # We just test that these run without errors - the models are randomly initialized # and so the actual text outputs will be garbage model.generate(**snake_case_ , do_sample=snake_case_ , max_new_tokens=4 ) model.generate(**snake_case_ , do_sample=snake_case_ , max_new_tokens=4 ) model.generate(**snake_case_ , num_beams=2 , max_new_tokens=4 ) @slow def lowerCamelCase__ ( self ): # The big models are way too big for the CI, so we use tiny random models that resemble their # architectures but with much smaller and fewer layers with torch.no_grad(): for repo in [ "Rocketknight1/falcon-rw-1b", "Rocketknight1/tiny-random-falcon-7b", "Rocketknight1/tiny-random-falcon-40b", ]: _snake_case : Any = AutoTokenizer.from_pretrained(snake_case_ ) _snake_case : Tuple = FalconForCausalLM.from_pretrained(snake_case_ ) model.eval() model.to(device=snake_case_ ) _snake_case : List[Any] = tokenizer("My favorite food is" , return_tensors="pt" ).to(snake_case_ ) # Test results are the same with and without cache _snake_case : Dict = model.generate(**snake_case_ , do_sample=snake_case_ , max_new_tokens=20 , use_cache=snake_case_ ) _snake_case : Optional[int] = model.generate(**snake_case_ , do_sample=snake_case_ , max_new_tokens=20 , use_cache=snake_case_ ) self.assertTrue((outputs_cache - outputs_no_cache).sum().item() == 0 )
701
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import json import os from typing import Optional, Tuple import regex as re from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging _a : Optional[Any] = logging.get_logger(__name__) _a : str = { """vocab_file""": """vocab.json""", """merges_file""": """merges.txt""", } _a : Tuple = { """vocab_file""": {"""ctrl""": """https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-vocab.json"""}, """merges_file""": {"""ctrl""": """https://raw.githubusercontent.com/salesforce/ctrl/master/ctrl-merges.txt"""}, } _a : Optional[int] = { """ctrl""": 256, } _a : Any = { """Pregnancy""": 168_629, """Christianity""": 7_675, """Explain""": 106_423, """Fitness""": 63_440, """Saving""": 63_163, """Ask""": 27_171, """Ass""": 95_985, """Joke""": 163_509, """Questions""": 45_622, """Thoughts""": 49_605, """Retail""": 52_342, """Feminism""": 164_338, """Writing""": 11_992, """Atheism""": 192_263, """Netflix""": 48_616, """Computing""": 39_639, """Opinion""": 43_213, """Alone""": 44_967, """Funny""": 58_917, """Gaming""": 40_358, """Human""": 4_088, """India""": 1_331, """Joker""": 77_138, """Diet""": 36_206, """Legal""": 11_859, """Norman""": 4_939, """Tip""": 72_689, """Weight""": 52_343, """Movies""": 46_273, """Running""": 23_425, """Science""": 2_090, """Horror""": 37_793, """Confession""": 60_572, """Finance""": 12_250, """Politics""": 16_360, """Scary""": 191_985, """Support""": 12_654, """Technologies""": 32_516, """Teenage""": 66_160, """Event""": 32_769, """Learned""": 67_460, """Notion""": 182_770, """Wikipedia""": 37_583, """Books""": 6_665, """Extract""": 76_050, """Confessions""": 102_701, """Conspiracy""": 75_932, """Links""": 63_674, """Narcissus""": 150_425, """Relationship""": 54_766, """Relationships""": 134_796, """Reviews""": 41_671, """News""": 4_256, """Translation""": 26_820, """multilingual""": 128_406, } def a__ ( a : List[str] ): """simple docstring""" _snake_case : List[str] = set() _snake_case : Dict = word[0] for char in word[1:]: pairs.add((prev_char, char) ) _snake_case : Union[str, Any] = char _snake_case : Dict = set(a ) return pairs class _UpperCAmelCase ( _snake_case): __lowercase : int = VOCAB_FILES_NAMES __lowercase : str = PRETRAINED_VOCAB_FILES_MAP __lowercase : List[str] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __lowercase : List[Any] = CONTROL_CODES def __init__( self , snake_case_ , snake_case_ , snake_case_="<unk>" , **snake_case_ ): super().__init__(unk_token=snake_case_ , **snake_case_ ) with open(snake_case_ , encoding="utf-8" ) as vocab_handle: _snake_case : Dict = json.load(snake_case_ ) _snake_case : Dict = {v: k for k, v in self.encoder.items()} with open(snake_case_ , encoding="utf-8" ) as merges_handle: _snake_case : str = merges_handle.read().split("\n" )[1:-1] _snake_case : int = [tuple(merge.split() ) for merge in merges] _snake_case : Tuple = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) _snake_case : int = {} @property def lowerCamelCase__ ( self ): return len(self.encoder ) def lowerCamelCase__ ( self ): return dict(self.encoder , **self.added_tokens_encoder ) def lowerCamelCase__ ( self , snake_case_ ): if token in self.cache: return self.cache[token] _snake_case : List[str] = tuple(snake_case_ ) _snake_case : Tuple = tuple(list(word[:-1] ) + [word[-1] + "</w>"] ) _snake_case : List[Any] = get_pairs(snake_case_ ) if not pairs: return token while True: _snake_case : Any = min(snake_case_ , key=lambda snake_case_ : self.bpe_ranks.get(snake_case_ , float("inf" ) ) ) if bigram not in self.bpe_ranks: break _snake_case : List[str] = bigram _snake_case : Any = [] _snake_case : Dict = 0 while i < len(snake_case_ ): try: _snake_case : int = word.index(snake_case_ , snake_case_ ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) _snake_case : Optional[Any] = j if word[i] == first and i < len(snake_case_ ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 _snake_case : str = tuple(snake_case_ ) _snake_case : Tuple = new_word if len(snake_case_ ) == 1: break else: _snake_case : Tuple = get_pairs(snake_case_ ) _snake_case : Dict = "@@ ".join(snake_case_ ) _snake_case : int = word[:-4] _snake_case : str = word return word def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Optional[int] = [] _snake_case : List[str] = re.findall(r"\S+\n?" , snake_case_ ) for token in words: split_tokens.extend(list(self.bpe(snake_case_ ).split(" " ) ) ) return split_tokens def lowerCamelCase__ ( self , snake_case_ ): return self.encoder.get(snake_case_ , self.encoder.get(self.unk_token ) ) def lowerCamelCase__ ( self , snake_case_ ): return self.decoder.get(snake_case_ , self.unk_token ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Optional[int] = " ".join(snake_case_ ).replace("@@ " , "" ).strip() return out_string def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): if not os.path.isdir(snake_case_ ): logger.error(F'Vocabulary path ({save_directory}) should be a directory' ) return _snake_case : Any = os.path.join( snake_case_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) _snake_case : List[Any] = os.path.join( snake_case_ , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(snake_case_ , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=snake_case_ , ensure_ascii=snake_case_ ) + "\n" ) _snake_case : Any = 0 with open(snake_case_ , "w" , encoding="utf-8" ) as writer: writer.write("#version: 0.2\n" ) for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda snake_case_ : kv[1] ): if index != token_index: logger.warning( F'Saving vocabulary to {merge_file}: BPE merge indices are not consecutive.' " Please check that the tokenizer is not corrupted!" ) _snake_case : str = token_index writer.write(" ".join(snake_case_ ) + "\n" ) index += 1 return vocab_file, merge_file # def decode(self, token_ids, skip_special_tokens=False, clean_up_tokenization_spaces=True): # filtered_tokens = ' '.join(self.convert_ids_to_tokens(token_ids, skip_special_tokens=skip_special_tokens)) # tokens_generated_so_far = re.sub('(@@ )', '', string=filtered_tokens) # tokens_generated_so_far = re.sub('(@@ ?$)', '', string=tokens_generated_so_far) # return ''.join(tokens_generated_so_far)
702
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Any = TextToVideoSDPipeline __lowercase : str = TEXT_TO_IMAGE_PARAMS __lowercase : int = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __lowercase : Optional[int] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) _snake_case : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : Tuple = CLIPTextModel(snake_case_ ) _snake_case : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : str = torch.manual_seed(snake_case_ ) else: _snake_case : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def lowerCamelCase__ ( self ): _snake_case : int = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : Optional[Any] = self.get_dummy_components() _snake_case : Tuple = TextToVideoSDPipeline(**snake_case_ ) _snake_case : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ ) _snake_case : Union[str, Any] = "np" _snake_case : Dict = sd_pipe(**snake_case_ ).frames _snake_case : Any = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _snake_case : Dict = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=1E-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) _snake_case : int = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _snake_case : Tuple = pipe.to("cuda" ) _snake_case : List[Any] = "Spiderman is surfing" _snake_case : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : int = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="pt" ).frames _snake_case : int = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def lowerCamelCase__ ( self ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) _snake_case : str = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : int = pipe.to("cuda" ) _snake_case : Any = "Spiderman is surfing" _snake_case : str = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Any = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type="pt" ).frames _snake_case : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_torch_available, is_vision_available, ) _a : int = { """configuration_blip""": [ """BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BlipConfig""", """BlipTextConfig""", """BlipVisionConfig""", ], """processing_blip""": ["""BlipProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Any = ["""BlipImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """BlipModel""", """BlipPreTrainedModel""", """BlipForConditionalGeneration""", """BlipForQuestionAnswering""", """BlipVisionModel""", """BlipTextModel""", """BlipForImageTextRetrieval""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = [ """TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBlipModel""", """TFBlipPreTrainedModel""", """TFBlipForConditionalGeneration""", """TFBlipForQuestionAnswering""", """TFBlipVisionModel""", """TFBlipTextModel""", """TFBlipForImageTextRetrieval""", ] if TYPE_CHECKING: from .configuration_blip import BLIP_PRETRAINED_CONFIG_ARCHIVE_MAP, BlipConfig, BlipTextConfig, BlipVisionConfig from .processing_blip import BlipProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_blip import BlipImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_blip import ( BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, BlipForConditionalGeneration, BlipForImageTextRetrieval, BlipForQuestionAnswering, BlipModel, BlipPreTrainedModel, BlipTextModel, BlipVisionModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_blip import ( TF_BLIP_PRETRAINED_MODEL_ARCHIVE_LIST, TFBlipForConditionalGeneration, TFBlipForImageTextRetrieval, TFBlipForQuestionAnswering, TFBlipModel, TFBlipPreTrainedModel, TFBlipTextModel, TFBlipVisionModel, ) else: import sys _a : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
703
"""simple docstring""" from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _UpperCAmelCase ( _snake_case): __lowercase : int = """EncodecFeatureExtractor""" __lowercase : str = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) _snake_case : Dict = self.feature_extractor _snake_case : Any = False def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_=True ): return self.tokenizer.get_decoder_prompt_ids(task=snake_case_ , language=snake_case_ , no_timestamps=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*snake_case_ , **snake_case_ ) _snake_case : str = kwargs.pop("audio" , snake_case_ ) _snake_case : Optional[int] = kwargs.pop("sampling_rate" , snake_case_ ) _snake_case : Optional[Any] = kwargs.pop("text" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Any = args[0] _snake_case : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: _snake_case : Any = self.tokenizer(snake_case_ , **snake_case_ ) if audio is not None: _snake_case : Any = self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) if audio is None: return inputs elif text is None: return audio_inputs else: _snake_case : str = audio_inputs["input_values"] if "padding_mask" in audio_inputs: _snake_case : List[str] = audio_inputs["padding_mask"] return inputs def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): _snake_case : Tuple = kwargs.pop("audio" , snake_case_ ) _snake_case : List[str] = kwargs.pop("padding_mask" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Dict = args[1:] if audio_values is not None: return self._decode_audio(snake_case_ , padding_mask=snake_case_ ) else: return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Optional[int] = to_numpy(snake_case_ ) _snake_case , _snake_case , _snake_case : Tuple = audio_values.shape if padding_mask is None: return list(snake_case_ ) _snake_case : Optional[int] = to_numpy(snake_case_ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _snake_case : Any = seq_len - padding_mask.shape[-1] _snake_case : Optional[Any] = 1 - self.feature_extractor.padding_value _snake_case : Optional[int] = np.pad(snake_case_ , ((0, 0), (0, difference)) , "constant" , constant_values=snake_case_ ) _snake_case : Any = audio_values.tolist() for i in range(snake_case_ ): _snake_case : Tuple = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _snake_case : Tuple = sliced_audio.reshape(snake_case_ , -1 ) return audio_values
87
0
from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _UpperCAmelCase ( _snake_case): __lowercase : int = """EncodecFeatureExtractor""" __lowercase : str = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) _snake_case : Dict = self.feature_extractor _snake_case : Any = False def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_=True ): return self.tokenizer.get_decoder_prompt_ids(task=snake_case_ , language=snake_case_ , no_timestamps=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*snake_case_ , **snake_case_ ) _snake_case : str = kwargs.pop("audio" , snake_case_ ) _snake_case : Optional[int] = kwargs.pop("sampling_rate" , snake_case_ ) _snake_case : Optional[Any] = kwargs.pop("text" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Any = args[0] _snake_case : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: _snake_case : Any = self.tokenizer(snake_case_ , **snake_case_ ) if audio is not None: _snake_case : Any = self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) if audio is None: return inputs elif text is None: return audio_inputs else: _snake_case : str = audio_inputs["input_values"] if "padding_mask" in audio_inputs: _snake_case : List[str] = audio_inputs["padding_mask"] return inputs def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): _snake_case : Tuple = kwargs.pop("audio" , snake_case_ ) _snake_case : List[str] = kwargs.pop("padding_mask" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Dict = args[1:] if audio_values is not None: return self._decode_audio(snake_case_ , padding_mask=snake_case_ ) else: return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Optional[int] = to_numpy(snake_case_ ) _snake_case : Tuple = audio_values.shape if padding_mask is None: return list(snake_case_ ) _snake_case : Optional[int] = to_numpy(snake_case_ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _snake_case : Any = seq_len - padding_mask.shape[-1] _snake_case : Optional[Any] = 1 - self.feature_extractor.padding_value _snake_case : Optional[int] = np.pad(snake_case_ , ((0, 0), (0, difference)) , "constant" , constant_values=snake_case_ ) _snake_case : Any = audio_values.tolist() for i in range(snake_case_ ): _snake_case : Tuple = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _snake_case : Tuple = sliced_audio.reshape(snake_case_ , -1 ) return audio_values
704
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : str = {"""configuration_yolos""": ["""YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """YolosConfig""", """YolosOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""YolosFeatureExtractor"""] _a : List[Any] = ["""YolosImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST""", """YolosForObjectDetection""", """YolosModel""", """YolosPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from ..models.auto import AutoModelForSeqaSeqLM, AutoTokenizer from .base import PipelineTool class _UpperCAmelCase ( _snake_case): __lowercase : Dict = """philschmid/bart-large-cnn-samsum""" __lowercase : Tuple = ( """This is a tool that summarizes an English text. It takes an input `text` containing the text to summarize, """ """and returns a summary of the text.""" ) __lowercase : Any = """summarizer""" __lowercase : List[Any] = AutoTokenizer __lowercase : Optional[int] = AutoModelForSeqaSeqLM __lowercase : int = ["""text"""] __lowercase : List[str] = ["""text"""] def lowerCamelCase__ ( self , snake_case_ ): return self.pre_processor(snake_case_ , return_tensors="pt" , truncation=snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): return self.model.generate(**snake_case_ )[0] def lowerCamelCase__ ( self , snake_case_ ): return self.pre_processor.decode(snake_case_ , skip_special_tokens=snake_case_ , clean_up_tokenization_spaces=snake_case_ )
705
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = dataset _snake_case : str = process _snake_case : int = params def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): _snake_case : Union[str, Any] = self.dataset[i] _snake_case : Optional[Any] = self.process(snake_case_ , **self.params ) return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): _snake_case : Union[str, Any] = loader _snake_case : Tuple = infer _snake_case : List[Any] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _snake_case : int = None _snake_case : int = loader_batch_size # Internal bookkeeping _snake_case : Any = None _snake_case : Dict = None def __len__( self ): return len(self.loader ) def __iter__( self ): _snake_case : int = iter(self.loader ) return self def lowerCamelCase__ ( self ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _snake_case : List[Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _snake_case : int = {} for k, element in self._loader_batch_data.items(): if isinstance(snake_case_ , snake_case_ ): # Convert ModelOutput to tuple first _snake_case : Tuple = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _snake_case : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(snake_case_ , snake_case_ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _snake_case : Tuple = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _snake_case : Tuple = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : List[Any] = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : Union[str, Any] = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _snake_case : List[Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _snake_case : int = self._loader_batch_data.__class__(snake_case_ ) self._loader_batch_index += 1 return result def lowerCamelCase__ ( self ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _snake_case : Tuple = next(self.iterator ) _snake_case : Any = self.infer(snake_case_ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Optional[int] = list(processed.keys() )[0] _snake_case : List[str] = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = len(snake_case_ ) else: _snake_case : Optional[int] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Union[str, Any] = observed_batch_size # Setting internal index to unwrap the batch _snake_case : str = processed _snake_case : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): super().__init__(snake_case_ , snake_case_ , snake_case_ ) def __iter__( self ): _snake_case : Tuple = iter(self.loader ) _snake_case : List[Any] = None return self def lowerCamelCase__ ( self ): if self.subiterator is None: _snake_case : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _snake_case : Union[str, Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _snake_case : str = self.infer(next(self.iterator ) , **self.params ) _snake_case : Tuple = next(self.subiterator ) return processed class _UpperCAmelCase ( _snake_case): def __iter__( self ): _snake_case : Optional[Any] = iter(self.loader ) return self def lowerCamelCase__ ( self ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _snake_case : Optional[Any] = False _snake_case : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : str = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator while not is_last: _snake_case : List[str] = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Tuple = list(processed.keys() )[0] _snake_case : Tuple = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Any = len(snake_case_ ) else: _snake_case : List[Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Dict = observed_batch_size _snake_case : List[Any] = processed _snake_case : List[str] = 0 while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : int = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator else: _snake_case : Dict = processed _snake_case : Dict = item.pop("is_last" ) accumulator.append(snake_case_ ) return accumulator class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ ): _snake_case : str = dataset _snake_case : Any = key def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return self.dataset[i][self.key] class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = dataset _snake_case : Any = keya _snake_case : int = keya def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
87
0
"""simple docstring""" import argparse import torch from transformers import ( WavaVecaConfig, WavaVecaFeatureExtractor, WavaVecaForAudioFrameClassification, WavaVecaForSequenceClassification, WavaVecaForXVector, logging, ) logging.set_verbosity_info() lowerCamelCase_ : Optional[int] = logging.get_logger(__name__) def a__ ( a : List[str] , a : Dict , a : List[str] ): """simple docstring""" _snake_case : Tuple = WavaVecaForSequenceClassification.from_pretrained(a , config=a ) _snake_case : Tuple = downstream_dict["projector.weight"] _snake_case : int = downstream_dict["projector.bias"] _snake_case : int = downstream_dict["model.post_net.linear.weight"] _snake_case : Union[str, Any] = downstream_dict["model.post_net.linear.bias"] return model def a__ ( a : int , a : str , a : Tuple ): """simple docstring""" _snake_case : Any = WavaVecaForAudioFrameClassification.from_pretrained(a , config=a ) _snake_case : int = downstream_dict["model.linear.weight"] _snake_case : List[str] = downstream_dict["model.linear.bias"] return model def a__ ( a : str , a : Union[str, Any] , a : Dict ): """simple docstring""" _snake_case : Any = WavaVecaForXVector.from_pretrained(a , config=a ) _snake_case : int = downstream_dict["connector.weight"] _snake_case : Dict = downstream_dict["connector.bias"] for i, kernel_size in enumerate(hf_config.tdnn_kernel ): _snake_case : Optional[Any] = downstream_dict[ f'model.framelevel_feature_extractor.module.{i}.kernel.weight' ] _snake_case : Union[str, Any] = downstream_dict[f'model.framelevel_feature_extractor.module.{i}.kernel.bias'] _snake_case : List[Any] = downstream_dict["model.utterancelevel_feature_extractor.linear1.weight"] _snake_case : List[Any] = downstream_dict["model.utterancelevel_feature_extractor.linear1.bias"] _snake_case : List[Any] = downstream_dict["model.utterancelevel_feature_extractor.linear2.weight"] _snake_case : Dict = downstream_dict["model.utterancelevel_feature_extractor.linear2.bias"] _snake_case : Dict = downstream_dict["objective.W"] return model @torch.no_grad() def a__ ( a : List[Any] , a : List[Any] , a : List[str] , a : int ): """simple docstring""" _snake_case : Union[str, Any] = torch.load(a , map_location="cpu" ) _snake_case : List[str] = checkpoint["Downstream"] _snake_case : Any = WavaVecaConfig.from_pretrained(a ) _snake_case : Dict = WavaVecaFeatureExtractor.from_pretrained( a , return_attention_mask=a , do_normalize=a ) _snake_case : Optional[Any] = hf_config.architectures[0] if arch.endswith("ForSequenceClassification" ): _snake_case : Any = convert_classification(a , a , a ) elif arch.endswith("ForAudioFrameClassification" ): _snake_case : List[Any] = convert_diarization(a , a , a ) elif arch.endswith("ForXVector" ): _snake_case : str = convert_xvector(a , a , a ) else: raise NotImplementedError(f'S3PRL weights conversion is not supported for {arch}' ) if hf_config.use_weighted_layer_sum: _snake_case : str = checkpoint["Featurizer"]["weights"] hf_feature_extractor.save_pretrained(a ) hf_model.save_pretrained(a ) if __name__ == "__main__": lowerCamelCase_ : List[str] = argparse.ArgumentParser() parser.add_argument( """--base_model_name""", default=None, type=str, help="""Name of the huggingface pretrained base model.""" ) parser.add_argument("""--config_path""", default=None, type=str, help="""Path to the huggingface classifier config.""") parser.add_argument("""--checkpoint_path""", default=None, type=str, help="""Path to the s3prl checkpoint.""") parser.add_argument("""--model_dump_path""", default=None, type=str, help="""Path to the final converted model.""") lowerCamelCase_ : int = parser.parse_args() convert_saprl_checkpoint(args.base_model_name, args.config_path, args.checkpoint_path, args.model_dump_path)
706
"""simple docstring""" def a__ ( a : int ): """simple docstring""" if not isinstance(a , a ): raise TypeError("Input value must be an 'int' type" ) _snake_case : Union[str, Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import warnings from typing import Dict import numpy as np from ..utils import ExplicitEnum, add_end_docstrings, is_tf_available, is_torch_available from .base import PIPELINE_INIT_ARGS, GenericTensor, Pipeline if is_tf_available(): from ..models.auto.modeling_tf_auto import TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if is_torch_available(): from ..models.auto.modeling_auto import MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING def a__ ( a : Dict ): """simple docstring""" return 1.0 / (1.0 + np.exp(-_outputs )) def a__ ( a : List[str] ): """simple docstring""" _snake_case : Tuple = np.max(_outputs , axis=-1 , keepdims=a ) _snake_case : Tuple = np.exp(_outputs - maxes ) return shifted_exp / shifted_exp.sum(axis=-1 , keepdims=a ) class _UpperCAmelCase ( _snake_case): __lowercase : Dict = """sigmoid""" __lowercase : Any = """softmax""" __lowercase : Optional[int] = """none""" @add_end_docstrings( _snake_case , r""" return_all_scores (`bool`, *optional*, defaults to `False`): Whether to return all prediction scores or just the one of the predicted class. function_to_apply (`str`, *optional*, defaults to `\"default\"`): The function to apply to the model outputs in order to retrieve the scores. Accepts four different values: - `\"default\"`: if the model has a single label, will apply the sigmoid function on the output. If the model has several labels, will apply the softmax function on the output. - `\"sigmoid\"`: Applies the sigmoid function on the output. - `\"softmax\"`: Applies the softmax function on the output. - `\"none\"`: Does not apply any function on the output. """ , ) class _UpperCAmelCase ( _snake_case): __lowercase : Any = False __lowercase : Dict = ClassificationFunction.NONE def __init__( self , **snake_case_ ): super().__init__(**snake_case_ ) self.check_model_type( TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING if self.framework == "tf" else MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING ) def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_="" , **snake_case_ ): # Using "" as default argument because we're going to use `top_k=None` in user code to declare # "No top_k" _snake_case : str = tokenizer_kwargs _snake_case : str = {} if hasattr(self.model.config , "return_all_scores" ) and return_all_scores is None: _snake_case : str = self.model.config.return_all_scores if isinstance(snake_case_ , snake_case_ ) or top_k is None: _snake_case : Optional[Any] = top_k _snake_case : List[str] = False elif return_all_scores is not None: warnings.warn( "`return_all_scores` is now deprecated, if want a similar functionality use `top_k=None` instead of" " `return_all_scores=True` or `top_k=1` instead of `return_all_scores=False`." , snake_case_ , ) if return_all_scores: _snake_case : Optional[int] = None else: _snake_case : Dict = 1 if isinstance(snake_case_ , snake_case_ ): _snake_case : List[Any] = ClassificationFunction[function_to_apply.upper()] if function_to_apply is not None: _snake_case : List[str] = function_to_apply return preprocess_params, {}, postprocess_params def __call__( self , *snake_case_ , **snake_case_ ): _snake_case : Dict = super().__call__(*snake_case_ , **snake_case_ ) # TODO try and retrieve it in a nicer way from _sanitize_parameters. _snake_case : Tuple = "top_k" not in kwargs if isinstance(args[0] , snake_case_ ) and _legacy: # This pipeline is odd, and return a list when single item is run return [result] else: return result def lowerCamelCase__ ( self , snake_case_ , **snake_case_ ): _snake_case : str = self.framework if isinstance(snake_case_ , snake_case_ ): return self.tokenizer(**snake_case_ , return_tensors=snake_case_ , **snake_case_ ) elif isinstance(snake_case_ , snake_case_ ) and len(snake_case_ ) == 1 and isinstance(inputs[0] , snake_case_ ) and len(inputs[0] ) == 2: # It used to be valid to use a list of list of list for text pairs, keeping this path for BC return self.tokenizer( text=inputs[0][0] , text_pair=inputs[0][1] , return_tensors=snake_case_ , **snake_case_ ) elif isinstance(snake_case_ , snake_case_ ): # This is likely an invalid usage of the pipeline attempting to pass text pairs. raise ValueError( "The pipeline received invalid inputs, if you are trying to send text pairs, you can try to send a" " dictionary `{\"text\": \"My text\", \"text_pair\": \"My pair\"}` in order to send a text pair." ) return self.tokenizer(snake_case_ , return_tensors=snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): return self.model(**snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_=None , snake_case_=1 , snake_case_=True ): # `_legacy` is used to determine if we're running the naked pipeline and in backward # compatibility mode, or if running the pipeline with `pipeline(..., top_k=1)` we're running # the more natural result containing the list. # Default value before `set_parameters` if function_to_apply is None: if self.model.config.problem_type == "multi_label_classification" or self.model.config.num_labels == 1: _snake_case : Dict = ClassificationFunction.SIGMOID elif self.model.config.problem_type == "single_label_classification" or self.model.config.num_labels > 1: _snake_case : Union[str, Any] = ClassificationFunction.SOFTMAX elif hasattr(self.model.config , "function_to_apply" ) and function_to_apply is None: _snake_case : Any = self.model.config.function_to_apply else: _snake_case : str = ClassificationFunction.NONE _snake_case : Optional[Any] = model_outputs["logits"][0] _snake_case : Optional[Any] = outputs.numpy() if function_to_apply == ClassificationFunction.SIGMOID: _snake_case : Optional[Any] = sigmoid(snake_case_ ) elif function_to_apply == ClassificationFunction.SOFTMAX: _snake_case : Tuple = softmax(snake_case_ ) elif function_to_apply == ClassificationFunction.NONE: _snake_case : Dict = outputs else: raise ValueError(F'Unrecognized `function_to_apply` argument: {function_to_apply}' ) if top_k == 1 and _legacy: return {"label": self.model.config.idalabel[scores.argmax().item()], "score": scores.max().item()} _snake_case : List[Any] = [ {"label": self.model.config.idalabel[i], "score": score.item()} for i, score in enumerate(snake_case_ ) ] if not _legacy: dict_scores.sort(key=lambda snake_case_ : x["score"] , reverse=snake_case_ ) if top_k is not None: _snake_case : int = dict_scores[:top_k] return dict_scores
707
"""simple docstring""" from __future__ import annotations import requests _a : List[str] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def a__ ( a : str , a : int = 1 , a : str = "new" , a : list | None = None ): """simple docstring""" _snake_case : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(a ) - valid_terms ) ): _snake_case : Optional[int] = f'Invalid search term: {invalid_search_terms}' raise ValueError(a ) _snake_case : int = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"User-agent": "A random string"} , ) if response.status_code == 429: raise requests.HTTPError _snake_case : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(a )} _snake_case : Tuple = {} for id_ in range(a ): _snake_case : List[str] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
87
0
"""simple docstring""" import gc import random import unittest import numpy as np import torch from transformers import ( CLIPImageProcessor, CLIPTextConfig, CLIPTextModelWithProjection, CLIPTokenizer, CLIPVisionConfig, CLIPVisionModelWithProjection, ) from diffusers import ( DiffusionPipeline, UnCLIPImageVariationPipeline, UnCLIPScheduler, UNetaDConditionModel, UNetaDModel, ) from diffusers.pipelines.unclip.text_proj import UnCLIPTextProjModel from diffusers.utils import floats_tensor, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, load_image, require_torch_gpu, skip_mps from ..pipeline_params import IMAGE_VARIATION_BATCH_PARAMS, IMAGE_VARIATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin, assert_mean_pixel_difference enable_full_determinism() class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Union[str, Any] = UnCLIPImageVariationPipeline __lowercase : Optional[int] = IMAGE_VARIATION_PARAMS - {"""height""", """width""", """guidance_scale"""} __lowercase : List[Any] = IMAGE_VARIATION_BATCH_PARAMS __lowercase : Any = [ """generator""", """return_dict""", """decoder_num_inference_steps""", """super_res_num_inference_steps""", ] __lowercase : str = False @property def lowerCamelCase__ ( self ): return 32 @property def lowerCamelCase__ ( self ): return 32 @property def lowerCamelCase__ ( self ): return self.time_input_dim @property def lowerCamelCase__ ( self ): return self.time_input_dim * 4 @property def lowerCamelCase__ ( self ): return 1_00 @property def lowerCamelCase__ ( self ): _snake_case : str = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) return tokenizer @property def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : Optional[int] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , ) return CLIPTextModelWithProjection(snake_case_ ) @property def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : Optional[int] = CLIPVisionConfig( hidden_size=self.text_embedder_hidden_size , projection_dim=self.text_embedder_hidden_size , num_hidden_layers=5 , num_attention_heads=4 , image_size=32 , intermediate_size=37 , patch_size=1 , ) return CLIPVisionModelWithProjection(snake_case_ ) @property def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : int = { "clip_embeddings_dim": self.text_embedder_hidden_size, "time_embed_dim": self.time_embed_dim, "cross_attention_dim": self.cross_attention_dim, } _snake_case : int = UnCLIPTextProjModel(**snake_case_ ) return model @property def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : Union[str, Any] = { "sample_size": 32, # RGB in channels "in_channels": 3, # Out channels is double in channels because predicts mean and variance "out_channels": 6, "down_block_types": ("ResnetDownsampleBlock2D", "SimpleCrossAttnDownBlock2D"), "up_block_types": ("SimpleCrossAttnUpBlock2D", "ResnetUpsampleBlock2D"), "mid_block_type": "UNetMidBlock2DSimpleCrossAttn", "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "layers_per_block": 1, "cross_attention_dim": self.cross_attention_dim, "attention_head_dim": 4, "resnet_time_scale_shift": "scale_shift", "class_embed_type": "identity", } _snake_case : Any = UNetaDConditionModel(**snake_case_ ) return model @property def lowerCamelCase__ ( self ): return { "sample_size": 64, "layers_per_block": 1, "down_block_types": ("ResnetDownsampleBlock2D", "ResnetDownsampleBlock2D"), "up_block_types": ("ResnetUpsampleBlock2D", "ResnetUpsampleBlock2D"), "block_out_channels": (self.block_out_channels_a, self.block_out_channels_a * 2), "in_channels": 6, "out_channels": 3, } @property def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : Tuple = UNetaDModel(**self.dummy_super_res_kwargs ) return model @property def lowerCamelCase__ ( self ): # seeded differently to get different unet than `self.dummy_super_res_first` torch.manual_seed(1 ) _snake_case : List[str] = UNetaDModel(**self.dummy_super_res_kwargs ) return model def lowerCamelCase__ ( self ): _snake_case : Any = self.dummy_decoder _snake_case : Union[str, Any] = self.dummy_text_proj _snake_case : Union[str, Any] = self.dummy_text_encoder _snake_case : List[str] = self.dummy_tokenizer _snake_case : Union[str, Any] = self.dummy_super_res_first _snake_case : str = self.dummy_super_res_last _snake_case : int = UnCLIPScheduler( variance_type="learned_range" , prediction_type="epsilon" , num_train_timesteps=10_00 , ) _snake_case : Any = UnCLIPScheduler( variance_type="fixed_small_log" , prediction_type="epsilon" , num_train_timesteps=10_00 , ) _snake_case : Any = CLIPImageProcessor(crop_size=32 , size=32 ) _snake_case : Optional[Any] = self.dummy_image_encoder return { "decoder": decoder, "text_encoder": text_encoder, "tokenizer": tokenizer, "text_proj": text_proj, "feature_extractor": feature_extractor, "image_encoder": image_encoder, "super_res_first": super_res_first, "super_res_last": super_res_last, "decoder_scheduler": decoder_scheduler, "super_res_scheduler": super_res_scheduler, } def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 , snake_case_=True ): _snake_case : int = floats_tensor((1, 3, 32, 32) , rng=random.Random(snake_case_ ) ).to(snake_case_ ) if str(snake_case_ ).startswith("mps" ): _snake_case : List[Any] = torch.manual_seed(snake_case_ ) else: _snake_case : int = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) if pil_image: _snake_case : Optional[int] = input_image * 0.5 + 0.5 _snake_case : Union[str, Any] = input_image.clamp(0 , 1 ) _snake_case : List[str] = input_image.cpu().permute(0 , 2 , 3 , 1 ).float().numpy() _snake_case : Optional[int] = DiffusionPipeline.numpy_to_pil(snake_case_ )[0] return { "image": input_image, "generator": generator, "decoder_num_inference_steps": 2, "super_res_num_inference_steps": 2, "output_type": "np", } def lowerCamelCase__ ( self ): _snake_case : List[str] = "cpu" _snake_case : List[Any] = self.get_dummy_components() _snake_case : Any = self.pipeline_class(**snake_case_ ) _snake_case : Tuple = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : List[Any] = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : Any = pipe(**snake_case_ ) _snake_case : int = output.images _snake_case : Optional[Any] = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : Optional[Any] = pipe( **snake_case_ , return_dict=snake_case_ , )[0] _snake_case : str = image[0, -3:, -3:, -1] _snake_case : Any = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _snake_case : List[str] = np.array( [ 0.9997, 0.0002, 0.9997, 0.9997, 0.9969, 0.0023, 0.9997, 0.9969, 0.9970, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): _snake_case : int = "cpu" _snake_case : str = self.get_dummy_components() _snake_case : Union[str, Any] = self.pipeline_class(**snake_case_ ) _snake_case : List[Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : List[str] = pipe(**snake_case_ ) _snake_case : Optional[Any] = output.images _snake_case : int = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : Optional[int] = pipe( **snake_case_ , return_dict=snake_case_ , )[0] _snake_case : Optional[int] = image[0, -3:, -3:, -1] _snake_case : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (1, 64, 64, 3) _snake_case : Any = np.array([0.9997, 0.0003, 0.9997, 0.9997, 0.9970, 0.0024, 0.9997, 0.9971, 0.9971] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): _snake_case : List[str] = "cpu" _snake_case : str = self.get_dummy_components() _snake_case : Tuple = self.pipeline_class(**snake_case_ ) _snake_case : Optional[Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Optional[Any] = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : Tuple = [ pipeline_inputs["image"], pipeline_inputs["image"], ] _snake_case : int = pipe(**snake_case_ ) _snake_case : Dict = output.images _snake_case : str = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : Tuple = [ tuple_pipeline_inputs["image"], tuple_pipeline_inputs["image"], ] _snake_case : List[Any] = pipe( **snake_case_ , return_dict=snake_case_ , )[0] _snake_case : Optional[Any] = image[0, -3:, -3:, -1] _snake_case : Tuple = image_from_tuple[0, -3:, -3:, -1] assert image.shape == (2, 64, 64, 3) _snake_case : Optional[Any] = np.array( [ 0.9997, 0.9989, 0.0008, 0.0021, 0.9960, 0.0018, 0.0014, 0.0002, 0.9933, ] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 assert np.abs(image_from_tuple_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): _snake_case : Optional[int] = torch.device("cpu" ) class _UpperCAmelCase : __lowercase : List[str] = 1 _snake_case : Any = self.get_dummy_components() _snake_case : List[Any] = self.pipeline_class(**snake_case_ ) _snake_case : Union[str, Any] = pipe.to(snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Dict = torch.Generator(device=snake_case_ ).manual_seed(0 ) _snake_case : List[Any] = pipe.decoder.dtype _snake_case : int = 1 _snake_case : Optional[int] = ( batch_size, pipe.decoder.config.in_channels, pipe.decoder.config.sample_size, pipe.decoder.config.sample_size, ) _snake_case : List[str] = pipe.prepare_latents( snake_case_ , dtype=snake_case_ , device=snake_case_ , generator=snake_case_ , latents=snake_case_ , scheduler=DummyScheduler() ) _snake_case : Optional[int] = ( batch_size, pipe.super_res_first.config.in_channels // 2, pipe.super_res_first.config.sample_size, pipe.super_res_first.config.sample_size, ) _snake_case : Optional[int] = pipe.prepare_latents( snake_case_ , dtype=snake_case_ , device=snake_case_ , generator=snake_case_ , latents=snake_case_ , scheduler=DummyScheduler() ) _snake_case : Any = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) _snake_case : str = pipe( **snake_case_ , decoder_latents=snake_case_ , super_res_latents=snake_case_ ).images _snake_case : Optional[int] = self.get_dummy_inputs(snake_case_ , pil_image=snake_case_ ) # Don't pass image, instead pass embedding _snake_case : List[Any] = pipeline_inputs.pop("image" ) _snake_case : List[str] = pipe.image_encoder(snake_case_ ).image_embeds _snake_case : Dict = pipe( **snake_case_ , decoder_latents=snake_case_ , super_res_latents=snake_case_ , image_embeddings=snake_case_ , ).images # make sure passing text embeddings manually is identical assert np.abs(img_out_a - img_out_a ).max() < 1E-4 @skip_mps def lowerCamelCase__ ( self ): _snake_case : str = torch_device == "cpu" # Check is relaxed because there is not a torch 2.0 sliced attention added kv processor _snake_case : Optional[int] = 1E-2 self._test_attention_slicing_forward_pass( test_max_difference=snake_case_ , expected_max_diff=snake_case_ ) @skip_mps def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = torch_device == "cpu" _snake_case : Optional[Any] = True _snake_case : List[Any] = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] self._test_inference_batch_single_identical( test_max_difference=snake_case_ , relax_max_difference=snake_case_ , additional_params_copy_to_batched_inputs=snake_case_ , ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = [ "decoder_num_inference_steps", "super_res_num_inference_steps", ] if torch_device == "mps": # TODO: MPS errors with larger batch sizes _snake_case : Dict = [2, 3] self._test_inference_batch_consistent( batch_sizes=snake_case_ , additional_params_copy_to_batched_inputs=snake_case_ , ) else: self._test_inference_batch_consistent( additional_params_copy_to_batched_inputs=snake_case_ ) @skip_mps def lowerCamelCase__ ( self ): return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowerCamelCase__ ( self ): return super().test_save_load_local() @skip_mps def lowerCamelCase__ ( self ): return super().test_save_load_optional_components() @slow @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): # clean up the VRAM after each test super().tearDown() gc.collect() torch.cuda.empty_cache() def lowerCamelCase__ ( self ): _snake_case : Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/unclip/cat.png" ) _snake_case : Dict = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/unclip/karlo_v1_alpha_cat_variation_fp16.npy" ) _snake_case : Any = UnCLIPImageVariationPipeline.from_pretrained( "kakaobrain/karlo-v1-alpha-image-variations" , torch_dtype=torch.floataa ) _snake_case : Union[str, Any] = pipeline.to(snake_case_ ) pipeline.set_progress_bar_config(disable=snake_case_ ) _snake_case : Dict = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Tuple = pipeline( snake_case_ , generator=snake_case_ , output_type="np" , ) _snake_case : Tuple = output.images[0] assert image.shape == (2_56, 2_56, 3) assert_mean_pixel_difference(snake_case_ , snake_case_ , 15 )
708
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def a__ ( a : float , a : float , a : bool = False ): """simple docstring""" if radian_mode: return [magnitude * cos(a ), magnitude * sin(a )] return [magnitude * cos(radians(a ) ), magnitude * sin(radians(a ) )] def a__ ( a : NDArray[floataa] , a : NDArray[floataa] , a : float = 10**-1 ): """simple docstring""" _snake_case : NDArray[floataa] = cross(a , a ) _snake_case : float = sum(a ) return abs(a ) < eps if __name__ == "__main__": # Test to check if it works _a : Tuple = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) _a : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg _a : List[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) _a : List[Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg _a : List[str] = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) _a : Optional[int] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
87
0
"""simple docstring""" from torch import nn def a__ ( a : Optional[Any] ): """simple docstring""" if act_fn in ["swish", "silu"]: return nn.SiLU() elif act_fn == "mish": return nn.Mish() elif act_fn == "gelu": return nn.GELU() else: raise ValueError(f'Unsupported activation function: {act_fn}' )
709
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[int] = logging.get_logger(__name__) _a : str = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class _UpperCAmelCase ( _snake_case): __lowercase : Optional[Any] = """openai-gpt""" __lowercase : Dict = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , snake_case_=4_04_78 , snake_case_=5_12 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_="cls_index" , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=0.1 , **snake_case_ , ): _snake_case : Tuple = vocab_size _snake_case : Dict = n_positions _snake_case : Any = n_embd _snake_case : Any = n_layer _snake_case : Optional[int] = n_head _snake_case : Union[str, Any] = afn _snake_case : Dict = resid_pdrop _snake_case : str = embd_pdrop _snake_case : Union[str, Any] = attn_pdrop _snake_case : str = layer_norm_epsilon _snake_case : Union[str, Any] = initializer_range _snake_case : Any = summary_type _snake_case : List[str] = summary_use_proj _snake_case : Optional[int] = summary_activation _snake_case : Union[str, Any] = summary_first_dropout _snake_case : Optional[int] = summary_proj_to_labels super().__init__(**snake_case_ )
87
0
"""simple docstring""" import numpy as np import pandas as pd from sklearn.preprocessing import MinMaxScaler from tensorflow.keras.layers import LSTM, Dense from tensorflow.keras.models import Sequential if __name__ == "__main__": _a : Tuple = pd.read_csv("""sample_data.csv""", header=None) _a : List[Any] = df.shape[:1][0] # If you're using some other dataset input the target column _a : int = df.iloc[:, 1:2] _a : List[Any] = actual_data.values.reshape(len_data, 1) _a : Optional[Any] = MinMaxScaler().fit_transform(actual_data) _a : Optional[Any] = 10 _a : int = 5 _a : List[Any] = 20 _a : Dict = len_data - periods * look_back _a : Any = actual_data[:division] _a : str = actual_data[division - look_back :] _a : Any = [], [] _a : Any = [], [] for i in range(0, len(train_data) - forward_days - look_back + 1): train_x.append(train_data[i : i + look_back]) train_y.append(train_data[i + look_back : i + look_back + forward_days]) for i in range(0, len(test_data) - forward_days - look_back + 1): test_x.append(test_data[i : i + look_back]) test_y.append(test_data[i + look_back : i + look_back + forward_days]) _a : Dict = np.array(train_x) _a : Tuple = np.array(test_x) _a : str = np.array([list(i.ravel()) for i in train_y]) _a : Optional[int] = np.array([list(i.ravel()) for i in test_y]) _a : Optional[int] = Sequential() model.add(LSTM(128, input_shape=(look_back, 1), return_sequences=True)) model.add(LSTM(64, input_shape=(128, 1))) model.add(Dense(forward_days)) model.compile(loss="""mean_squared_error""", optimizer="""adam""") _a : List[str] = model.fit( x_train, y_train, epochs=150, verbose=1, shuffle=True, batch_size=4 ) _a : Tuple = model.predict(x_test)
710
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _a : Tuple = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _a : int = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def a__ ( a : List[str] , a : int , a : int ): """simple docstring""" _snake_case : Union[str, Any] = state_dict.pop(a ) _snake_case : Union[str, Any] = val def a__ ( a : Tuple ): """simple docstring""" _snake_case : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case : Dict = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) _snake_case : Tuple = value else: _snake_case : Dict = value return new_state_dict def a__ ( a : int ): """simple docstring""" _snake_case : Any = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[:256, :] _snake_case : List[str] = in_proj_bias[:256] _snake_case : Optional[Any] = in_proj_weight[256:512, :] _snake_case : List[str] = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : Dict = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case : List[Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : Union[str, Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : Union[str, Any] = in_proj_weight[:256, :] _snake_case : Tuple = in_proj_bias[:256] _snake_case : int = in_proj_weight[256:512, :] _snake_case : int = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : str = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _snake_case : Dict = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _snake_case : Optional[int] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case : Dict = in_proj_weight_cross_attn[:256, :] _snake_case : Any = in_proj_bias_cross_attn[:256] _snake_case : Union[str, Any] = in_proj_weight_cross_attn[256:512, :] _snake_case : Optional[int] = in_proj_bias_cross_attn[256:512] _snake_case : Any = in_proj_weight_cross_attn[-256:, :] _snake_case : str = in_proj_bias_cross_attn[-256:] def a__ ( a : str , a : int ): """simple docstring""" _snake_case , _snake_case : List[str] = image.size _snake_case : Dict = max(a , a ) _snake_case : Union[str, Any] = 800 if "detection" in checkpoint_url else 1_000 _snake_case : Any = target_max_size / current_max_size _snake_case : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def a__ ( a : str ): """simple docstring""" _snake_case : str = F.to_tensor(a ) _snake_case : Union[str, Any] = F.normalize(a , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def a__ ( a : Optional[Any] , a : Any , a : Union[str, Any] ): """simple docstring""" logger.info("Converting model..." ) # load original state dict _snake_case : Tuple = torch.hub.load_state_dict_from_url(a , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(a , a , a ) _snake_case : Union[str, Any] = rename_backbone_keys(a ) # query, key and value matrices need special treatment read_in_q_k_v(a ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case : int = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): _snake_case : Optional[int] = state_dict.pop(a ) _snake_case : Any = val # create HuggingFace model and load state dict _snake_case : Tuple = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case : Any = 15 _snake_case : int = 2 _snake_case : Optional[Any] = {0: "table", 1: "table rotated"} _snake_case : Union[str, Any] = idalabel _snake_case : Union[str, Any] = {v: k for k, v in idalabel.items()} else: _snake_case : Any = 125 _snake_case : Union[str, Any] = 6 _snake_case : List[str] = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } _snake_case : Any = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} _snake_case : Union[str, Any] = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1_000 ) _snake_case : str = TableTransformerForObjectDetection(a ) model.load_state_dict(a ) model.eval() # verify our conversion _snake_case : Optional[int] = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" _snake_case : Optional[Any] = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=a ) _snake_case : Dict = Image.open(a ).convert("RGB" ) _snake_case : Union[str, Any] = normalize(resize(a , a ) ).unsqueeze(0 ) _snake_case : str = model(a ) if "detection" in checkpoint_url: _snake_case : int = (1, 15, 3) _snake_case : List[str] = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) _snake_case : List[str] = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: _snake_case : Union[str, Any] = (1, 125, 7) _snake_case : str = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) _snake_case : Optional[Any] = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , a , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , a , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) _snake_case : int = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(a ) image_processor.push_to_hub(a ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _a : Any = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
87
0
"""simple docstring""" import argparse import json from pathlib import Path import requests import timm import torch from huggingface_hub import hf_hub_download from PIL import Image from timm.data import resolve_data_config from timm.data.transforms_factory import create_transform from transformers import ( BitConfig, ViTHybridConfig, ViTHybridForImageClassification, ViTHybridImageProcessor, ViTHybridModel, ) from transformers.image_utils import PILImageResampling from transformers.utils import logging logging.set_verbosity_info() _a : Any = logging.get_logger(__name__) def a__ ( a : Optional[Any] , a : int=False ): """simple docstring""" _snake_case : int = [] # fmt: off # stem: rename_keys.append(("cls_token", "vit.embeddings.cls_token") ) rename_keys.append(("pos_embed", "vit.embeddings.position_embeddings") ) rename_keys.append(("patch_embed.proj.weight", "vit.embeddings.patch_embeddings.projection.weight") ) rename_keys.append(("patch_embed.proj.bias", "vit.embeddings.patch_embeddings.projection.bias") ) # backbone rename_keys.append(("patch_embed.backbone.stem.conv.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.convolution.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.weight", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.weight") ) rename_keys.append(("patch_embed.backbone.stem.norm.bias", "vit.embeddings.patch_embeddings.backbone.bit.embedder.norm.bias") ) for stage_idx in range(len(config.backbone_config.depths ) ): for layer_idx in range(config.backbone_config.depths[stage_idx] ): rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv1.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv1.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm1.bias', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm1.bias') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv2.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv2.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm2.bias', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm2.bias') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.conv3.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.conv3.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.{layer_idx}.norm3.bias', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.{layer_idx}.norm3.bias') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.conv.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.conv.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.weight', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.weight') ) rename_keys.append((f'patch_embed.backbone.stages.{stage_idx}.blocks.0.downsample.norm.bias', f'vit.embeddings.patch_embeddings.backbone.bit.encoder.stages.{stage_idx}.layers.0.downsample.norm.bias') ) # transformer encoder for i in range(config.num_hidden_layers ): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append((f'blocks.{i}.norm1.weight', f'vit.encoder.layer.{i}.layernorm_before.weight') ) rename_keys.append((f'blocks.{i}.norm1.bias', f'vit.encoder.layer.{i}.layernorm_before.bias') ) rename_keys.append((f'blocks.{i}.attn.proj.weight', f'vit.encoder.layer.{i}.attention.output.dense.weight') ) rename_keys.append((f'blocks.{i}.attn.proj.bias', f'vit.encoder.layer.{i}.attention.output.dense.bias') ) rename_keys.append((f'blocks.{i}.norm2.weight', f'vit.encoder.layer.{i}.layernorm_after.weight') ) rename_keys.append((f'blocks.{i}.norm2.bias', f'vit.encoder.layer.{i}.layernorm_after.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc1.weight', f'vit.encoder.layer.{i}.intermediate.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc1.bias', f'vit.encoder.layer.{i}.intermediate.dense.bias') ) rename_keys.append((f'blocks.{i}.mlp.fc2.weight', f'vit.encoder.layer.{i}.output.dense.weight') ) rename_keys.append((f'blocks.{i}.mlp.fc2.bias', f'vit.encoder.layer.{i}.output.dense.bias') ) if base_model: # layernorm + pooler rename_keys.extend( [ ("norm.weight", "layernorm.weight"), ("norm.bias", "layernorm.bias"), ("pre_logits.fc.weight", "pooler.dense.weight"), ("pre_logits.fc.bias", "pooler.dense.bias"), ] ) # if just the base model, we should remove "vit" from all keys that start with "vit" _snake_case : Any = [(pair[0], pair[1][4:]) if pair[1].startswith("vit" ) else pair for pair in rename_keys] else: # layernorm + classification head rename_keys.extend( [ ("norm.weight", "vit.layernorm.weight"), ("norm.bias", "vit.layernorm.bias"), ("head.weight", "classifier.weight"), ("head.bias", "classifier.bias"), ] ) # fmt: on return rename_keys def a__ ( a : Any , a : Tuple , a : int=False ): """simple docstring""" for i in range(config.num_hidden_layers ): if base_model: _snake_case : Optional[int] = "" else: _snake_case : Tuple = "vit." # read in weights + bias of input projection layer (in timm, this is a single matrix + bias) _snake_case : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.weight' ) _snake_case : Tuple = state_dict.pop(f'blocks.{i}.attn.qkv.bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[ : config.hidden_size, : ] _snake_case : int = in_proj_bias[: config.hidden_size] _snake_case : int = in_proj_weight[ config.hidden_size : config.hidden_size * 2, : ] _snake_case : Optional[Any] = in_proj_bias[ config.hidden_size : config.hidden_size * 2 ] _snake_case : Union[str, Any] = in_proj_weight[ -config.hidden_size :, : ] _snake_case : Optional[int] = in_proj_bias[-config.hidden_size :] def a__ ( a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = ["head.weight", "head.bias"] for k in ignore_keys: state_dict.pop(a , a ) def a__ ( a : List[Any] , a : Tuple , a : int ): """simple docstring""" _snake_case : int = dct.pop(a ) _snake_case : Optional[int] = val def a__ ( ): """simple docstring""" _snake_case : Union[str, Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" _snake_case : List[Any] = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def a__ ( a : Any , a : Any , a : Optional[int]=False ): """simple docstring""" _snake_case : Optional[int] = BitConfig( global_padding="same" , layer_type="bottleneck" , depths=(3, 4, 9) , out_features=["stage3"] , embedding_dynamic_padding=a , ) _snake_case : int = ViTHybridConfig(backbone_config=a , image_size=384 , num_labels=1_000 ) _snake_case : int = False # load original model from timm _snake_case : Optional[int] = timm.create_model(a , pretrained=a ) timm_model.eval() # load state_dict of original model, remove and rename some keys _snake_case : Any = timm_model.state_dict() if base_model: remove_classification_head_(a ) _snake_case : List[Any] = create_rename_keys(a , a ) for src, dest in rename_keys: rename_key(a , a , a ) read_in_q_k_v(a , a , a ) _snake_case : str = "huggingface/label-files" _snake_case : str = "imagenet-1k-id2label.json" _snake_case : str = json.load(open(hf_hub_download(a , a , repo_type="dataset" ) , "r" ) ) _snake_case : Optional[int] = {int(a ): v for k, v in idalabel.items()} _snake_case : Dict = idalabel _snake_case : int = {v: k for k, v in idalabel.items()} # load HuggingFace model if vit_name[-5:] == "in21k": _snake_case : int = ViTHybridModel(a ).eval() else: _snake_case : Any = ViTHybridForImageClassification(a ).eval() model.load_state_dict(a ) # create image processor _snake_case : Dict = create_transform(**resolve_data_config({} , model=a ) ) _snake_case : List[Any] = transform.transforms _snake_case : List[str] = { "bilinear": PILImageResampling.BILINEAR, "bicubic": PILImageResampling.BICUBIC, "nearest": PILImageResampling.NEAREST, } _snake_case : Tuple = ViTHybridImageProcessor( do_resize=a , size={"shortest_edge": timm_transforms[0].size} , resample=pillow_resamplings[timm_transforms[0].interpolation.value] , do_center_crop=a , crop_size={"height": timm_transforms[1].size[0], "width": timm_transforms[1].size[1]} , do_normalize=a , image_mean=timm_transforms[-1].mean.tolist() , image_std=timm_transforms[-1].std.tolist() , ) _snake_case : Any = prepare_img() _snake_case : Tuple = transform(a ).unsqueeze(0 ) _snake_case : List[Any] = processor(a , return_tensors="pt" ).pixel_values # verify pixel values assert torch.allclose(a , a ) # verify logits with torch.no_grad(): _snake_case : Optional[int] = model(a ) _snake_case : Any = outputs.logits print("Predicted class:" , logits.argmax(-1 ).item() ) if base_model: _snake_case : List[Any] = timm_model.forward_features(a ) assert timm_pooled_output.shape == outputs.pooler_output.shape assert torch.allclose(a , outputs.pooler_output , atol=1e-3 ) else: _snake_case : Optional[Any] = timm_model(a ) assert timm_logits.shape == outputs.logits.shape assert torch.allclose(a , outputs.logits , atol=1e-3 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: Path(a ).mkdir(exist_ok=a ) print(f'Saving model {vit_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a ) print(f'Saving processor to {pytorch_dump_folder_path}' ) processor.save_pretrained(a ) if push_to_hub: print(f'Pushing model and processor to the hub {vit_name}' ) model.push_to_hub(f'ybelkada/{vit_name}' ) processor.push_to_hub(f'ybelkada/{vit_name}' ) if __name__ == "__main__": _a : str = argparse.ArgumentParser() # Required parameters parser.add_argument( """--vit_name""", default="""vit_base_r50_s16_384""", type=str, help="""Name of the hybrid ViT timm model you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the output PyTorch model directory.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether to upload the model to the HuggingFace hub.""" ) _a : Tuple = parser.parse_args() convert_vit_checkpoint(args.vit_name, args.pytorch_dump_folder_path, args.push_to_hub)
711
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
87
0
import argparse import collections import json from pathlib import Path import requests import torch import yaml from huggingface_hub import hf_hub_download from PIL import Image from transformers import ( MobileViTImageProcessor, MobileViTVaConfig, MobileViTVaForImageClassification, MobileViTVaForSemanticSegmentation, ) from transformers.utils import logging logging.set_verbosity_info() _a : List[Any] = logging.get_logger(__name__) def a__ ( a : Optional[Any] ): """simple docstring""" print("Loading config file..." ) def flatten_yaml_as_dict(a : Tuple , a : Union[str, Any]="" , a : Dict="." ): _snake_case : Tuple = [] for k, v in d.items(): _snake_case : List[Any] = parent_key + sep + k if parent_key else k if isinstance(a , collections.abc.MutableMapping ): items.extend(flatten_yaml_as_dict(a , a , sep=a ).items() ) else: items.append((new_key, v) ) return dict(a ) _snake_case : Optional[int] = argparse.Namespace() with open(a , "r" ) as yaml_file: try: _snake_case : Any = yaml.load(a , Loader=yaml.FullLoader ) _snake_case : List[str] = flatten_yaml_as_dict(a ) for k, v in flat_cfg.items(): setattr(a , a , a ) except yaml.YAMLError as exc: logger.error("Error while loading config file: {}. Error message: {}".format(a , str(a ) ) ) return config def a__ ( a : Any , a : Dict ): """simple docstring""" _snake_case : Union[str, Any] = MobileViTVaConfig() _snake_case : Union[str, Any] = False # dataset if task_name.startswith("imagenet1k_" ): _snake_case : Union[str, Any] = 1_000 if int(task_name.strip().split("_" )[-1] ) == 384: _snake_case : Dict = 384 else: _snake_case : Any = 256 _snake_case : List[str] = "imagenet-1k-id2label.json" elif task_name.startswith("imagenet21k_to_1k_" ): _snake_case : Any = 21_000 if int(task_name.strip().split("_" )[-1] ) == 384: _snake_case : Tuple = 384 else: _snake_case : str = 256 _snake_case : str = "imagenet-22k-id2label.json" elif task_name.startswith("ade20k_" ): _snake_case : Optional[Any] = 151 _snake_case : List[str] = 512 _snake_case : List[Any] = "ade20k-id2label.json" _snake_case : Optional[Any] = True elif task_name.startswith("voc_" ): _snake_case : List[str] = 21 _snake_case : Dict = 512 _snake_case : str = "pascal-voc-id2label.json" _snake_case : List[Any] = True # orig_config _snake_case : List[str] = load_orig_config_file(a ) assert getattr(a , "model.classification.name" , -1 ) == "mobilevit_v2", "Invalid model" _snake_case : int = getattr(a , "model.classification.mitv2.width_multiplier" , 1.0 ) assert ( getattr(a , "model.classification.mitv2.attn_norm_layer" , -1 ) == "layer_norm_2d" ), "Norm layers other than layer_norm_2d is not supported" _snake_case : Any = getattr(a , "model.classification.activation.name" , "swish" ) # config.image_size == getattr(orig_config, 'sampler.bs.crop_size_width', 256) if is_segmentation_model: _snake_case : Optional[Any] = getattr(a , "model.segmentation.output_stride" , 16 ) if "_deeplabv3" in task_name: _snake_case : Optional[Any] = getattr(a , "model.segmentation.deeplabv3.aspp_rates" , [12, 24, 36] ) _snake_case : Tuple = getattr(a , "model.segmentation.deeplabv3.aspp_out_channels" , 512 ) _snake_case : List[Any] = getattr(a , "model.segmentation.deeplabv3.aspp_dropout" , 0.1 ) # id2label _snake_case : Optional[int] = "huggingface/label-files" _snake_case : Tuple = json.load(open(hf_hub_download(a , a , repo_type="dataset" ) , "r" ) ) _snake_case : Any = {int(a ): v for k, v in idalabel.items()} _snake_case : Dict = idalabel _snake_case : str = {v: k for k, v in idalabel.items()} return config def a__ ( a : str , a : Dict , a : Union[str, Any] ): """simple docstring""" _snake_case : int = dct.pop(a ) _snake_case : Tuple = val def a__ ( a : Tuple , a : int=False ): """simple docstring""" if base_model: _snake_case : List[Any] = "" else: _snake_case : Optional[int] = "mobilevitv2." _snake_case : Tuple = [] for k in state_dict.keys(): if k[:8] == "encoder.": _snake_case : List[Any] = k[8:] else: _snake_case : Tuple = k if ".block." in k: _snake_case : Optional[int] = k_new.replace(".block." , "." ) if ".conv." in k: _snake_case : Dict = k_new.replace(".conv." , ".convolution." ) if ".norm." in k: _snake_case : Union[str, Any] = k_new.replace(".norm." , ".normalization." ) if "conv_1." in k: _snake_case : Union[str, Any] = k_new.replace("conv_1." , f'{model_prefix}conv_stem.' ) for i in [1, 2]: if f'layer_{i}.' in k: _snake_case : Any = k_new.replace(f'layer_{i}.' , f'{model_prefix}encoder.layer.{i-1}.layer.' ) if ".exp_1x1." in k: _snake_case : Optional[int] = k_new.replace(".exp_1x1." , ".expand_1x1." ) if ".red_1x1." in k: _snake_case : Tuple = k_new.replace(".red_1x1." , ".reduce_1x1." ) for i in [3, 4, 5]: if f'layer_{i}.0.' in k: _snake_case : List[Any] = k_new.replace(f'layer_{i}.0.' , f'{model_prefix}encoder.layer.{i-1}.downsampling_layer.' ) if f'layer_{i}.1.local_rep.0.' in k: _snake_case : Optional[int] = k_new.replace(f'layer_{i}.1.local_rep.0.' , f'{model_prefix}encoder.layer.{i-1}.conv_kxk.' ) if f'layer_{i}.1.local_rep.1.' in k: _snake_case : List[str] = k_new.replace(f'layer_{i}.1.local_rep.1.' , f'{model_prefix}encoder.layer.{i-1}.conv_1x1.' ) for i in [3, 4, 5]: if i == 3: _snake_case : Tuple = [0, 1] elif i == 4: _snake_case : str = [0, 1, 2, 3] elif i == 5: _snake_case : Optional[Any] = [0, 1, 2] for j in j_in: if f'layer_{i}.1.global_rep.{j}.' in k: _snake_case : Tuple = k_new.replace( f'layer_{i}.1.global_rep.{j}.' , f'{model_prefix}encoder.layer.{i-1}.transformer.layer.{j}.' ) if f'layer_{i}.1.global_rep.{j+1}.' in k: _snake_case : int = k_new.replace( f'layer_{i}.1.global_rep.{j+1}.' , f'{model_prefix}encoder.layer.{i-1}.layernorm.' ) if f'layer_{i}.1.conv_proj.' in k: _snake_case : List[str] = k_new.replace(f'layer_{i}.1.conv_proj.' , f'{model_prefix}encoder.layer.{i-1}.conv_projection.' ) if "pre_norm_attn.0." in k: _snake_case : Any = k_new.replace("pre_norm_attn.0." , "layernorm_before." ) if "pre_norm_attn.1." in k: _snake_case : Union[str, Any] = k_new.replace("pre_norm_attn.1." , "attention." ) if "pre_norm_ffn.0." in k: _snake_case : str = k_new.replace("pre_norm_ffn.0." , "layernorm_after." ) if "pre_norm_ffn.1." in k: _snake_case : str = k_new.replace("pre_norm_ffn.1." , "ffn.conv1." ) if "pre_norm_ffn.3." in k: _snake_case : Tuple = k_new.replace("pre_norm_ffn.3." , "ffn.conv2." ) if "classifier.1." in k: _snake_case : Union[str, Any] = k_new.replace("classifier.1." , "classifier." ) if "seg_head." in k: _snake_case : Union[str, Any] = k_new.replace("seg_head." , "segmentation_head." ) if ".aspp_layer." in k: _snake_case : Optional[Any] = k_new.replace(".aspp_layer." , "." ) if ".aspp_pool." in k: _snake_case : Dict = k_new.replace(".aspp_pool." , "." ) rename_keys.append((k, k_new) ) return rename_keys def a__ ( a : int ): """simple docstring""" _snake_case : Optional[int] = [] for k in state_dict.keys(): if k.startswith("seg_head.aux_head." ): keys_to_ignore.append(a ) for k in keys_to_ignore: state_dict.pop(a , a ) def a__ ( ): """simple docstring""" _snake_case : List[Any] = "http://images.cocodataset.org/val2017/000000039769.jpg" # url = "https://cdn.britannica.com/86/141086-050-9D7C75EE/Gulfstream-G450-business-jet-passengers.jpg" _snake_case : List[Any] = Image.open(requests.get(a , stream=a ).raw ) return im @torch.no_grad() def a__ ( a : Dict , a : Optional[int] , a : Any , a : List[str] ): """simple docstring""" _snake_case : Optional[int] = get_mobilevitva_config(a , a ) # load original state_dict _snake_case : List[str] = torch.load(a , map_location="cpu" ) # load huggingface model if task_name.startswith("ade20k_" ) or task_name.startswith("voc_" ): _snake_case : Any = MobileViTVaForSemanticSegmentation(a ).eval() _snake_case : List[str] = False else: _snake_case : Any = MobileViTVaForImageClassification(a ).eval() _snake_case : Dict = False # remove and rename some keys of load the original model _snake_case : int = checkpoint remove_unused_keys(a ) _snake_case : Dict = create_rename_keys(a , base_model=a ) for rename_key_src, rename_key_dest in rename_keys: rename_key(a , a , a ) # load modified state_dict model.load_state_dict(a ) # Check outputs on an image, prepared by MobileViTImageProcessor _snake_case : int = MobileViTImageProcessor(crop_size=config.image_size , size=config.image_size + 32 ) _snake_case : str = image_processor(images=prepare_img() , return_tensors="pt" ) _snake_case : Tuple = model(**a ) # verify classification model if task_name.startswith("imagenet" ): _snake_case : List[str] = outputs.logits _snake_case : Optional[int] = logits.argmax(-1 ).item() print("Predicted class:" , model.config.idalabel[predicted_class_idx] ) if task_name.startswith("imagenet1k_256" ) and config.width_multiplier == 1.0: # expected_logits for base variant _snake_case : List[Any] = torch.tensor([-1.6_3_3_6e0_0, -7.3_2_0_4e-0_2, -5.1_8_8_3e-0_1] ) assert torch.allclose(logits[0, :3] , a , atol=1e-4 ) Path(a ).mkdir(exist_ok=a ) print(f'Saving model {task_name} to {pytorch_dump_folder_path}' ) model.save_pretrained(a ) print(f'Saving image processor to {pytorch_dump_folder_path}' ) image_processor.save_pretrained(a ) if __name__ == "__main__": _a : List[str] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--task""", default="""imagenet1k_256""", type=str, help=( """Name of the task for which the MobileViTV2 model you'd like to convert is trained on . """ """ Classification (ImageNet-1k) - MobileViTV2 (256x256) : imagenet1k_256 - MobileViTV2 (Trained on 256x256 and Finetuned on 384x384) : imagenet1k_384 - MobileViTV2 (Trained on ImageNet-21k and Finetuned on ImageNet-1k 256x256) : imagenet21k_to_1k_256 - MobileViTV2 (Trained on ImageNet-21k, Finetuned on ImageNet-1k 256x256, and Finetuned on ImageNet-1k 384x384) : imagenet21k_to_1k_384 Segmentation - ADE20K Dataset : ade20k_deeplabv3 - Pascal VOC 2012 Dataset: voc_deeplabv3 """ ), choices=[ """imagenet1k_256""", """imagenet1k_384""", """imagenet21k_to_1k_256""", """imagenet21k_to_1k_384""", """ade20k_deeplabv3""", """voc_deeplabv3""", ], ) parser.add_argument( """--orig_checkpoint_path""", required=True, type=str, help="""Path to the original state dict (.pt file).""" ) parser.add_argument("""--orig_config_path""", required=True, type=str, help="""Path to the original config file.""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, type=str, help="""Path to the output PyTorch model directory.""" ) _a : Any = parser.parse_args() convert_mobilevitva_checkpoint( args.task, args.orig_checkpoint_path, args.orig_config_path, args.pytorch_dump_folder_path )
712
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _UpperCAmelCase ( _snake_case , _snake_case): __lowercase : List[Any] = """convnextv2""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=4 , snake_case_=None , snake_case_=None , snake_case_="gelu" , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=0.0 , snake_case_=2_24 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Tuple = num_channels _snake_case : Optional[int] = patch_size _snake_case : Tuple = num_stages _snake_case : int = [96, 1_92, 3_84, 7_68] if hidden_sizes is None else hidden_sizes _snake_case : str = [3, 3, 9, 3] if depths is None else depths _snake_case : int = hidden_act _snake_case : Tuple = initializer_range _snake_case : Union[str, Any] = layer_norm_eps _snake_case : Optional[int] = drop_path_rate _snake_case : Union[str, Any] = image_size _snake_case : List[Any] = ["stem"] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] _snake_case , _snake_case : Dict = get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names )
87
0
"""simple docstring""" import unittest from transformers import JukeboxTokenizer from transformers.testing_utils import require_torch class _UpperCAmelCase ( unittest.TestCase): __lowercase : int = JukeboxTokenizer __lowercase : Any = { """artist""": """Zac Brown Band""", """genres""": """Country""", """lyrics""": """I met a traveller from an antique land, Who said \"Two vast and trunkless legs of stone Stand in the desert. . . . Near them, on the sand, Half sunk a shattered visage lies, whose frown, And wrinkled lip, and sneer of cold command, Tell that its sculptor well those passions read Which yet survive, stamped on these lifeless things, The hand that mocked them, and the heart that fed; And on the pedestal, these words appear: My name is Ozymandias, King of Kings; Look on my Works, ye Mighty, and despair! Nothing beside remains. Round the decay Of that colossal Wreck, boundless and bare The lone and level sands stretch far away """, } @require_torch def lowerCamelCase__ ( self ): import torch _snake_case : List[str] = JukeboxTokenizer.from_pretrained("openai/jukebox-1b-lyrics" ) _snake_case : Dict = tokenizer(**self.metas )["input_ids"] # fmt: off _snake_case : Optional[int] = [ torch.tensor([[ 0, 0, 0, 71_69, 5_07, 9, 76, 39, 31, 46, 76, 27, 76, 46, 44, 27, 48, 31, 38, 38, 31, 44, 76, 32, 44, 41, 39, 76, 27, 40, 76, 27, 40, 46, 35, 43, 47, 31, 76, 38, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 41, 76, 45, 27, 35, 30, 76, 71, 20, 49, 41, 76, 48, 27, 45, 46, 76, 27, 40, 30, 76, 46, 44, 47, 40, 37, 38, 31, 45, 45, 76, 38, 31, 33, 45, 76, 41, 32, 76, 45, 46, 41, 40, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 19, 46, 27, 40, 30, 76, 35, 40, 76, 46, 34, 31, 76, 30, 31, 45, 31, 44, 46, 63, 76, 63, 76, 63, 76, 63, 76, 14, 31, 27, 44, 76, 46, 34, 31, 39, 64, 76, 41, 40, 76, 46, 34, 31, 76, 45, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 8, 27, 38, 32, 76, 45, 47, 40, 37, 76, 27, 76, 45, 34, 27, 46, 46, 31, 44, 31, 30, 76, 48, 35, 45, 27, 33, 31, 76, 38, 35, 31, 45, 64, 76, 49, 34, 41, 45, 31, 76, 32, 44, 41, 49, 40, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 49, 44, 35, 40, 37, 38, 31, 30, 76, 38, 35, 42, 64, 76, 27, 40, 30, 76, 45, 40, 31, 31, 44, 76, 41, 32, 76, 29, 41, 38, 30, 76, 29, 41, 39, 39, 27, 40, 30, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 31, 38, 38, 76, 46, 34, 27, 46, 76, 35, 46, 45, 76, 45, 29, 47, 38, 42, 46, 41, 44, 76, 49, 31, 38, 38, 76, 46, 34, 41, 45, 31, 76, 42, 27, 45, 45, 35, 41, 40, 45, 76, 44, 31, 27, 30, 78, 76, 76, 76, 76, 76, 76, 76, 76, 23, 34, 35, 29, 34, 76, 51, 31, 46, 76, 45, 47, 44, 48, 35, 48, 31, 64, 76, 45, 46, 27, 39, 42, 31, 30, 76, 41, 40, 76, 46, 34, 31, 45, 31, 76, 38, 35, 32, 31, 38, 31, 45, 45, 76, 46, 34, 35, 40, 33, 45, 64, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 34, 27, 40, 30, 76, 46, 34, 27, 46, 76, 39, 41, 29, 37, 31, 30, 76, 46, 34, 31, 39, 64, 76, 27, 40, 30, 76, 46, 34, 31, 76, 34, 31, 27, 44, 46, 76, 46, 34, 27, 46, 76, 32, 31, 30, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 1, 40, 30, 76, 41, 40, 76, 46, 34, 31, 76, 42, 31, 30, 31, 45, 46, 27, 38, 64, 76, 46, 34, 31, 45, 31, 76, 49, 41, 44, 30, 45, 76, 27, 42, 42, 31, 27, 44, 65, 78, 76, 76, 76, 76, 76, 76, 76, 76, 13, 51, 76, 40, 27, 39, 31, 76, 35, 45, 76, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 76, 11, 35, 40, 33, 76, 41, 32, 76, 11, 35, 40, 33, 45, 66, 78, 76, 76, 76, 76, 76, 76, 76, 76, 12, 41, 41, 37, 76, 41, 40, 76, 39, 51, 76, 23, 41, 44, 37, 45, 64, 76, 51, 31, 76, 13, 35, 33, 34, 46, 51, 64, 76, 27, 40, 30, 76, 30, 31, 45, 42, 27, 35, 44, 67, 78, 76, 76, 76, 76, 76, 76, 76, 76, 14, 41, 46, 34, 35, 40, 33, 76, 28, 31, 45, 35, 30, 31, 76, 44, 31, 39, 27, 35, 40, 45, 63, 76, 18, 41, 47, 40, 30, 76, 46, 34, 31, 76, 30, 31, 29, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76, 15, 32, 76, 46, 34, 27, 46, 76, 29, 41, 38, 41, 45, 45, 27, 38, 76, 23, 44, 31, 29, 37, 64, 76, 28, 41, 47, 40, 30, 38, 31, 45, 45, 76, 27, 40, 30, 76, 28, 27, 44, 31, 78, 76, 76, 76, 76, 76, 76, 76, 76, 20, 34, 31, 76, 38, 41, 40, 31, 76, 27, 40, 30, 76, 38, 31, 48, 31, 38, 76, 45, 27, 40, 30, 45, 76, 45, 46, 44, 31, 46, 29, 34, 76, 32, 27, 44, 76, 27, 49, 27, 51, 78, 76, 76, 76, 76, 76, 76, 76, 76]] ), torch.tensor([[0, 0, 0, 10_69, 11]] ), torch.tensor([[0, 0, 0, 10_69, 11]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) ) @require_torch def lowerCamelCase__ ( self ): import torch _snake_case : List[str] = JukeboxTokenizer.from_pretrained("openai/jukebox-5b-lyrics" ) _snake_case : Dict = tokenizer(**self.metas )["input_ids"] # fmt: off _snake_case : Tuple = [ torch.tensor([[ 0, 0, 0, 10_69, 11, -1, -1, -1, -1, 9, 77, 39, 31, 46, 77, 27, 77, 46, 44, 27, 48, 31, 38, 38, 31, 44, 77, 32, 44, 41, 39, 77, 27, 40, 77, 27, 40, 46, 35, 43, 47, 31, 77, 38, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 41, 77, 45, 27, 35, 30, 77, 72, 20, 49, 41, 77, 48, 27, 45, 46, 77, 27, 40, 30, 77, 46, 44, 47, 40, 37, 38, 31, 45, 45, 77, 38, 31, 33, 45, 77, 41, 32, 77, 45, 46, 41, 40, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 19, 46, 27, 40, 30, 77, 35, 40, 77, 46, 34, 31, 77, 30, 31, 45, 31, 44, 46, 63, 77, 63, 77, 63, 77, 63, 77, 14, 31, 27, 44, 77, 46, 34, 31, 39, 64, 77, 41, 40, 77, 46, 34, 31, 77, 45, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 8, 27, 38, 32, 77, 45, 47, 40, 37, 77, 27, 77, 45, 34, 27, 46, 46, 31, 44, 31, 30, 77, 48, 35, 45, 27, 33, 31, 77, 38, 35, 31, 45, 64, 77, 49, 34, 41, 45, 31, 77, 32, 44, 41, 49, 40, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 49, 44, 35, 40, 37, 38, 31, 30, 77, 38, 35, 42, 64, 77, 27, 40, 30, 77, 45, 40, 31, 31, 44, 77, 41, 32, 77, 29, 41, 38, 30, 77, 29, 41, 39, 39, 27, 40, 30, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 31, 38, 38, 77, 46, 34, 27, 46, 77, 35, 46, 45, 77, 45, 29, 47, 38, 42, 46, 41, 44, 77, 49, 31, 38, 38, 77, 46, 34, 41, 45, 31, 77, 42, 27, 45, 45, 35, 41, 40, 45, 77, 44, 31, 27, 30, 79, 77, 77, 77, 77, 77, 77, 77, 77, 23, 34, 35, 29, 34, 77, 51, 31, 46, 77, 45, 47, 44, 48, 35, 48, 31, 64, 77, 45, 46, 27, 39, 42, 31, 30, 77, 41, 40, 77, 46, 34, 31, 45, 31, 77, 38, 35, 32, 31, 38, 31, 45, 45, 77, 46, 34, 35, 40, 33, 45, 64, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 34, 27, 40, 30, 77, 46, 34, 27, 46, 77, 39, 41, 29, 37, 31, 30, 77, 46, 34, 31, 39, 64, 77, 27, 40, 30, 77, 46, 34, 31, 77, 34, 31, 27, 44, 46, 77, 46, 34, 27, 46, 77, 32, 31, 30, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 1, 40, 30, 77, 41, 40, 77, 46, 34, 31, 77, 42, 31, 30, 31, 45, 46, 27, 38, 64, 77, 46, 34, 31, 45, 31, 77, 49, 41, 44, 30, 45, 77, 27, 42, 42, 31, 27, 44, 65, 79, 77, 77, 77, 77, 77, 77, 77, 77, 13, 51, 77, 40, 27, 39, 31, 77, 35, 45, 77, 15, 52, 51, 39, 27, 40, 30, 35, 27, 45, 64, 77, 11, 35, 40, 33, 77, 41, 32, 77, 11, 35, 40, 33, 45, 66, 79, 77, 77, 77, 77, 77, 77, 77, 77, 12, 41, 41, 37, 77, 41, 40, 77, 39, 51, 77, 23, 41, 44, 37, 45, 64, 77, 51, 31, 77, 13, 35, 33, 34, 46, 51, 64, 77, 27, 40, 30, 77, 30, 31, 45, 42, 27, 35, 44, 67, 79, 77, 77, 77, 77, 77, 77, 77, 77, 14, 41, 46, 34, 35, 40, 33, 77, 28, 31, 45, 35, 30, 31, 77, 44, 31, 39, 27, 35, 40, 45, 63, 77, 18, 41, 47, 40, 30, 77, 46, 34, 31, 77, 30, 31, 29, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77, 15, 32, 77, 46, 34, 27, 46, 77, 29, 41, 38, 41, 45, 45, 27, 38, 77, 23, 44, 31, 29, 37, 64, 77, 28, 41, 47, 40, 30, 38, 31, 45, 45, 77, 27, 40, 30, 77, 28, 27, 44, 31, 79, 77, 77, 77, 77, 77, 77, 77, 77, 20, 34, 31, 77, 38, 41, 40, 31, 77, 27, 40, 30, 77, 38, 31, 48, 31, 38, 77, 45, 27, 40, 30, 45, 77, 45, 46, 44, 31, 46, 29, 34, 77, 32, 27, 44, 77, 27, 49, 27, 51, 79, 77, 77, 77, 77, 77, 77, 77, 77]] ), torch.tensor([[0, 0, 0, 10_69, 11, -1, -1, -1, -1]] ), torch.tensor([[0, 0, 0, 10_69, 11, -1, -1, -1, -1]] ), ] # fmt: on self.assertTrue(torch.allclose(tokens[0] , EXPECTED_OUTPUT[0] ) ) self.assertTrue(torch.allclose(tokens[1] , EXPECTED_OUTPUT[1] ) ) self.assertTrue(torch.allclose(tokens[2] , EXPECTED_OUTPUT[2] ) )
713
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def a__ ( a : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) _a : int = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : Dict = parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=snake_case_ , required=snake_case_ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=snake_case_ , required=snake_case_ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=snake_case_ , required=snake_case_ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=snake_case_ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=snake_case_ , default=snake_case_ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : str = logging.get_logger("transformers-cli/converting" ) self._logger.info(F'Loading model {model_type}' ) _snake_case : Optional[int] = model_type _snake_case : Any = tf_checkpoint _snake_case : Optional[int] = pytorch_dump_output _snake_case : Tuple = config _snake_case : Tuple = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) if "ckpt" in self._tf_checkpoint.lower(): _snake_case : int = self._tf_checkpoint _snake_case : Optional[Any] = "" else: _snake_case : Optional[int] = self._tf_checkpoint _snake_case : List[str] = "" convert_transfo_xl_checkpoint_to_pytorch( snake_case_ , self._config , self._pytorch_dump_output , snake_case_ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
87
0
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tensorflow_text_available, is_tf_available, is_tokenizers_available, is_torch_available, ) _a : List[str] = { """configuration_bert""": ["""BERT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """BertConfig""", """BertOnnxConfig"""], """tokenization_bert""": ["""BasicTokenizer""", """BertTokenizer""", """WordpieceTokenizer"""], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ["""BertTokenizerFast"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : str = [ """BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """BertForMaskedLM""", """BertForMultipleChoice""", """BertForNextSentencePrediction""", """BertForPreTraining""", """BertForQuestionAnswering""", """BertForSequenceClassification""", """BertForTokenClassification""", """BertLayer""", """BertLMHeadModel""", """BertModel""", """BertPreTrainedModel""", """load_tf_weights_in_bert""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Tuple = [ """TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST""", """TFBertEmbeddings""", """TFBertForMaskedLM""", """TFBertForMultipleChoice""", """TFBertForNextSentencePrediction""", """TFBertForPreTraining""", """TFBertForQuestionAnswering""", """TFBertForSequenceClassification""", """TFBertForTokenClassification""", """TFBertLMHeadModel""", """TFBertMainLayer""", """TFBertModel""", """TFBertPreTrainedModel""", ] try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = ["""TFBertTokenizer"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Any = [ """FlaxBertForCausalLM""", """FlaxBertForMaskedLM""", """FlaxBertForMultipleChoice""", """FlaxBertForNextSentencePrediction""", """FlaxBertForPreTraining""", """FlaxBertForQuestionAnswering""", """FlaxBertForSequenceClassification""", """FlaxBertForTokenClassification""", """FlaxBertModel""", """FlaxBertPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_bert import BERT_PRETRAINED_CONFIG_ARCHIVE_MAP, BertConfig, BertOnnxConfig from .tokenization_bert import BasicTokenizer, BertTokenizer, WordpieceTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_fast import BertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_bert import ( BERT_PRETRAINED_MODEL_ARCHIVE_LIST, BertForMaskedLM, BertForMultipleChoice, BertForNextSentencePrediction, BertForPreTraining, BertForQuestionAnswering, BertForSequenceClassification, BertForTokenClassification, BertLayer, BertLMHeadModel, BertModel, BertPreTrainedModel, load_tf_weights_in_bert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_bert import ( TF_BERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFBertEmbeddings, TFBertForMaskedLM, TFBertForMultipleChoice, TFBertForNextSentencePrediction, TFBertForPreTraining, TFBertForQuestionAnswering, TFBertForSequenceClassification, TFBertForTokenClassification, TFBertLMHeadModel, TFBertMainLayer, TFBertModel, TFBertPreTrainedModel, ) try: if not is_tensorflow_text_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bert_tf import TFBertTokenizer try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_bert import ( FlaxBertForCausalLM, FlaxBertForMaskedLM, FlaxBertForMultipleChoice, FlaxBertForNextSentencePrediction, FlaxBertForPreTraining, FlaxBertForQuestionAnswering, FlaxBertForSequenceClassification, FlaxBertForTokenClassification, FlaxBertModel, FlaxBertPreTrainedModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
714
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def a__ ( a : List[str] , a : Any ): """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _snake_case : Any = flax_key_tuple[:-1] + ("weight",) _snake_case : str = torch.permute(a , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(a ): # linear layer _snake_case : Optional[int] = flax_key_tuple[:-1] + ("weight",) _snake_case : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _snake_case : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) return flax_key_tuple, flax_tensor def a__ ( a : List[Any] , a : Union[str, Any] , a : List[str] ): """simple docstring""" if "metadata" in layer: _snake_case : Optional[int] = layer.split("metadata" ) _snake_case : Optional[int] = "".join(split_layer[0] )[:-1] _snake_case : int = [tuple(("metadata" + split_layer[1]).split("/" ) )] elif "kvstore" in layer: _snake_case : Any = layer.split("kvstore" ) _snake_case : str = "".join(split_layer[0] )[:-1] _snake_case : Any = [tuple(("kvstore" + split_layer[1]).split("/" ) )] else: _snake_case : List[Any] = layer.split("/" ) _snake_case : Tuple = "/".join(split_layer[:-1] ) _snake_case : int = (split_layer[-1],) if "kvstore/path" in layer: _snake_case : Optional[Any] = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _snake_case : Tuple = "file" else: _snake_case : Optional[int] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def a__ ( a : List[Any] , a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = rename_keys(a ) _snake_case : int = {} for k, v in current_block.items(): _snake_case : Optional[int] = v _snake_case : Optional[int] = new_current_block torch.save(a , a ) def a__ ( a : Dict , a : Tuple , a : List[str] , a : int , a : str = WEIGHTS_NAME ): """simple docstring""" _snake_case : Any = convert_file_size_to_int(a ) _snake_case : Tuple = [] _snake_case : Optional[int] = {} _snake_case : Tuple = 0 _snake_case : Optional[Any] = 0 os.makedirs(a , exist_ok=a ) with gfile.GFile(switch_checkpoint_path + "/checkpoint" , "rb" ) as fp: _snake_case : Any = serialization.msgpack_restore(fp.read() )["optimizer"]["target"] _snake_case : Optional[Any] = flatten_dict(a , sep="/" ) _snake_case : Optional[Any] = {} for layer in checkpoint_info.keys(): _snake_case , _snake_case , _snake_case : int = get_key_and_tensorstore_dict( a , a , a ) if curr_real_layer_name in all_layers: _snake_case : Dict = content else: _snake_case : Tuple = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _snake_case : List[str] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _snake_case : Dict = torch.tensor(a ) _snake_case : Dict = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _snake_case , _snake_case : Optional[int] = rename_base_flax_keys(tuple(key.split("/" ) ) , a ) _snake_case : Optional[Any] = "/".join(a ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _snake_case : Any = os.path.join( a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) del current_block _snake_case : List[Any] = {} _snake_case : str = 0 _snake_case : List[str] = raw_weights.to(getattr(a , a ) ) current_block_size += weight_size total_size += weight_size # Add the last block _snake_case : int = os.path.join(a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(a ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _snake_case : str = {} _snake_case : Any = {} for idx, shard in enumerate(a ): _snake_case : Optional[int] = weights_name.replace( ".bin" , f'-{idx+1:05d}-of-{len(a ):05d}.bin' ) # len(sharded_state_dicts):05d} _snake_case : Dict = os.path.join(a , weights_name.replace(".bin" , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(a , os.path.join(a , a ) ) _snake_case : Dict = shard for key in shard: _snake_case : int = shard_file # Add the metadata _snake_case : List[Any] = {"total_size": total_size} _snake_case : Any = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(a , a ) , "w" , encoding="utf-8" ) as f: _snake_case : Union[str, Any] = json.dumps(a , indent=2 , sort_keys=a ) + "\n" f.write(a ) return metadata, index if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) _a : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def a__ ( ): """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _snake_case : List[str] = SwitchTransformersConfig.from_pretrained("google/switch-base-8" ) config.save_pretrained("/home/arthur_huggingface_co/transformers/switch_converted" ) _snake_case : str = SwitchTransformersForConditionalGeneration.from_pretrained( "/home/arthur_huggingface_co/transformers/switch_converted" , device_map="auto" ) _snake_case : List[Any] = TaTokenizer.from_pretrained("t5-small" ) _snake_case : Optional[Any] = "A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>." _snake_case : Dict = tokenizer(a , return_tensors="pt" ).input_ids _snake_case : List[Any] = model.generate(a , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) if is_sentencepiece_available(): from ..ta.tokenization_ta import TaTokenizer else: from ...utils.dummy_sentencepiece_objects import TaTokenizer _a : List[str] = TaTokenizer if is_tokenizers_available(): from ..ta.tokenization_ta_fast import TaTokenizerFast else: from ...utils.dummy_tokenizers_objects import TaTokenizerFast _a : Optional[Any] = TaTokenizerFast _a : Union[str, Any] = {"""configuration_mt5""": ["""MT5Config""", """MT5OnnxConfig"""]} try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = [ """MT5EncoderModel""", """MT5ForConditionalGeneration""", """MT5ForQuestionAnswering""", """MT5Model""", """MT5PreTrainedModel""", """MT5Stack""", ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""TFMT5EncoderModel""", """TFMT5ForConditionalGeneration""", """TFMT5Model"""] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[str] = ["""FlaxMT5EncoderModel""", """FlaxMT5ForConditionalGeneration""", """FlaxMT5Model"""] if TYPE_CHECKING: from .configuration_mta import MTaConfig, MTaOnnxConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mta import ( MTaEncoderModel, MTaForConditionalGeneration, MTaForQuestionAnswering, MTaModel, MTaPreTrainedModel, MTaStack, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_mta import TFMTaEncoderModel, TFMTaForConditionalGeneration, TFMTaModel try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_mta import FlaxMTaEncoderModel, FlaxMTaForConditionalGeneration, FlaxMTaModel else: import sys _a : List[str] = _LazyModule( __name__, globals()["""__file__"""], _import_structure, extra_objects={"""MT5Tokenizer""": MTaTokenizer, """MT5TokenizerFast""": MTaTokenizerFast}, module_spec=__spec__, )
715
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
87
0
"""simple docstring""" import unittest from pathlib import Path from shutil import copyfile from transformers import SPIECE_UNDERLINE, is_sentencepiece_available from transformers.models.speech_to_text import SpeechaTextTokenizer from transformers.models.speech_to_text.tokenization_speech_to_text import VOCAB_FILES_NAMES, save_json from transformers.testing_utils import get_tests_dir, require_sentencepiece, require_tokenizers, slow from ...test_tokenization_common import TokenizerTesterMixin _a : str = get_tests_dir("""fixtures/test_sentencepiece.model""") if is_sentencepiece_available(): import sentencepiece as sp _a : int = 5 _a : Dict = 10 @require_sentencepiece @require_tokenizers class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : str = SpeechaTextTokenizer __lowercase : int = False __lowercase : List[str] = True def lowerCamelCase__ ( self ): super().setUp() _snake_case : Optional[Any] = sp.SentencePieceProcessor() spm_model.Load(snake_case_ ) _snake_case : Any = ["<s>", "<pad>", "</s>", "<unk>"] vocab += [spm_model.IdToPiece(id_ ) for id_ in range(len(snake_case_ ) )] _snake_case : Union[str, Any] = dict(zip(snake_case_ , range(len(snake_case_ ) ) ) ) _snake_case : Dict = Path(self.tmpdirname ) save_json(snake_case_ , save_dir / VOCAB_FILES_NAMES["vocab_file"] ) if not (save_dir / VOCAB_FILES_NAMES["spm_file"]).exists(): copyfile(snake_case_ , save_dir / VOCAB_FILES_NAMES["spm_file"] ) _snake_case : List[Any] = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) tokenizer.save_pretrained(self.tmpdirname ) def lowerCamelCase__ ( self ): _snake_case : Tuple = "<pad>" _snake_case : Optional[Any] = 1 self.assertEqual(self.get_tokenizer()._convert_token_to_id(snake_case_ ) , snake_case_ ) self.assertEqual(self.get_tokenizer()._convert_id_to_token(snake_case_ ) , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : int = list(self.get_tokenizer().get_vocab().keys() ) self.assertEqual(vocab_keys[0] , "<s>" ) self.assertEqual(vocab_keys[1] , "<pad>" ) self.assertEqual(vocab_keys[-1] , "j" ) self.assertEqual(len(snake_case_ ) , 10_01 ) def lowerCamelCase__ ( self ): self.assertEqual(self.get_tokenizer().vocab_size , 10_01 ) def lowerCamelCase__ ( self ): _snake_case : List[str] = SpeechaTextTokenizer.from_pretrained(self.tmpdirname ) _snake_case : Optional[Any] = tokenizer.tokenize("This is a test" ) self.assertListEqual(snake_case_ , ["▁This", "▁is", "▁a", "▁t", "est"] ) self.assertListEqual( tokenizer.convert_tokens_to_ids(snake_case_ ) , [2_89, 50, 14, 1_74, 3_86] , ) _snake_case : Optional[int] = tokenizer.tokenize("I was born in 92000, and this is falsé." ) self.assertListEqual( snake_case_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "9", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "é", "."] , ) _snake_case : Optional[Any] = tokenizer.convert_tokens_to_ids(snake_case_ ) self.assertListEqual(snake_case_ , [12, 25, 88, 59, 28, 23, 11, 4, 6_06, 3_51, 3_51, 3_51, 7, 16, 70, 50, 76, 84, 10, 4, 8] ) _snake_case : List[Any] = tokenizer.convert_ids_to_tokens(snake_case_ ) self.assertListEqual( snake_case_ , [SPIECE_UNDERLINE + "I", SPIECE_UNDERLINE + "was", SPIECE_UNDERLINE + "b", "or", "n", SPIECE_UNDERLINE + "in", SPIECE_UNDERLINE + "", "<unk>", "2", "0", "0", "0", ",", SPIECE_UNDERLINE + "and", SPIECE_UNDERLINE + "this", SPIECE_UNDERLINE + "is", SPIECE_UNDERLINE + "f", "al", "s", "<unk>", "."] , ) @slow def lowerCamelCase__ ( self ): # fmt: off _snake_case : Tuple = {"input_ids": [[37_91, 7_97, 31, 11, 64, 7_97, 31, 24_29, 4_33, 12, 11_76, 12, 20, 7_86, 9_15, 1_42, 24_13, 2_40, 37, 32_38, 7_97, 31, 11, 35, 93, 9_15, 1_42, 24_13, 2_40, 37, 55_40, 5_67, 12_76, 93, 37, 6_10, 40, 62, 4_55, 6_57, 10_42, 1_23, 7_80, 1_77, 37, 3_09, 2_41, 12_98, 5_14, 20, 2_92, 27_37, 1_14, 24_69, 2_41, 85, 64, 3_02, 5_48, 5_28, 4_23, 4, 5_09, 4_06, 4_23, 37, 6_01, 4, 7_77, 3_02, 5_48, 5_28, 4_23, 2_84, 4, 33_88, 5_11, 4_59, 4, 35_55, 40, 3_21, 3_02, 7_05, 4, 33_88, 5_11, 5_83, 3_26, 5, 5, 5, 62, 33_10, 5_60, 1_77, 26_80, 2_17, 15_08, 32, 31, 8_53, 4_18, 64, 5_83, 5_11, 16_05, 62, 35, 93, 5_60, 1_77, 26_80, 2_17, 15_08, 15_21, 64, 5_83, 5_11, 5_19, 62, 20, 15_15, 7_64, 20, 1_49, 2_61, 56_25, 79_72, 20, 55_40, 5_67, 12_76, 93, 39_25, 16_75, 11, 15, 8_02, 79_72, 5_76, 2_17, 15_08, 11, 35, 93, 12_53, 24_41, 15, 2_89, 6_52, 31, 4_16, 3_21, 38_42, 1_15, 40, 9_11, 8, 4_76, 6_19, 4, 3_80, 1_42, 4_23, 3_35, 2_40, 35, 93, 2_64, 8, 11, 3_35, 5_69, 4_20, 1_63, 5, 2], [2_60, 5_48, 5_28, 4_23, 20, 4_51, 20, 26_81, 11_53, 34_34, 20, 55_40, 37, 5_67, 1_26, 12_53, 24_41, 33_76, 4_49, 2_10, 4_31, 15_63, 1_77, 7_67, 55_40, 11, 12_03, 4_72, 11, 29_53, 6_85, 2_85, 3_64, 7_06, 11_53, 20, 67_99, 20, 28_69, 20, 44_64, 1_26, 40, 24_29, 20, 10_40, 8_66, 26_64, 4_18, 20, 3_18, 20, 17_26, 1_86, 20, 2_65, 5_22, 35, 93, 21_91, 46_34, 20, 10_40, 12, 67_99, 15, 2_28, 23_56, 1_42, 31, 11, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [25_75, 26_66, 6_84, 15_82, 11_76, 12, 6_27, 1_49, 6_19, 20, 49_02, 5_63, 11, 20, 1_49, 2_61, 34_20, 23_56, 1_74, 1_42, 47_14, 1_31, 5, 2, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]], "attention_mask": [[1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0], [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0]]} # noqa: E501 # fmt: on self.tokenizer_integration_test_util( expected_encoding=snake_case_ , model_name="facebook/s2t-small-mustc-en-de-st" , revision="a14f04cf0776c02f62a8cb800cf7909e15ea23ad" , ) @require_sentencepiece class _UpperCAmelCase ( unittest.TestCase): __lowercase : Tuple = """valhalla/s2t_mustc_multilinguial_medium""" __lowercase : Union[str, Any] = """C'est trop cool""" __lowercase : Any = """Esto es genial""" @classmethod def lowerCamelCase__ ( cls ): _snake_case : SpeechaTextTokenizer = SpeechaTextTokenizer.from_pretrained(cls.checkpoint_name ) return cls def lowerCamelCase__ ( self ): self.assertEqual(self.tokenizer.lang_code_to_id["pt"] , 4 ) self.assertEqual(self.tokenizer.lang_code_to_id["ru"] , 6 ) self.assertEqual(self.tokenizer.lang_code_to_id["it"] , 9 ) self.assertEqual(self.tokenizer.lang_code_to_id["de"] , 11 ) def lowerCamelCase__ ( self ): self.assertEqual(self.tokenizer.vocab_size , 1_00_00 ) def lowerCamelCase__ ( self ): self.assertIn(snake_case_ , self.tokenizer.all_special_ids ) _snake_case : Optional[int] = [ES_CODE, 4, 16_01, 47, 76_47, 2] _snake_case : Tuple = self.tokenizer.decode(snake_case_ , skip_special_tokens=snake_case_ ) _snake_case : int = self.tokenizer.decode(generated_ids[1:] , skip_special_tokens=snake_case_ ) self.assertEqual(snake_case_ , snake_case_ ) self.assertNotIn(self.tokenizer.eos_token , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = "fr" _snake_case : int = self.tokenizer(self.french_text ).input_ids self.assertEqual(encoded[0] , snake_case_ ) self.assertEqual(encoded[-1] , self.tokenizer.eos_token_id ) def lowerCamelCase__ ( self ): _snake_case : Tuple = "fr" self.assertListEqual(self.tokenizer.prefix_tokens , [FR_CODE] ) _snake_case : List[Any] = "es" self.assertListEqual(self.tokenizer.prefix_tokens , [ES_CODE] )
716
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" import argparse import os import shutil import torch from emmental.modules import MagnitudeBinarizer, ThresholdBinarizer, TopKBinarizer def a__ ( a : Any ): """simple docstring""" _snake_case : str = args.pruning_method _snake_case : Any = args.threshold _snake_case : Union[str, Any] = args.model_name_or_path.rstrip("/" ) _snake_case : Dict = args.target_model_path print(f'Load fine-pruned model from {model_name_or_path}' ) _snake_case : Dict = torch.load(os.path.join(a , "pytorch_model.bin" ) ) _snake_case : Union[str, Any] = {} for name, tensor in model.items(): if "embeddings" in name or "LayerNorm" in name or "pooler" in name: _snake_case : str = tensor print(f'Copied layer {name}' ) elif "classifier" in name or "qa_output" in name: _snake_case : List[Any] = tensor print(f'Copied layer {name}' ) elif "bias" in name: _snake_case : Tuple = tensor print(f'Copied layer {name}' ) else: if pruning_method == "magnitude": _snake_case : Any = MagnitudeBinarizer.apply(inputs=a , threshold=a ) _snake_case : Union[str, Any] = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "topK": if "mask_scores" in name: continue _snake_case : List[Any] = name[:-6] _snake_case : Any = model[f'{prefix_}mask_scores'] _snake_case : Tuple = TopKBinarizer.apply(a , a ) _snake_case : List[str] = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "sigmoied_threshold": if "mask_scores" in name: continue _snake_case : str = name[:-6] _snake_case : int = model[f'{prefix_}mask_scores'] _snake_case : Optional[int] = ThresholdBinarizer.apply(a , a , a ) _snake_case : List[str] = tensor * mask print(f'Pruned layer {name}' ) elif pruning_method == "l0": if "mask_scores" in name: continue _snake_case : Optional[int] = name[:-6] _snake_case : str = model[f'{prefix_}mask_scores'] _snake_case : Dict = -0.1, 1.1 _snake_case : Union[str, Any] = torch.sigmoid(a ) _snake_case : List[str] = s * (r - l) + l _snake_case : Tuple = s_bar.clamp(min=0.0 , max=1.0 ) _snake_case : str = tensor * mask print(f'Pruned layer {name}' ) else: raise ValueError("Unknown pruning method" ) if target_model_path is None: _snake_case : Optional[Any] = os.path.join( os.path.dirname(a ) , f'bertarized_{os.path.basename(a )}' ) if not os.path.isdir(a ): shutil.copytree(a , a ) print(f'\nCreated folder {target_model_path}' ) torch.save(a , os.path.join(a , "pytorch_model.bin" ) ) print("\nPruned model saved! See you later!" ) if __name__ == "__main__": _a : Optional[int] = argparse.ArgumentParser() parser.add_argument( """--pruning_method""", choices=["""l0""", """magnitude""", """topK""", """sigmoied_threshold"""], type=str, required=True, help=( """Pruning Method (l0 = L0 regularization, magnitude = Magnitude pruning, topK = Movement pruning,""" """ sigmoied_threshold = Soft movement pruning)""" ), ) parser.add_argument( """--threshold""", type=float, required=False, help=( """For `magnitude` and `topK`, it is the level of remaining weights (in %) in the fine-pruned model.""" """For `sigmoied_threshold`, it is the threshold \tau against which the (sigmoied) scores are compared.""" """Not needed for `l0`""" ), ) parser.add_argument( """--model_name_or_path""", type=str, required=True, help="""Folder containing the model that was previously fine-pruned""", ) parser.add_argument( """--target_model_path""", default=None, type=str, required=False, help="""Folder containing the model that was previously fine-pruned""", ) _a : Optional[int] = parser.parse_args() main(args)
717
"""simple docstring""" def a__ ( a : list , a : int , a : int = 0 , a : int = 0 ): """simple docstring""" _snake_case : Optional[int] = right or len(a ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(a , a , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import io import itertools import json from dataclasses import dataclass from typing import Optional import pyarrow as pa import pyarrow.json as paj import datasets from datasets.table import table_cast from datasets.utils.file_utils import readline _a : Tuple = datasets.utils.logging.get_logger(__name__) @dataclass class _UpperCAmelCase ( datasets.BuilderConfig): __lowercase : Optional[datasets.Features] = None __lowercase : str = "utf-8" __lowercase : Optional[str] = None __lowercase : Optional[str] = None __lowercase : bool = True # deprecated __lowercase : Optional[int] = None # deprecated __lowercase : int = 1_0 << 2_0 # 10MB __lowercase : Optional[bool] = None class _UpperCAmelCase ( datasets.ArrowBasedBuilder): __lowercase : int = JsonConfig def lowerCamelCase__ ( self ): if self.config.block_size is not None: logger.warning("The JSON loader parameter `block_size` is deprecated. Please use `chunksize` instead" ) _snake_case : List[str] = self.config.block_size if self.config.use_threads is not True: logger.warning( "The JSON loader parameter `use_threads` is deprecated and doesn't have any effect anymore." ) if self.config.newlines_in_values is not None: raise ValueError("The JSON loader parameter `newlines_in_values` is no longer supported" ) return datasets.DatasetInfo(features=self.config.features ) def lowerCamelCase__ ( self , snake_case_ ): if not self.config.data_files: raise ValueError(F'At least one data file must be specified, but got data_files={self.config.data_files}' ) _snake_case : Union[str, Any] = dl_manager.download_and_extract(self.config.data_files ) if isinstance(snake_case_ , (str, list, tuple) ): _snake_case : List[Any] = data_files if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = [files] _snake_case : str = [dl_manager.iter_files(snake_case_ ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] _snake_case : List[str] = [] for split_name, files in data_files.items(): if isinstance(snake_case_ , snake_case_ ): _snake_case : int = [files] _snake_case : Optional[int] = [dl_manager.iter_files(snake_case_ ) for file in files] splits.append(datasets.SplitGenerator(name=snake_case_ , gen_kwargs={"files": files} ) ) return splits def lowerCamelCase__ ( self , snake_case_ ): if self.config.features is not None: # adding missing columns for column_name in set(self.config.features ) - set(pa_table.column_names ): _snake_case : Optional[int] = self.config.features.arrow_schema.field(snake_case_ ).type _snake_case : Union[str, Any] = pa_table.append_column(snake_case_ , pa.array([None] * len(snake_case_ ) , type=snake_case_ ) ) # more expensive cast to support nested structures with keys in a different order # allows str <-> int/float or str to Audio for example _snake_case : Tuple = table_cast(snake_case_ , self.config.features.arrow_schema ) return pa_table def lowerCamelCase__ ( self , snake_case_ ): for file_idx, file in enumerate(itertools.chain.from_iterable(snake_case_ ) ): # If the file is one json object and if we need to look at the list of items in one specific field if self.config.field is not None: with open(snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: _snake_case : Union[str, Any] = json.load(snake_case_ ) # We keep only the field we are interested in _snake_case : List[str] = dataset[self.config.field] # We accept two format: a list of dicts or a dict of lists if isinstance(snake_case_ , (list, tuple) ): _snake_case : Any = set().union(*[row.keys() for row in dataset] ) _snake_case : str = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} else: _snake_case : List[str] = dataset _snake_case : Tuple = pa.Table.from_pydict(snake_case_ ) yield file_idx, self._cast_table(snake_case_ ) # If the file has one json object per line else: with open(snake_case_ , "rb" ) as f: _snake_case : Union[str, Any] = 0 # Use block_size equal to the chunk size divided by 32 to leverage multithreading # Set a default minimum value of 16kB if the chunk size is really small _snake_case : Optional[Any] = max(self.config.chunksize // 32 , 16 << 10 ) _snake_case : Tuple = ( self.config.encoding_errors if self.config.encoding_errors is not None else "strict" ) while True: _snake_case : int = f.read(self.config.chunksize ) if not batch: break # Finish current line try: batch += f.readline() except (AttributeError, io.UnsupportedOperation): batch += readline(snake_case_ ) # PyArrow only accepts utf-8 encoded bytes if self.config.encoding != "utf-8": _snake_case : Optional[int] = batch.decode(self.config.encoding , errors=snake_case_ ).encode("utf-8" ) try: while True: try: _snake_case : List[Any] = paj.read_json( io.BytesIO(snake_case_ ) , read_options=paj.ReadOptions(block_size=snake_case_ ) ) break except (pa.ArrowInvalid, pa.ArrowNotImplementedError) as e: if ( isinstance(snake_case_ , pa.ArrowInvalid ) and "straddling" not in str(snake_case_ ) or block_size > len(snake_case_ ) ): raise else: # Increase the block size in case it was too small. # The block size will be reset for the next file. logger.debug( F'Batch of {len(snake_case_ )} bytes couldn\'t be parsed with block_size={block_size}. Retrying with block_size={block_size * 2}.' ) block_size *= 2 except pa.ArrowInvalid as e: try: with open( snake_case_ , encoding=self.config.encoding , errors=self.config.encoding_errors ) as f: _snake_case : Optional[Any] = json.load(snake_case_ ) except json.JSONDecodeError: logger.error(F'Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}' ) raise e # If possible, parse the file as a list of json objects and exit the loop if isinstance(snake_case_ , snake_case_ ): # list is the only sequence type supported in JSON try: _snake_case : Union[str, Any] = set().union(*[row.keys() for row in dataset] ) _snake_case : Union[str, Any] = {col: [row.get(snake_case_ ) for row in dataset] for col in keys} _snake_case : Union[str, Any] = pa.Table.from_pydict(snake_case_ ) except (pa.ArrowInvalid, AttributeError) as e: logger.error(F'Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}' ) raise ValueError(F'Not able to read records in the JSON file at {file}.' ) from None yield file_idx, self._cast_table(snake_case_ ) break else: logger.error(F'Failed to read file \'{file}\' with error {type(snake_case_ )}: {e}' ) raise ValueError( F'Not able to read records in the JSON file at {file}. ' F'You should probably indicate the field of the JSON file containing your records. ' F'This JSON file contain the following fields: {str(list(dataset.keys() ) )}. ' F'Select the correct one and provide it as `field=\'XXX\'` to the dataset loading method. ' ) from None # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield (file_idx, batch_idx), self._cast_table(snake_case_ ) batch_idx += 1
718
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ ): _snake_case , _snake_case : Dict = text, pattern _snake_case , _snake_case : int = len(snake_case_ ), len(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCamelCase__ ( self ): # searches pattern in text and returns index positions _snake_case : List[str] = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case : Union[str, Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: _snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) _snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _a : List[Any] = """ABAABA""" _a : str = """AB""" _a : List[Any] = BoyerMooreSearch(text, pattern) _a : Any = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
87
0
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
719
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
87
0
"""simple docstring""" import json import os from typing import Optional import numpy as np from ...feature_extraction_utils import BatchFeature from ...processing_utils import ProcessorMixin from ...utils import logging from ...utils.hub import get_file_from_repo from ..auto import AutoTokenizer _a : Dict = logging.get_logger(__name__) class _UpperCAmelCase ( _snake_case): __lowercase : Tuple = """AutoTokenizer""" __lowercase : Dict = ["""tokenizer"""] __lowercase : Optional[Any] = { """semantic_prompt""": 1, """coarse_prompt""": 2, """fine_prompt""": 2, } def __init__( self , snake_case_ , snake_case_=None ): super().__init__(snake_case_ ) _snake_case : List[str] = speaker_embeddings @classmethod def lowerCamelCase__ ( cls , snake_case_ , snake_case_="speaker_embeddings_path.json" , **snake_case_ ): if speaker_embeddings_dict_path is not None: _snake_case : List[Any] = get_file_from_repo( snake_case_ , snake_case_ , subfolder=kwargs.pop("subfolder" , snake_case_ ) , cache_dir=kwargs.pop("cache_dir" , snake_case_ ) , force_download=kwargs.pop("force_download" , snake_case_ ) , proxies=kwargs.pop("proxies" , snake_case_ ) , resume_download=kwargs.pop("resume_download" , snake_case_ ) , local_files_only=kwargs.pop("local_files_only" , snake_case_ ) , use_auth_token=kwargs.pop("use_auth_token" , snake_case_ ) , revision=kwargs.pop("revision" , snake_case_ ) , ) if speaker_embeddings_path is None: logger.warning( F'`{os.path.join(snake_case_ , snake_case_ )}` does not exists\n , no preloaded speaker embeddings will be used - Make sure to provide a correct path to the json\n dictionnary if wanted, otherwise set `speaker_embeddings_dict_path=None`.' ) _snake_case : Dict = None else: with open(snake_case_ ) as speaker_embeddings_json: _snake_case : int = json.load(snake_case_ ) else: _snake_case : int = None _snake_case : Dict = AutoTokenizer.from_pretrained(snake_case_ , **snake_case_ ) return cls(tokenizer=snake_case_ , speaker_embeddings=snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_="speaker_embeddings_path.json" , snake_case_="speaker_embeddings" , snake_case_ = False , **snake_case_ , ): if self.speaker_embeddings is not None: os.makedirs(os.path.join(snake_case_ , snake_case_ , "v2" ) , exist_ok=snake_case_ ) _snake_case : int = {} _snake_case : int = save_directory for prompt_key in self.speaker_embeddings: if prompt_key != "repo_or_path": _snake_case : Union[str, Any] = self._load_voice_preset(snake_case_ ) _snake_case : List[str] = {} for key in self.speaker_embeddings[prompt_key]: np.save( os.path.join( embeddings_dict["repo_or_path"] , snake_case_ , F'{prompt_key}_{key}' ) , voice_preset[key] , allow_pickle=snake_case_ , ) _snake_case : Optional[int] = os.path.join(snake_case_ , F'{prompt_key}_{key}.npy' ) _snake_case : Union[str, Any] = tmp_dict with open(os.path.join(snake_case_ , snake_case_ ) , "w" ) as fp: json.dump(snake_case_ , snake_case_ ) super().save_pretrained(snake_case_ , snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ = None , **snake_case_ ): _snake_case : Dict = self.speaker_embeddings[voice_preset] _snake_case : Optional[Any] = {} for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset_paths: raise ValueError( F'Voice preset unrecognized, missing {key} as a key in self.speaker_embeddings[{voice_preset}].' ) _snake_case : str = get_file_from_repo( self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] , subfolder=kwargs.pop("subfolder" , snake_case_ ) , cache_dir=kwargs.pop("cache_dir" , snake_case_ ) , force_download=kwargs.pop("force_download" , snake_case_ ) , proxies=kwargs.pop("proxies" , snake_case_ ) , resume_download=kwargs.pop("resume_download" , snake_case_ ) , local_files_only=kwargs.pop("local_files_only" , snake_case_ ) , use_auth_token=kwargs.pop("use_auth_token" , snake_case_ ) , revision=kwargs.pop("revision" , snake_case_ ) , ) if path is None: raise ValueError( F'`{os.path.join(self.speaker_embeddings.get("repo_or_path" , "/" ) , voice_preset_paths[key] )}` does not exists\n , no preloaded voice preset will be used - Make sure to provide correct paths to the {voice_preset}\n embeddings.' ) _snake_case : str = np.load(snake_case_ ) return voice_preset_dict def lowerCamelCase__ ( self , snake_case_ = None ): for key in ["semantic_prompt", "coarse_prompt", "fine_prompt"]: if key not in voice_preset: raise ValueError(F'Voice preset unrecognized, missing {key} as a key.' ) if not isinstance(voice_preset[key] , np.ndarray ): raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) if len(voice_preset[key].shape ) != self.preset_shape[key]: raise ValueError(F'{key} voice preset must be a {str(self.preset_shape[key] )}D ndarray.' ) def __call__( self , snake_case_=None , snake_case_=None , snake_case_="pt" , snake_case_=2_56 , snake_case_=False , snake_case_=True , snake_case_=False , **snake_case_ , ): if voice_preset is not None and not isinstance(snake_case_ , snake_case_ ): if ( isinstance(snake_case_ , snake_case_ ) and self.speaker_embeddings is not None and voice_preset in self.speaker_embeddings ): _snake_case : Any = self._load_voice_preset(snake_case_ ) else: if isinstance(snake_case_ , snake_case_ ) and not voice_preset.endswith(".npz" ): _snake_case : Any = voice_preset + ".npz" _snake_case : List[Any] = np.load(snake_case_ ) if voice_preset is not None: self._validate_voice_preset_dict(snake_case_ , **snake_case_ ) _snake_case : List[Any] = BatchFeature(data=snake_case_ , tensor_type=snake_case_ ) _snake_case : Tuple = self.tokenizer( snake_case_ , return_tensors=snake_case_ , padding="max_length" , max_length=snake_case_ , return_attention_mask=snake_case_ , return_token_type_ids=snake_case_ , add_special_tokens=snake_case_ , **snake_case_ , ) if voice_preset is not None: _snake_case : Union[str, Any] = voice_preset return encoded_text
720
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" def a__ ( a : list[list] ): """simple docstring""" _snake_case : Any = current_set.copy() for row_index, row in enumerate(a ): _snake_case : List[Any] = row[0] for column_index, column in enumerate(a ): if magnitude == 0: _snake_case : Union[str, Any] = column continue _snake_case : List[str] = column / magnitude # Subtract to cancel term _snake_case : Optional[int] = current_set[0] _snake_case : Optional[int] = [first_row] _snake_case : List[str] = current_set[1::] for row in current_set: _snake_case : Tuple = [] # If first term is 0, it is already in form we want, so we preserve it if row[0] == 0: final_set.append(a ) continue for column_index in range(len(a ) ): temp_row.append(first_row[column_index] - row[column_index] ) final_set.append(a ) # Create next recursion iteration set if len(final_set[0] ) != 3: _snake_case : List[str] = final_set[0] _snake_case : Dict = [] _snake_case : Tuple = [] for row in final_set[1::]: current_first_column.append(row[0] ) next_iteration.append(row[1::] ) _snake_case : Optional[int] = simplify(a ) for i in range(len(a ) ): resultant[i].insert(0 , current_first_column[i] ) resultant.insert(0 , a ) _snake_case : Any = resultant return final_set def a__ ( a : list[list] ): """simple docstring""" if len(a ) == 0: raise IndexError("solve_simultaneous() requires n lists of length n+1" ) _snake_case : Optional[int] = len(a ) + 1 if any(len(a ) != _length for item in equations ): raise IndexError("solve_simultaneous() requires n lists of length n+1" ) for row in equations: if any(not isinstance(a , (int, float) ) for column in row ): raise ValueError("solve_simultaneous() requires lists of integers" ) if len(a ) == 1: return [equations[0][-1] / equations[0][0]] _snake_case : Dict = equations.copy() if any(0 in row for row in data_set ): _snake_case : Tuple = data_set.copy() _snake_case : Any = [] for row_index, row in enumerate(a ): if 0 not in row: _snake_case : Optional[Any] = data_set.pop(a ) break if not full_row: raise ValueError("solve_simultaneous() requires at least 1 full equation" ) data_set.insert(0 , a ) _snake_case : Optional[int] = data_set.copy() _snake_case : Union[str, Any] = simplify(a ) _snake_case : Optional[Any] = simplified[::-1] _snake_case : list = [] for row in simplified: _snake_case : str = row[-1] if not solutions: if row[-2] == 0: solutions.append(0 ) continue solutions.append(current_solution / row[-2] ) continue _snake_case : str = row.copy()[: len(a ) - 1 :] while temp_row[0] == 0: temp_row.pop(0 ) if len(a ) == 0: solutions.append(0 ) continue _snake_case : Any = temp_row[1::] _snake_case : Any = temp_row[::-1] for column_index, column in enumerate(a ): current_solution -= column * solutions[column_index] solutions.append(a ) _snake_case : Optional[Any] = [] for item in solutions: final.append(float(round(a , 5 ) ) ) return final[::-1] if __name__ == "__main__": import doctest doctest.testmod() _a : Optional[int] = [ [2, 1, 1, 1, 1, 4], [1, 2, 1, 1, 1, 5], [1, 1, 2, 1, 1, 6], [1, 1, 1, 2, 1, 7], [1, 1, 1, 1, 2, 8], ] print(solve_simultaneous(eq)) print(solve_simultaneous([[4, 2]]))
721
"""simple docstring""" import argparse import json import subprocess def a__ ( a : Optional[Any] , a : Optional[int] ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[Any] = ( f'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _snake_case : Dict = subprocess.run(a , shell=a , stdout=subprocess.PIPE ) _snake_case : Tuple = output.stdout.decode("utf-8" ) _snake_case : List[str] = json.loads(a ) _snake_case : Any = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(a ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(a ) ) if len(a ) > 0: _snake_case : Any = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def a__ ( a : Optional[int] ): """simple docstring""" return values.split("," ) _a : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _a : List[str] = parser.parse_args() get_runner_status(args.target_runners, args.token)
87
0
from scipy.stats import spearmanr import datasets _a : Optional[Any] = """ The Spearman rank-order correlation coefficient is a measure of the relationship between two datasets. Like other correlation coefficients, this one varies between -1 and +1 with 0 implying no correlation. Positive correlations imply that as data in dataset x increases, so does data in dataset y. Negative correlations imply that as x increases, y decreases. Correlations of -1 or +1 imply an exact monotonic relationship. Unlike the Pearson correlation, the Spearman correlation does not assume that both datasets are normally distributed. The p-value roughly indicates the probability of an uncorrelated system producing datasets that have a Spearman correlation at least as extreme as the one computed from these datasets. The p-values are not entirely reliable but are probably reasonable for datasets larger than 500 or so. """ _a : Optional[Any] = """ Args: predictions (`List[float]`): Predicted labels, as returned by a model. references (`List[float]`): Ground truth labels. return_pvalue (`bool`): If `True`, returns the p-value. If `False`, returns only the spearmanr score. Defaults to `False`. Returns: spearmanr (`float`): Spearman correlation coefficient. p-value (`float`): p-value. **Note**: is only returned if `return_pvalue=True` is input. Examples: Example 1: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], predictions=[10, 9, 2.5, 6, 4]) >>> print(results) {'spearmanr': -0.7} Example 2: >>> spearmanr_metric = datasets.load_metric(\"spearmanr\") >>> results = spearmanr_metric.compute(references=[1, 2, 3, 4, 5], ... predictions=[10, 9, 2.5, 6, 4], ... return_pvalue=True) >>> print(results['spearmanr']) -0.7 >>> print(round(results['spearmanr_pvalue'], 2)) 0.19 """ _a : Tuple = r"""\ @book{kokoska2000crc, title={CRC standard probability and statistics tables and formulae}, author={Kokoska, Stephen and Zwillinger, Daniel}, year={2000}, publisher={Crc Press} } @article{2020SciPy-NMeth, author = {Virtanen, Pauli and Gommers, Ralf and Oliphant, Travis E. and Haberland, Matt and Reddy, Tyler and Cournapeau, David and Burovski, Evgeni and Peterson, Pearu and Weckesser, Warren and Bright, Jonathan and {van der Walt}, St{\'e}fan J. and Brett, Matthew and Wilson, Joshua and Millman, K. Jarrod and Mayorov, Nikolay and Nelson, Andrew R. J. and Jones, Eric and Kern, Robert and Larson, Eric and Carey, C J and Polat, {\.I}lhan and Feng, Yu and Moore, Eric W. and {VanderPlas}, Jake and Laxalde, Denis and Perktold, Josef and Cimrman, Robert and Henriksen, Ian and Quintero, E. A. and Harris, Charles R. and Archibald, Anne M. and Ribeiro, Ant{\^o}nio H. and Pedregosa, Fabian and {van Mulbregt}, Paul and {SciPy 1.0 Contributors}}, title = {{{SciPy} 1.0: Fundamental Algorithms for Scientific Computing in Python}}, journal = {Nature Methods}, year = {2020}, volume = {17}, pages = {261--272}, adsurl = {https://rdcu.be/b08Wh}, doi = {10.1038/s41592-019-0686-2}, } """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _UpperCAmelCase ( datasets.Metric): def lowerCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Value("float" ), "references": datasets.Value("float" ), } ) , reference_urls=["https://docs.scipy.org/doc/scipy/reference/generated/scipy.stats.spearmanr.html"] , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Optional[Any] = spearmanr(snake_case_ , snake_case_ ) if return_pvalue: return {"spearmanr": results[0], "spearmanr_pvalue": results[1]} else: return {"spearmanr": results[0]}
700
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
87
0
import sys from collections.abc import Mapping from typing import TYPE_CHECKING, Dict, Optional import numpy as np import pyarrow as pa from .. import config from ..utils.logging import get_logger from ..utils.py_utils import map_nested from .formatting import TensorFormatter if TYPE_CHECKING: import jax import jaxlib _a : Optional[int] = get_logger() _a : Optional[dict] = None class _UpperCAmelCase ( TensorFormatter[Mapping, """jax.Array""", Mapping]): def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): super().__init__(features=snake_case_ ) import jax from jaxlib.xla_client import Device if isinstance(snake_case_ , snake_case_ ): raise ValueError( F'Expected {device} to be a `str` not {type(snake_case_ )}, as `jaxlib.xla_extension.Device` ' "is not serializable neither with `pickle` nor with `dill`. Instead you can surround " "the device with `str()` to get its string identifier that will be internally mapped " "to the actual `jaxlib.xla_extension.Device`." ) _snake_case : Union[str, Any] = device if isinstance(snake_case_ , snake_case_ ) else str(jax.devices()[0] ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: _snake_case : Tuple = self._map_devices_to_str() if self.device not in list(DEVICE_MAPPING.keys() ): logger.warning( F'Device with string identifier {self.device} not listed among the available ' F'devices: {list(DEVICE_MAPPING.keys() )}, so falling back to the default ' F'device: {str(jax.devices()[0] )}.' ) _snake_case : Dict = str(jax.devices()[0] ) _snake_case : List[Any] = jnp_array_kwargs @staticmethod def lowerCamelCase__ ( ): import jax return {str(snake_case_ ): device for device in jax.devices()} def lowerCamelCase__ ( self , snake_case_ ): import jax import jax.numpy as jnp if isinstance(snake_case_ , snake_case_ ) and column: if all( isinstance(snake_case_ , jax.Array ) and x.shape == column[0].shape and x.dtype == column[0].dtype for x in column ): return jnp.stack(snake_case_ , axis=0 ) return column def lowerCamelCase__ ( self , snake_case_ ): import jax import jax.numpy as jnp if isinstance(snake_case_ , (str, bytes, type(snake_case_ )) ): return value elif isinstance(snake_case_ , (np.character, np.ndarray) ) and np.issubdtype(value.dtype , np.character ): return value.tolist() _snake_case : Union[str, Any] = {} if isinstance(snake_case_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.integer ): # the default int precision depends on the jax config # see https://jax.readthedocs.io/en/latest/notebooks/Common_Gotchas_in_JAX.html#double-64bit-precision if jax.config.jax_enable_xaa: _snake_case : Optional[int] = {"dtype": jnp.intaa} else: _snake_case : Union[str, Any] = {"dtype": jnp.intaa} elif isinstance(snake_case_ , (np.number, np.ndarray) ) and np.issubdtype(value.dtype , np.floating ): _snake_case : List[str] = {"dtype": jnp.floataa} elif config.PIL_AVAILABLE and "PIL" in sys.modules: import PIL.Image if isinstance(snake_case_ , PIL.Image.Image ): _snake_case : Optional[Any] = np.asarray(snake_case_ ) # using global variable since `jaxlib.xla_extension.Device` is not serializable neither # with `pickle` nor with `dill`, so we need to use a global variable instead global DEVICE_MAPPING if DEVICE_MAPPING is None: _snake_case : str = self._map_devices_to_str() with jax.default_device(DEVICE_MAPPING[self.device] ): # calling jnp.array on a np.ndarray does copy the data # see https://github.com/google/jax/issues/4486 return jnp.array(snake_case_ , **{**default_dtype, **self.jnp_array_kwargs} ) def lowerCamelCase__ ( self , snake_case_ ): import jax # support for torch, tf, jax etc. if config.TORCH_AVAILABLE and "torch" in sys.modules: import torch if isinstance(snake_case_ , torch.Tensor ): return self._tensorize(data_struct.detach().cpu().numpy()[()] ) if hasattr(snake_case_ , "__array__" ) and not isinstance(snake_case_ , jax.Array ): _snake_case : Tuple = data_struct.__array__() # support for nested types like struct of list of struct if isinstance(snake_case_ , np.ndarray ): if data_struct.dtype == object: # jax arrays cannot be instantied from an array of objects return self._consolidate([self.recursive_tensorize(snake_case_ ) for substruct in data_struct] ) elif isinstance(snake_case_ , (list, tuple) ): return self._consolidate([self.recursive_tensorize(snake_case_ ) for substruct in data_struct] ) return self._tensorize(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): return map_nested(self._recursive_tensorize , snake_case_ , map_list=snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : List[str] = self.numpy_arrow_extractor().extract_row(snake_case_ ) _snake_case : Optional[int] = self.python_features_decoder.decode_row(snake_case_ ) return self.recursive_tensorize(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Any = self.numpy_arrow_extractor().extract_column(snake_case_ ) _snake_case : Any = self.python_features_decoder.decode_column(snake_case_ , pa_table.column_names[0] ) _snake_case : Any = self.recursive_tensorize(snake_case_ ) _snake_case : str = self._consolidate(snake_case_ ) return column def lowerCamelCase__ ( self , snake_case_ ): _snake_case : int = self.numpy_arrow_extractor().extract_batch(snake_case_ ) _snake_case : Optional[int] = self.python_features_decoder.decode_batch(snake_case_ ) _snake_case : Union[str, Any] = self.recursive_tensorize(snake_case_ ) for column_name in batch: _snake_case : int = self._consolidate(batch[column_name] ) return batch
701
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
702
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Any = TextToVideoSDPipeline __lowercase : str = TEXT_TO_IMAGE_PARAMS __lowercase : int = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __lowercase : Optional[int] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) _snake_case : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : Tuple = CLIPTextModel(snake_case_ ) _snake_case : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : str = torch.manual_seed(snake_case_ ) else: _snake_case : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def lowerCamelCase__ ( self ): _snake_case : int = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : Optional[Any] = self.get_dummy_components() _snake_case : Tuple = TextToVideoSDPipeline(**snake_case_ ) _snake_case : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ ) _snake_case : Union[str, Any] = "np" _snake_case : Dict = sd_pipe(**snake_case_ ).frames _snake_case : Any = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _snake_case : Dict = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=1E-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) _snake_case : int = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _snake_case : Tuple = pipe.to("cuda" ) _snake_case : List[Any] = "Spiderman is surfing" _snake_case : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : int = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="pt" ).frames _snake_case : int = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def lowerCamelCase__ ( self ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) _snake_case : str = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : int = pipe.to("cuda" ) _snake_case : Any = "Spiderman is surfing" _snake_case : str = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Any = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type="pt" ).frames _snake_case : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
87
0
"""simple docstring""" import json import os from collections import Counter import torch import torchvision import torchvision.transforms as transforms from PIL import Image from torch import nn from torch.utils.data import Dataset _a : Union[str, Any] = {1: (1, 1), 2: (2, 1), 3: (3, 1), 4: (2, 2), 5: (5, 1), 6: (3, 2), 7: (7, 1), 8: (4, 2), 9: (3, 3)} class _UpperCAmelCase ( nn.Module): def __init__( self , snake_case_ ): super().__init__() _snake_case : str = torchvision.models.resnetaaa(pretrained=snake_case_ ) _snake_case : List[str] = list(model.children() )[:-2] _snake_case : List[str] = nn.Sequential(*snake_case_ ) _snake_case : List[str] = nn.AdaptiveAvgPoolad(POOLING_BREAKDOWN[args.num_image_embeds] ) def lowerCamelCase__ ( self , snake_case_ ): # Bx3x224x224 -> Bx2048x7x7 -> Bx2048xN -> BxNx2048 _snake_case : int = self.pool(self.model(snake_case_ ) ) _snake_case : List[str] = torch.flatten(snake_case_ , start_dim=2 ) _snake_case : Optional[int] = out.transpose(1 , 2 ).contiguous() return out # BxNx2048 class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Tuple = [json.loads(snake_case_ ) for l in open(snake_case_ )] _snake_case : Dict = os.path.dirname(snake_case_ ) _snake_case : List[str] = tokenizer _snake_case : List[Any] = labels _snake_case : str = len(snake_case_ ) _snake_case : Tuple = max_seq_length _snake_case : str = transforms def __len__( self ): return len(self.data ) def __getitem__( self , snake_case_ ): _snake_case : List[Any] = torch.LongTensor(self.tokenizer.encode(self.data[index]["text"] , add_special_tokens=snake_case_ ) ) _snake_case : int = sentence[0], sentence[1:-1], sentence[-1] _snake_case : Union[str, Any] = sentence[: self.max_seq_length] _snake_case : str = torch.zeros(self.n_classes ) _snake_case : str = 1 _snake_case : Any = Image.open(os.path.join(self.data_dir , self.data[index]["img"] ) ).convert("RGB" ) _snake_case : List[Any] = self.transforms(snake_case_ ) return { "image_start_token": start_token, "image_end_token": end_token, "sentence": sentence, "image": image, "label": label, } def lowerCamelCase__ ( self ): _snake_case : str = Counter() for row in self.data: label_freqs.update(row["label"] ) return label_freqs def a__ ( a : Tuple ): """simple docstring""" _snake_case : Optional[Any] = [len(row["sentence"] ) for row in batch] _snake_case : Optional[Any] = len(a ), max(a ) _snake_case : Any = torch.zeros(a , a , dtype=torch.long ) _snake_case : List[Any] = torch.zeros(a , a , dtype=torch.long ) for i_batch, (input_row, length) in enumerate(zip(a , a ) ): _snake_case : str = input_row["sentence"] _snake_case : List[Any] = 1 _snake_case : Tuple = torch.stack([row["image"] for row in batch] ) _snake_case : Dict = torch.stack([row["label"] for row in batch] ) _snake_case : Any = torch.stack([row["image_start_token"] for row in batch] ) _snake_case : str = torch.stack([row["image_end_token"] for row in batch] ) return text_tensor, mask_tensor, img_tensor, img_start_token, img_end_token, tgt_tensor def a__ ( ): """simple docstring""" return [ "Crime", "Drama", "Thriller", "Action", "Comedy", "Romance", "Documentary", "Short", "Mystery", "History", "Family", "Adventure", "Fantasy", "Sci-Fi", "Western", "Horror", "Sport", "War", "Music", "Musical", "Animation", "Biography", "Film-Noir", ] def a__ ( ): """simple docstring""" return transforms.Compose( [ transforms.Resize(256 ), transforms.CenterCrop(224 ), transforms.ToTensor(), transforms.Normalize( mean=[0.46777044, 0.44531429, 0.40661017] , std=[0.12221994, 0.12145835, 0.14380469] , ), ] )
703
"""simple docstring""" from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _UpperCAmelCase ( _snake_case): __lowercase : int = """EncodecFeatureExtractor""" __lowercase : str = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) _snake_case : Dict = self.feature_extractor _snake_case : Any = False def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_=True ): return self.tokenizer.get_decoder_prompt_ids(task=snake_case_ , language=snake_case_ , no_timestamps=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*snake_case_ , **snake_case_ ) _snake_case : str = kwargs.pop("audio" , snake_case_ ) _snake_case : Optional[int] = kwargs.pop("sampling_rate" , snake_case_ ) _snake_case : Optional[Any] = kwargs.pop("text" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Any = args[0] _snake_case : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: _snake_case : Any = self.tokenizer(snake_case_ , **snake_case_ ) if audio is not None: _snake_case : Any = self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) if audio is None: return inputs elif text is None: return audio_inputs else: _snake_case : str = audio_inputs["input_values"] if "padding_mask" in audio_inputs: _snake_case : List[str] = audio_inputs["padding_mask"] return inputs def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): _snake_case : Tuple = kwargs.pop("audio" , snake_case_ ) _snake_case : List[str] = kwargs.pop("padding_mask" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Dict = args[1:] if audio_values is not None: return self._decode_audio(snake_case_ , padding_mask=snake_case_ ) else: return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Optional[int] = to_numpy(snake_case_ ) _snake_case , _snake_case , _snake_case : Tuple = audio_values.shape if padding_mask is None: return list(snake_case_ ) _snake_case : Optional[int] = to_numpy(snake_case_ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _snake_case : Any = seq_len - padding_mask.shape[-1] _snake_case : Optional[Any] = 1 - self.feature_extractor.padding_value _snake_case : Optional[int] = np.pad(snake_case_ , ((0, 0), (0, difference)) , "constant" , constant_values=snake_case_ ) _snake_case : Any = audio_values.tolist() for i in range(snake_case_ ): _snake_case : Tuple = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _snake_case : Tuple = sliced_audio.reshape(snake_case_ , -1 ) return audio_values
87
0
import inspect import unittest from typing import List import numpy as np from transformers import EfficientFormerConfig from transformers.testing_utils import require_tf, require_vision, slow from transformers.utils import cached_property, is_tf_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerModel, ) from transformers.models.efficientformer.modeling_tf_efficientformer import ( TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_vision_available(): from PIL import Image from transformers import EfficientFormerImageProcessor class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ = 13 , snake_case_ = 64 , snake_case_ = 2 , snake_case_ = 3 , snake_case_ = 3 , snake_case_ = True , snake_case_ = True , snake_case_ = 1_28 , snake_case_=[16, 32, 64, 1_28] , snake_case_ = 7 , snake_case_ = 4 , snake_case_ = 37 , snake_case_ = "gelu" , snake_case_ = 0.1 , snake_case_ = 0.1 , snake_case_ = 10 , snake_case_ = 0.02 , snake_case_ = 2 , snake_case_ = 1 , snake_case_ = 1_28 , snake_case_ = [2, 2, 2, 2] , snake_case_ = 2 , snake_case_ = 2 , ): _snake_case : str = parent _snake_case : List[Any] = batch_size _snake_case : Dict = image_size _snake_case : Dict = patch_size _snake_case : Optional[int] = num_channels _snake_case : List[str] = is_training _snake_case : int = use_labels _snake_case : Tuple = hidden_size _snake_case : str = num_hidden_layers _snake_case : Tuple = num_attention_heads _snake_case : List[str] = intermediate_size _snake_case : Optional[Any] = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : List[str] = type_sequence_label_size _snake_case : int = initializer_range _snake_case : List[Any] = encoder_stride _snake_case : Optional[int] = num_attention_outputs _snake_case : Optional[Any] = embed_dim _snake_case : Union[str, Any] = embed_dim + 1 _snake_case : Union[str, Any] = resolution _snake_case : Tuple = depths _snake_case : Any = hidden_sizes _snake_case : int = dim _snake_case : str = mlp_expansion_ratio def lowerCamelCase__ ( self ): _snake_case : List[str] = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case : int = None if self.use_labels: _snake_case : int = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : str = self.get_config() return config, pixel_values, labels def lowerCamelCase__ ( self ): return EfficientFormerConfig( image_size=self.image_size , patch_size=self.patch_size , num_channels=self.num_channels , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=snake_case_ , initializer_range=self.initializer_range , encoder_stride=self.encoder_stride , resolution=self.resolution , depths=self.depths , hidden_sizes=self.hidden_sizes , dim=self.dim , mlp_expansion_ratio=self.mlp_expansion_ratio , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = TFEfficientFormerModel(config=snake_case_ ) _snake_case : Tuple = model(snake_case_ , training=snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = self.type_sequence_label_size _snake_case : Any = TFEfficientFormerForImageClassification(snake_case_ ) _snake_case : Tuple = model(snake_case_ , labels=snake_case_ , training=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) # test greyscale images _snake_case : Tuple = 1 _snake_case : Optional[Any] = TFEfficientFormerForImageClassification(snake_case_ ) _snake_case : Tuple = floats_tensor([self.batch_size, 1, self.image_size, self.image_size] ) _snake_case : Tuple = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.type_sequence_label_size) ) def lowerCamelCase__ ( self ): _snake_case : str = self.prepare_config_and_inputs() _snake_case : Optional[Any] = config_and_inputs _snake_case : Any = {"pixel_values": pixel_values} return config, inputs_dict @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Tuple = ( ( TFEfficientFormerModel, TFEfficientFormerForImageClassificationWithTeacher, TFEfficientFormerForImageClassification, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFEfficientFormerModel, """image-classification""": ( TFEfficientFormerForImageClassification, TFEfficientFormerForImageClassificationWithTeacher, ), } if is_tf_available() else {} ) __lowercase : str = False __lowercase : str = False __lowercase : str = False __lowercase : str = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self ): _snake_case : List[Any] = TFEfficientFormerModelTester(self ) _snake_case : Optional[int] = ConfigTester( self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() @unittest.skip(reason="EfficientFormer does not use inputs_embeds" ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="EfficientFormer does not support input and output embeddings" ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): _snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : List[str] = model_class(snake_case_ ) _snake_case : Union[str, Any] = inspect.signature(model.call ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : Optional[Any] = [*signature.parameters.keys()] _snake_case : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def lowerCamelCase__ ( self ): def check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = model_class(snake_case_ ) _snake_case : Dict = model(**self._prepare_for_class(snake_case_ , snake_case_ ) , training=snake_case_ ) _snake_case : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case : int = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) if hasattr(self.model_tester , "encoder_seq_length" ): _snake_case : List[Any] = self.model_tester.encoder_seq_length if hasattr(self.model_tester , "chunk_length" ) and self.model_tester.chunk_length > 1: _snake_case : List[Any] = seq_length * self.model_tester.chunk_length else: _snake_case : List[str] = self.model_tester.seq_length self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [seq_length, self.model_tester.hidden_size] , ) if config.is_encoder_decoder: _snake_case : List[str] = outputs.decoder_hidden_states self.asseretIsInstance(snake_case_ , (list, tuple) ) self.assertEqual(len(snake_case_ ) , snake_case_ ) _snake_case : List[Any] = getattr(self.model_tester , "seq_length" , snake_case_ ) _snake_case : List[Any] = getattr(self.model_tester , "decoder_seq_length" , snake_case_ ) self.assertListEqual( list(hidden_states[-1].shape[-2:] ) , [decoder_seq_length, self.model_tester.hidden_size] , ) _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : str = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : List[str] = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : List[str] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class.__name__ == "TFEfficientFormerForImageClassificationWithTeacher": del inputs_dict["labels"] return inputs_dict def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) @unittest.skip(reason="EfficientFormer does not implement masked image modeling yet" ) def lowerCamelCase__ ( self ): _snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_image_modeling(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): for model_name in TF_EFFICIENTFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : Optional[int] = TFEfficientFormerModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Optional[int] = True _snake_case : str = getattr(self.model_tester , "seq_length" , snake_case_ ) _snake_case : Optional[int] = getattr(self.model_tester , "encoder_seq_length" , snake_case_ ) _snake_case : Dict = getattr(self.model_tester , "key_length" , snake_case_ ) _snake_case : Optional[int] = getattr(self.model_tester , "chunk_length" , snake_case_ ) if chunk_length is not None and hasattr(self.model_tester , "num_hashes" ): _snake_case : List[Any] = encoder_seq_length * self.model_tester.num_hashes for model_class in self.all_model_classes: _snake_case : List[str] = True _snake_case : Dict = False _snake_case : Tuple = True _snake_case : Optional[Any] = model_class(snake_case_ ) _snake_case : Optional[int] = model(**self._prepare_for_class(snake_case_ , snake_case_ ) , training=snake_case_ ) _snake_case : List[str] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_attention_outputs ) # check that output_attentions also work using config del inputs_dict["output_attentions"] _snake_case : Optional[Any] = True _snake_case : Union[str, Any] = model_class(snake_case_ ) _snake_case : List[Any] = model(**self._prepare_for_class(snake_case_ , snake_case_ ) , training=snake_case_ ) _snake_case : Optional[Any] = outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_attention_outputs ) if chunk_length is not None: self.assertListEqual( list(attentions[0].shape[-4:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, chunk_length, encoder_key_length] , ) else: self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads, encoder_seq_length, encoder_key_length] , ) def lowerCamelCase__ ( self ): # We use a simplified version of this test for EfficientFormer because it requires training=False # and Keras refuses to let us force that during functional construction _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: # Prepare our model _snake_case : Tuple = model_class(snake_case_ ) # These are maximally general inputs for the model, with multiple None dimensions # Hopefully this will catch any conditionals that fail for flexible shapes _snake_case : str = { key: tf.keras.Input(shape=val.shape[1:] , dtype=val.dtype , name=snake_case_ ) for key, val in model.input_signature.items() if key in model.dummy_inputs } _snake_case : str = model(snake_case_ ) self.assertTrue(outputs_dict is not None ) def a__ ( ): """simple docstring""" _snake_case : Any = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf @require_vision class _UpperCAmelCase ( unittest.TestCase): @cached_property def lowerCamelCase__ ( self ): return ( EfficientFormerImageProcessor.from_pretrained("snap-research/efficientformer-l1-300" ) if is_vision_available() else None ) @slow def lowerCamelCase__ ( self ): _snake_case : List[str] = TFEfficientFormerForImageClassification.from_pretrained("snap-research/efficientformer-l1-300" ) _snake_case : Any = self.default_image_processor _snake_case : Optional[Any] = prepare_img() _snake_case : str = image_processor(images=snake_case_ , return_tensors="tf" ) # forward pass _snake_case : List[str] = model(**snake_case_ , training=snake_case_ ) # verify the logits _snake_case : Tuple = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant([-0.0555, 0.4825, -0.0852] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , snake_case_ , atol=1E-4 ) ) @slow def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = TFEfficientFormerForImageClassificationWithTeacher.from_pretrained( "snap-research/efficientformer-l1-300" ) _snake_case : str = self.default_image_processor _snake_case : List[Any] = prepare_img() _snake_case : str = image_processor(images=snake_case_ , return_tensors="tf" ) # forward pass _snake_case : Dict = model(**snake_case_ , training=snake_case_ ) # verify the logits _snake_case : Any = tf.TensorShape((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case_ ) _snake_case : List[str] = tf.constant([-0.1312, 0.4353, -1.0499] ) self.assertTrue(np.allclose(outputs.logits[0, :3] , snake_case_ , atol=1E-4 ) )
704
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : str = {"""configuration_yolos""": ["""YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """YolosConfig""", """YolosOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""YolosFeatureExtractor"""] _a : List[Any] = ["""YolosImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST""", """YolosForObjectDetection""", """YolosModel""", """YolosPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" import unittest from transformers import BertGenerationConfig, is_torch_available from transformers.testing_utils import require_torch, slow, torch_device from ...generation.test_utils import GenerationTesterMixin from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import BertGenerationDecoder, BertGenerationEncoder class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=5 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=50 , snake_case_=0.02 , snake_case_=True , snake_case_=None , ): _snake_case : Optional[int] = parent _snake_case : Any = batch_size _snake_case : Union[str, Any] = seq_length _snake_case : List[str] = is_training _snake_case : Any = use_input_mask _snake_case : Optional[Any] = vocab_size _snake_case : str = hidden_size _snake_case : Any = num_hidden_layers _snake_case : Optional[int] = num_attention_heads _snake_case : Tuple = intermediate_size _snake_case : int = hidden_act _snake_case : int = hidden_dropout_prob _snake_case : Any = attention_probs_dropout_prob _snake_case : Tuple = max_position_embeddings _snake_case : List[str] = initializer_range _snake_case : Tuple = use_labels _snake_case : int = scope def lowerCamelCase__ ( self ): _snake_case : Tuple = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Any = None if self.use_input_mask: _snake_case : Tuple = random_attention_mask([self.batch_size, self.seq_length] ) if self.use_labels: _snake_case : str = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = self.get_config() return config, input_ids, input_mask, token_labels def lowerCamelCase__ ( self ): return BertGenerationConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , is_decoder=snake_case_ , initializer_range=self.initializer_range , ) def lowerCamelCase__ ( self ): ( _snake_case ) : int = self.prepare_config_and_inputs() _snake_case : Union[str, Any] = True _snake_case : Dict = floats_tensor([self.batch_size, self.seq_length, self.hidden_size] ) _snake_case : List[str] = ids_tensor([self.batch_size, self.seq_length] , vocab_size=2 ) return ( config, input_ids, input_mask, token_labels, encoder_hidden_states, encoder_attention_mask, ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): _snake_case : Tuple = BertGenerationEncoder(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : int = model(snake_case_ , attention_mask=snake_case_ ) _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): _snake_case : Any = True _snake_case : int = BertGenerationEncoder(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : int = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , ) _snake_case : int = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , **snake_case_ , ): _snake_case : Tuple = True _snake_case : Optional[Any] = True _snake_case : Optional[int] = BertGenerationDecoder(config=snake_case_ ).to(snake_case_ ).eval() # first forward pass _snake_case : Dict = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , use_cache=snake_case_ , ) _snake_case : Dict = outputs.past_key_values # create hypothetical multiple next token and extent to next_input_ids _snake_case : Optional[int] = ids_tensor((self.batch_size, 3) , config.vocab_size ) _snake_case : List[str] = ids_tensor((self.batch_size, 3) , vocab_size=2 ) # append to next input_ids and _snake_case : Optional[int] = torch.cat([input_ids, next_tokens] , dim=-1 ) _snake_case : List[Any] = torch.cat([input_mask, next_mask] , dim=-1 ) _snake_case : Optional[int] = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , output_hidden_states=snake_case_ , )["hidden_states"][0] _snake_case : Tuple = model( snake_case_ , attention_mask=snake_case_ , encoder_hidden_states=snake_case_ , encoder_attention_mask=snake_case_ , past_key_values=snake_case_ , output_hidden_states=snake_case_ , )["hidden_states"][0] # select random slice _snake_case : int = ids_tensor((1,) , output_from_past.shape[-1] ).item() _snake_case : List[str] = output_from_no_past[:, -3:, random_slice_idx].detach() _snake_case : int = output_from_past[:, :, random_slice_idx].detach() self.parent.assertTrue(output_from_past_slice.shape[1] == next_tokens.shape[1] ) # test that outputs are equal for slice self.parent.assertTrue(torch.allclose(snake_case_ , snake_case_ , atol=1E-3 ) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : List[str] = BertGenerationDecoder(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Dict = model(snake_case_ , attention_mask=snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask} return config, inputs_dict @require_torch class _UpperCAmelCase ( _snake_case , _snake_case , _snake_case , unittest.TestCase): __lowercase : Any = (BertGenerationEncoder, BertGenerationDecoder) if is_torch_available() else () __lowercase : List[Any] = (BertGenerationDecoder,) if is_torch_available() else () __lowercase : List[str] = ( {"""feature-extraction""": BertGenerationEncoder, """text-generation""": BertGenerationDecoder} if is_torch_available() else {} ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = BertGenerationEncoderTester(self ) _snake_case : Optional[int] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() _snake_case : Optional[Any] = "bert" self.model_tester.create_and_check_model(snake_case_ , snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_model_as_decoder(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_decoder_model_past_large_inputs(*snake_case_ ) def lowerCamelCase__ ( self ): # This regression test was failing with PyTorch < 1.3 ( _snake_case ) : Tuple = self.model_tester.prepare_config_and_inputs_for_decoder() _snake_case : List[str] = None self.model_tester.create_and_check_model_as_decoder( snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_decoder() self.model_tester.create_and_check_for_causal_lm(*snake_case_ ) @slow def lowerCamelCase__ ( self ): _snake_case : int = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) self.assertIsNotNone(snake_case_ ) @require_torch class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : str = BertGenerationEncoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) _snake_case : Optional[int] = torch.tensor([[1_01, 75_92, 10_10, 20_26, 38_99, 20_03, 1_01_40, 1_02]] ) with torch.no_grad(): _snake_case : int = model(snake_case_ )[0] _snake_case : str = torch.Size([1, 8, 10_24] ) self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[int] = torch.tensor( [[[0.1775, 0.0083, -0.0321], [1.6002, 0.1287, 0.3912], [2.1473, 0.5791, 0.6066]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case_ , atol=1E-4 ) ) @require_torch class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : List[Any] = BertGenerationDecoder.from_pretrained("google/bert_for_seq_generation_L-24_bbc_encoder" ) _snake_case : List[Any] = torch.tensor([[1_01, 75_92, 10_10, 20_26, 38_99, 20_03, 1_01_40, 1_02]] ) with torch.no_grad(): _snake_case : List[str] = model(snake_case_ )[0] _snake_case : str = torch.Size([1, 8, 5_03_58] ) self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[int] = torch.tensor( [[[-0.5788, -2.5994, -3.7054], [0.0438, 4.7997, 1.8795], [1.5862, 6.6409, 4.4638]]] ) self.assertTrue(torch.allclose(output[:, :3, :3] , snake_case_ , atol=1E-4 ) )
705
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = dataset _snake_case : str = process _snake_case : int = params def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): _snake_case : Union[str, Any] = self.dataset[i] _snake_case : Optional[Any] = self.process(snake_case_ , **self.params ) return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): _snake_case : Union[str, Any] = loader _snake_case : Tuple = infer _snake_case : List[Any] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _snake_case : int = None _snake_case : int = loader_batch_size # Internal bookkeeping _snake_case : Any = None _snake_case : Dict = None def __len__( self ): return len(self.loader ) def __iter__( self ): _snake_case : int = iter(self.loader ) return self def lowerCamelCase__ ( self ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _snake_case : List[Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _snake_case : int = {} for k, element in self._loader_batch_data.items(): if isinstance(snake_case_ , snake_case_ ): # Convert ModelOutput to tuple first _snake_case : Tuple = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _snake_case : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(snake_case_ , snake_case_ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _snake_case : Tuple = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _snake_case : Tuple = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : List[Any] = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : Union[str, Any] = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _snake_case : List[Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _snake_case : int = self._loader_batch_data.__class__(snake_case_ ) self._loader_batch_index += 1 return result def lowerCamelCase__ ( self ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _snake_case : Tuple = next(self.iterator ) _snake_case : Any = self.infer(snake_case_ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Optional[int] = list(processed.keys() )[0] _snake_case : List[str] = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = len(snake_case_ ) else: _snake_case : Optional[int] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Union[str, Any] = observed_batch_size # Setting internal index to unwrap the batch _snake_case : str = processed _snake_case : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): super().__init__(snake_case_ , snake_case_ , snake_case_ ) def __iter__( self ): _snake_case : Tuple = iter(self.loader ) _snake_case : List[Any] = None return self def lowerCamelCase__ ( self ): if self.subiterator is None: _snake_case : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _snake_case : Union[str, Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _snake_case : str = self.infer(next(self.iterator ) , **self.params ) _snake_case : Tuple = next(self.subiterator ) return processed class _UpperCAmelCase ( _snake_case): def __iter__( self ): _snake_case : Optional[Any] = iter(self.loader ) return self def lowerCamelCase__ ( self ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _snake_case : Optional[Any] = False _snake_case : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : str = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator while not is_last: _snake_case : List[str] = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Tuple = list(processed.keys() )[0] _snake_case : Tuple = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Any = len(snake_case_ ) else: _snake_case : List[Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Dict = observed_batch_size _snake_case : List[Any] = processed _snake_case : List[str] = 0 while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : int = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator else: _snake_case : Dict = processed _snake_case : Dict = item.pop("is_last" ) accumulator.append(snake_case_ ) return accumulator class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ ): _snake_case : str = dataset _snake_case : Any = key def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return self.dataset[i][self.key] class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = dataset _snake_case : Any = keya _snake_case : int = keya def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
87
0
"""simple docstring""" from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available lowerCamelCase_ : Optional[Any] = { """configuration_cpmant""": ["""CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """CpmAntConfig"""], """tokenization_cpmant""": ["""CpmAntTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[Any] = [ """CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST""", """CpmAntForCausalLM""", """CpmAntModel""", """CpmAntPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys lowerCamelCase_ : str = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
706
"""simple docstring""" def a__ ( a : int ): """simple docstring""" if not isinstance(a , a ): raise TypeError("Input value must be an 'int' type" ) _snake_case : Union[str, Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import json import os import shutil import sys import tempfile import unittest import unittest.mock as mock from pathlib import Path from huggingface_hub import HfFolder, delete_repo from requests.exceptions import HTTPError from transformers import AutoConfig, BertConfig, GPTaConfig from transformers.configuration_utils import PretrainedConfig from transformers.testing_utils import TOKEN, USER, is_staging_test sys.path.append(str(Path(__file__).parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 _a : List[Any] = { """return_dict""": False, """output_hidden_states""": True, """output_attentions""": True, """torchscript""": True, """torch_dtype""": """float16""", """use_bfloat16""": True, """tf_legacy_loss""": True, """pruned_heads""": {"""a""": 1}, """tie_word_embeddings""": False, """is_decoder""": True, """cross_attention_hidden_size""": 128, """add_cross_attention""": True, """tie_encoder_decoder""": True, """max_length""": 50, """min_length""": 3, """do_sample""": True, """early_stopping""": True, """num_beams""": 3, """num_beam_groups""": 3, """diversity_penalty""": 0.5, """temperature""": 2.0, """top_k""": 10, """top_p""": 0.7, """typical_p""": 0.2, """repetition_penalty""": 0.8, """length_penalty""": 0.8, """no_repeat_ngram_size""": 5, """encoder_no_repeat_ngram_size""": 5, """bad_words_ids""": [1, 2, 3], """num_return_sequences""": 3, """chunk_size_feed_forward""": 5, """output_scores""": True, """return_dict_in_generate""": True, """forced_bos_token_id""": 2, """forced_eos_token_id""": 3, """remove_invalid_values""": True, """architectures""": ["""BertModel"""], """finetuning_task""": """translation""", """id2label""": {0: """label"""}, """label2id""": {"""label""": """0"""}, """tokenizer_class""": """BertTokenizerFast""", """prefix""": """prefix""", """bos_token_id""": 6, """pad_token_id""": 7, """eos_token_id""": 8, """sep_token_id""": 9, """decoder_start_token_id""": 10, """exponential_decay_length_penalty""": (5, 1.01), """suppress_tokens""": [0, 1], """begin_suppress_tokens""": 2, """task_specific_params""": {"""translation""": """some_params"""}, """problem_type""": """regression""", } @is_staging_test class _UpperCAmelCase ( unittest.TestCase): @classmethod def lowerCamelCase__ ( cls ): _snake_case : Union[str, Any] = TOKEN HfFolder.save_token(snake_case_ ) @classmethod def lowerCamelCase__ ( cls ): try: delete_repo(token=cls._token , repo_id="test-config" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="valid_org/test-config-org" ) except HTTPError: pass try: delete_repo(token=cls._token , repo_id="test-dynamic-config" ) except HTTPError: pass def lowerCamelCase__ ( self ): _snake_case : str = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("test-config" , use_auth_token=self._token ) _snake_case : List[Any] = BertConfig.from_pretrained(F'{USER}/test-config' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(snake_case_ , getattr(snake_case_ , snake_case_ ) ) # Reset repo delete_repo(token=self._token , repo_id="test-config" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained(snake_case_ , repo_id="test-config" , push_to_hub=snake_case_ , use_auth_token=self._token ) _snake_case : Tuple = BertConfig.from_pretrained(F'{USER}/test-config' ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(snake_case_ , getattr(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Tuple = BertConfig( vocab_size=99 , hidden_size=32 , num_hidden_layers=5 , num_attention_heads=4 , intermediate_size=37 ) config.push_to_hub("valid_org/test-config-org" , use_auth_token=self._token ) _snake_case : List[Any] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(snake_case_ , getattr(snake_case_ , snake_case_ ) ) # Reset repo delete_repo(token=self._token , repo_id="valid_org/test-config-org" ) # Push to hub via save_pretrained with tempfile.TemporaryDirectory() as tmp_dir: config.save_pretrained( snake_case_ , repo_id="valid_org/test-config-org" , push_to_hub=snake_case_ , use_auth_token=self._token ) _snake_case : List[Any] = BertConfig.from_pretrained("valid_org/test-config-org" ) for k, v in config.to_dict().items(): if k != "transformers_version": self.assertEqual(snake_case_ , getattr(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): CustomConfig.register_for_auto_class() _snake_case : List[Any] = CustomConfig(attribute=42 ) config.push_to_hub("test-dynamic-config" , use_auth_token=self._token ) # This has added the proper auto_map field to the config self.assertDictEqual(config.auto_map , {"AutoConfig": "custom_configuration.CustomConfig"} ) _snake_case : Optional[int] = AutoConfig.from_pretrained(F'{USER}/test-dynamic-config' , trust_remote_code=snake_case_ ) # Can't make an isinstance check because the new_config is from the FakeConfig class of a dynamic module self.assertEqual(new_config.__class__.__name__ , "CustomConfig" ) self.assertEqual(new_config.attribute , 42 ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = GPTaConfig() # attempt to modify each of int/float/bool/str config records and verify they were updated _snake_case : Optional[Any] = c.n_embd + 1 # int _snake_case : Tuple = c.resid_pdrop + 1.0 # float _snake_case : int = not c.scale_attn_weights # bool _snake_case : str = c.summary_type + "foo" # str c.update_from_string( F'n_embd={n_embd},resid_pdrop={resid_pdrop},scale_attn_weights={scale_attn_weights},summary_type={summary_type}' ) self.assertEqual(snake_case_ , c.n_embd , "mismatch for key: n_embd" ) self.assertEqual(snake_case_ , c.resid_pdrop , "mismatch for key: resid_pdrop" ) self.assertEqual(snake_case_ , c.scale_attn_weights , "mismatch for key: scale_attn_weights" ) self.assertEqual(snake_case_ , c.summary_type , "mismatch for key: summary_type" ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = PretrainedConfig() _snake_case : Optional[int] = [key for key in base_config.__dict__ if key not in config_common_kwargs] # If this part of the test fails, you have arguments to addin config_common_kwargs above. self.assertListEqual( snake_case_ , ["is_encoder_decoder", "_name_or_path", "_commit_hash", "transformers_version"] ) _snake_case : int = [key for key, value in config_common_kwargs.items() if value == getattr(snake_case_ , snake_case_ )] if len(snake_case_ ) > 0: raise ValueError( "The following keys are set with the default values in" " `test_configuration_common.config_common_kwargs` pick another value for them:" F' {", ".join(snake_case_ )}.' ) def lowerCamelCase__ ( self ): with self.assertRaises(snake_case_ ): # config is in subfolder, the following should not work without specifying the subfolder _snake_case : Optional[int] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" ) _snake_case : Any = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert-subfolder" , subfolder="bert" ) self.assertIsNotNone(snake_case_ ) def lowerCamelCase__ ( self ): # A mock response for an HTTP head request to emulate server down _snake_case : Optional[int] = mock.Mock() _snake_case : Dict = 5_00 _snake_case : Union[str, Any] = {} _snake_case : List[Any] = HTTPError _snake_case : Union[str, Any] = {} # Download this model to make sure it's in the cache. _snake_case : Optional[Any] = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # Under the mock environment we get a 500 error when trying to reach the model. with mock.patch("requests.Session.request" , return_value=snake_case_ ) as mock_head: _snake_case : str = BertConfig.from_pretrained("hf-internal-testing/tiny-random-bert" ) # This check we did call the fake head request mock_head.assert_called() def lowerCamelCase__ ( self ): # This test is for deprecated behavior and can be removed in v5 _snake_case : Union[str, Any] = BertConfig.from_pretrained( "https://huggingface.co/hf-internal-testing/tiny-random-bert/resolve/main/config.json" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = AutoConfig.from_pretrained("bert-base-cased" ) _snake_case : Dict = ["config.4.0.0.json"] with tempfile.TemporaryDirectory() as tmp_dir: configuration.save_pretrained(snake_case_ ) _snake_case : Optional[Any] = 2 json.dump(configuration.to_dict() , open(os.path.join(snake_case_ , "config.4.0.0.json" ) , "w" ) ) # This should pick the new configuration file as the version of Transformers is > 4.0.0 _snake_case : List[Any] = AutoConfig.from_pretrained(snake_case_ ) self.assertEqual(new_configuration.hidden_size , 2 ) # Will need to be adjusted if we reach v42 and this test is still here. # Should pick the old configuration file as the version of Transformers is < 4.42.0 _snake_case : Optional[Any] = ["config.42.0.0.json"] _snake_case : str = 7_68 configuration.save_pretrained(snake_case_ ) shutil.move(os.path.join(snake_case_ , "config.4.0.0.json" ) , os.path.join(snake_case_ , "config.42.0.0.json" ) ) _snake_case : Any = AutoConfig.from_pretrained(snake_case_ ) self.assertEqual(new_configuration.hidden_size , 7_68 ) def lowerCamelCase__ ( self ): # This repo has two configuration files, one for v4.0.0 and above with a different hidden size. _snake_case : Optional[Any] = "hf-internal-testing/test-two-configs" import transformers as new_transformers _snake_case : Tuple = "v4.0.0" _snake_case : Union[str, Any] = new_transformers.models.auto.AutoConfig.from_pretrained( snake_case_ , return_unused_kwargs=snake_case_ ) self.assertEqual(new_configuration.hidden_size , 2 ) # This checks `_configuration_file` ia not kept in the kwargs by mistake. self.assertDictEqual(snake_case_ , {} ) # Testing an older version by monkey-patching the version in the module it's used. import transformers as old_transformers _snake_case : str = "v3.0.0" _snake_case : Optional[int] = old_transformers.models.auto.AutoConfig.from_pretrained(snake_case_ ) self.assertEqual(old_configuration.hidden_size , 7_68 )
707
"""simple docstring""" from __future__ import annotations import requests _a : List[str] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def a__ ( a : str , a : int = 1 , a : str = "new" , a : list | None = None ): """simple docstring""" _snake_case : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(a ) - valid_terms ) ): _snake_case : Optional[int] = f'Invalid search term: {invalid_search_terms}' raise ValueError(a ) _snake_case : int = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"User-agent": "A random string"} , ) if response.status_code == 429: raise requests.HTTPError _snake_case : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(a )} _snake_case : Tuple = {} for id_ in range(a ): _snake_case : List[str] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
87
0
"""simple docstring""" import random import unittest import numpy as np import torch from diffusers import ( DPMSolverMultistepScheduler, EulerAncestralDiscreteScheduler, EulerDiscreteScheduler, LMSDiscreteScheduler, OnnxStableDiffusionUpscalePipeline, PNDMScheduler, ) from diffusers.utils import floats_tensor from diffusers.utils.testing_utils import ( is_onnx_available, load_image, nightly, require_onnxruntime, require_torch_gpu, ) from ..test_pipelines_onnx_common import OnnxPipelineTesterMixin if is_onnx_available(): import onnxruntime as ort class _UpperCAmelCase ( _snake_case , unittest.TestCase): # TODO: is there an appropriate internal test set? __lowercase : int = """ssube/stable-diffusion-x4-upscaler-onnx""" def lowerCamelCase__ ( self , snake_case_=0 ): _snake_case : str = floats_tensor((1, 3, 1_28, 1_28) , rng=random.Random(snake_case_ ) ) _snake_case : Any = torch.manual_seed(snake_case_ ) _snake_case : List[str] = { "prompt": "A painting of a squirrel eating a burger", "image": image, "generator": generator, "num_inference_steps": 3, "guidance_scale": 7.5, "output_type": "numpy", } return inputs def lowerCamelCase__ ( self ): _snake_case : Dict = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Optional[int] = self.get_dummy_inputs() _snake_case : Optional[Any] = pipe(**snake_case_ ).images _snake_case : List[str] = image[0, -3:, -3:, -1].flatten() # started as 128, should now be 512 assert image.shape == (1, 5_12, 5_12, 3) _snake_case : Any = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice - expected_slice ).max() < 1E-1 def lowerCamelCase__ ( self ): _snake_case : List[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case : Optional[int] = PNDMScheduler.from_config(pipe.scheduler.config , skip_prk_steps=snake_case_ ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Tuple = self.get_dummy_inputs() _snake_case : List[Any] = pipe(**snake_case_ ).images _snake_case : Union[str, Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case : Optional[int] = np.array( [0.6898892, 0.59240556, 0.52499527, 0.58866215, 0.52258235, 0.52572715, 0.62414473, 0.6174387, 0.6214964] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase__ ( self ): _snake_case : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case : List[Any] = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : str = self.get_dummy_inputs() _snake_case : Union[str, Any] = pipe(**snake_case_ ).images _snake_case : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case : Any = np.array( [0.7659278, 0.76437664, 0.75579107, 0.7691116, 0.77666986, 0.7727672, 0.7758664, 0.7812226, 0.76942515] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase__ ( self ): _snake_case : Optional[int] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case : str = EulerDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Union[str, Any] = self.get_dummy_inputs() _snake_case : Optional[int] = pipe(**snake_case_ ).images _snake_case : Dict = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case : Any = np.array( [0.6974782, 0.68902093, 0.70135885, 0.7583618, 0.7804545, 0.7854912, 0.78667426, 0.78743863, 0.78070223] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained(self.hub_checkpoint , provider="CPUExecutionProvider" ) _snake_case : Union[str, Any] = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Dict = self.get_dummy_inputs() _snake_case : Optional[Any] = pipe(**snake_case_ ).images _snake_case : Optional[Any] = image[0, -3:, -3:, -1] assert image.shape == (1, 5_12, 5_12, 3) _snake_case : Tuple = np.array( [0.77424496, 0.773601, 0.7645288, 0.7769598, 0.7772739, 0.7738688, 0.78187233, 0.77879584, 0.767043] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-1 @nightly @require_onnxruntime @require_torch_gpu class _UpperCAmelCase ( unittest.TestCase): @property def lowerCamelCase__ ( self ): return ( "CUDAExecutionProvider", { "gpu_mem_limit": "15000000000", # 15GB "arena_extend_strategy": "kSameAsRequested", }, ) @property def lowerCamelCase__ ( self ): _snake_case : Tuple = ort.SessionOptions() _snake_case : str = False return options def lowerCamelCase__ ( self ): _snake_case : Dict = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) _snake_case : List[Any] = init_image.resize((1_28, 1_28) ) # using the PNDM scheduler by default _snake_case : List[str] = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Union[str, Any] = "A fantasy landscape, trending on artstation" _snake_case : Any = torch.manual_seed(0 ) _snake_case : Dict = pipe( prompt=snake_case_ , image=snake_case_ , guidance_scale=7.5 , num_inference_steps=10 , generator=snake_case_ , output_type="np" , ) _snake_case : Union[str, Any] = output.images _snake_case : List[Any] = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _snake_case : int = np.array([0.4883, 0.4947, 0.4980, 0.4975, 0.4982, 0.4980, 0.5000, 0.5006, 0.4972] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2 def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/img2img/sketch-mountains-input.jpg" ) _snake_case : str = init_image.resize((1_28, 1_28) ) _snake_case : Optional[Any] = LMSDiscreteScheduler.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , subfolder="scheduler" ) _snake_case : Optional[Any] = OnnxStableDiffusionUpscalePipeline.from_pretrained( "ssube/stable-diffusion-x4-upscaler-onnx" , scheduler=snake_case_ , provider=self.gpu_provider , sess_options=self.gpu_options , ) pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : Optional[int] = "A fantasy landscape, trending on artstation" _snake_case : Tuple = torch.manual_seed(0 ) _snake_case : List[Any] = pipe( prompt=snake_case_ , image=snake_case_ , guidance_scale=7.5 , num_inference_steps=20 , generator=snake_case_ , output_type="np" , ) _snake_case : str = output.images _snake_case : Dict = images[0, 2_55:2_58, 3_83:3_86, -1] assert images.shape == (1, 5_12, 5_12, 3) _snake_case : Optional[Any] = np.array( [0.50173753, 0.50223356, 0.502039, 0.50233036, 0.5023725, 0.5022601, 0.5018758, 0.50234085, 0.50241566] ) # TODO: lower the tolerance after finding the cause of onnxruntime reproducibility issues assert np.abs(image_slice.flatten() - expected_slice ).max() < 2E-2
708
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def a__ ( a : float , a : float , a : bool = False ): """simple docstring""" if radian_mode: return [magnitude * cos(a ), magnitude * sin(a )] return [magnitude * cos(radians(a ) ), magnitude * sin(radians(a ) )] def a__ ( a : NDArray[floataa] , a : NDArray[floataa] , a : float = 10**-1 ): """simple docstring""" _snake_case : NDArray[floataa] = cross(a , a ) _snake_case : float = sum(a ) return abs(a ) < eps if __name__ == "__main__": # Test to check if it works _a : Tuple = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) _a : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg _a : List[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) _a : List[Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg _a : List[str] = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) _a : Optional[int] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
87
0
"""simple docstring""" _a : Dict = { """Pillow""": """Pillow""", """accelerate""": """accelerate>=0.11.0""", """compel""": """compel==0.1.8""", """black""": """black~=23.1""", """datasets""": """datasets""", """filelock""": """filelock""", """flax""": """flax>=0.4.1""", """hf-doc-builder""": """hf-doc-builder>=0.3.0""", """huggingface-hub""": """huggingface-hub>=0.13.2""", """requests-mock""": """requests-mock==1.10.0""", """importlib_metadata""": """importlib_metadata""", """invisible-watermark""": """invisible-watermark""", """isort""": """isort>=5.5.4""", """jax""": """jax>=0.2.8,!=0.3.2""", """jaxlib""": """jaxlib>=0.1.65""", """Jinja2""": """Jinja2""", """k-diffusion""": """k-diffusion>=0.0.12""", """torchsde""": """torchsde""", """note_seq""": """note_seq""", """librosa""": """librosa""", """numpy""": """numpy""", """omegaconf""": """omegaconf""", """parameterized""": """parameterized""", """protobuf""": """protobuf>=3.20.3,<4""", """pytest""": """pytest""", """pytest-timeout""": """pytest-timeout""", """pytest-xdist""": """pytest-xdist""", """ruff""": """ruff>=0.0.241""", """safetensors""": """safetensors""", """sentencepiece""": """sentencepiece>=0.1.91,!=0.1.92""", """scipy""": """scipy""", """onnx""": """onnx""", """regex""": """regex!=2019.12.17""", """requests""": """requests""", """tensorboard""": """tensorboard""", """torch""": """torch>=1.4""", """torchvision""": """torchvision""", """transformers""": """transformers>=4.25.1""", """urllib3""": """urllib3<=2.0.0""", }
709
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[int] = logging.get_logger(__name__) _a : str = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class _UpperCAmelCase ( _snake_case): __lowercase : Optional[Any] = """openai-gpt""" __lowercase : Dict = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , snake_case_=4_04_78 , snake_case_=5_12 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_="cls_index" , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=0.1 , **snake_case_ , ): _snake_case : Tuple = vocab_size _snake_case : Dict = n_positions _snake_case : Any = n_embd _snake_case : Any = n_layer _snake_case : Optional[int] = n_head _snake_case : Union[str, Any] = afn _snake_case : Dict = resid_pdrop _snake_case : str = embd_pdrop _snake_case : Union[str, Any] = attn_pdrop _snake_case : str = layer_norm_epsilon _snake_case : Union[str, Any] = initializer_range _snake_case : Any = summary_type _snake_case : List[str] = summary_use_proj _snake_case : Optional[int] = summary_activation _snake_case : Union[str, Any] = summary_first_dropout _snake_case : Optional[int] = summary_proj_to_labels super().__init__(**snake_case_ )
87
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[Any] = logging.get_logger(__name__) _a : Dict = { """microsoft/cvt-13""": """https://huggingface.co/microsoft/cvt-13/resolve/main/config.json""", # See all Cvt models at https://huggingface.co/models?filter=cvt } class _UpperCAmelCase ( _snake_case): __lowercase : str = """cvt""" def __init__( self , snake_case_=3 , snake_case_=[7, 3, 3] , snake_case_=[4, 2, 2] , snake_case_=[2, 1, 1] , snake_case_=[64, 1_92, 3_84] , snake_case_=[1, 3, 6] , snake_case_=[1, 2, 10] , snake_case_=[4.0, 4.0, 4.0] , snake_case_=[0.0, 0.0, 0.0] , snake_case_=[0.0, 0.0, 0.0] , snake_case_=[0.0, 0.0, 0.1] , snake_case_=[True, True, True] , snake_case_=[False, False, True] , snake_case_=["dw_bn", "dw_bn", "dw_bn"] , snake_case_=[3, 3, 3] , snake_case_=[1, 1, 1] , snake_case_=[2, 2, 2] , snake_case_=[1, 1, 1] , snake_case_=[1, 1, 1] , snake_case_=0.02 , snake_case_=1E-12 , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Optional[Any] = num_channels _snake_case : str = patch_sizes _snake_case : Union[str, Any] = patch_stride _snake_case : Tuple = patch_padding _snake_case : Tuple = embed_dim _snake_case : Optional[Any] = num_heads _snake_case : Optional[int] = depth _snake_case : List[str] = mlp_ratio _snake_case : str = attention_drop_rate _snake_case : Optional[int] = drop_rate _snake_case : str = drop_path_rate _snake_case : Dict = qkv_bias _snake_case : Union[str, Any] = cls_token _snake_case : Any = qkv_projection_method _snake_case : Union[str, Any] = kernel_qkv _snake_case : List[str] = padding_kv _snake_case : Tuple = stride_kv _snake_case : List[Any] = padding_q _snake_case : Dict = stride_q _snake_case : Optional[Any] = initializer_range _snake_case : int = layer_norm_eps
710
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _a : Tuple = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _a : int = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def a__ ( a : List[str] , a : int , a : int ): """simple docstring""" _snake_case : Union[str, Any] = state_dict.pop(a ) _snake_case : Union[str, Any] = val def a__ ( a : Tuple ): """simple docstring""" _snake_case : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case : Dict = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) _snake_case : Tuple = value else: _snake_case : Dict = value return new_state_dict def a__ ( a : int ): """simple docstring""" _snake_case : Any = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[:256, :] _snake_case : List[str] = in_proj_bias[:256] _snake_case : Optional[Any] = in_proj_weight[256:512, :] _snake_case : List[str] = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : Dict = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case : List[Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : Union[str, Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : Union[str, Any] = in_proj_weight[:256, :] _snake_case : Tuple = in_proj_bias[:256] _snake_case : int = in_proj_weight[256:512, :] _snake_case : int = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : str = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _snake_case : Dict = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _snake_case : Optional[int] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case : Dict = in_proj_weight_cross_attn[:256, :] _snake_case : Any = in_proj_bias_cross_attn[:256] _snake_case : Union[str, Any] = in_proj_weight_cross_attn[256:512, :] _snake_case : Optional[int] = in_proj_bias_cross_attn[256:512] _snake_case : Any = in_proj_weight_cross_attn[-256:, :] _snake_case : str = in_proj_bias_cross_attn[-256:] def a__ ( a : str , a : int ): """simple docstring""" _snake_case , _snake_case : List[str] = image.size _snake_case : Dict = max(a , a ) _snake_case : Union[str, Any] = 800 if "detection" in checkpoint_url else 1_000 _snake_case : Any = target_max_size / current_max_size _snake_case : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def a__ ( a : str ): """simple docstring""" _snake_case : str = F.to_tensor(a ) _snake_case : Union[str, Any] = F.normalize(a , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def a__ ( a : Optional[Any] , a : Any , a : Union[str, Any] ): """simple docstring""" logger.info("Converting model..." ) # load original state dict _snake_case : Tuple = torch.hub.load_state_dict_from_url(a , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(a , a , a ) _snake_case : Union[str, Any] = rename_backbone_keys(a ) # query, key and value matrices need special treatment read_in_q_k_v(a ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case : int = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): _snake_case : Optional[int] = state_dict.pop(a ) _snake_case : Any = val # create HuggingFace model and load state dict _snake_case : Tuple = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case : Any = 15 _snake_case : int = 2 _snake_case : Optional[Any] = {0: "table", 1: "table rotated"} _snake_case : Union[str, Any] = idalabel _snake_case : Union[str, Any] = {v: k for k, v in idalabel.items()} else: _snake_case : Any = 125 _snake_case : Union[str, Any] = 6 _snake_case : List[str] = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } _snake_case : Any = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} _snake_case : Union[str, Any] = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1_000 ) _snake_case : str = TableTransformerForObjectDetection(a ) model.load_state_dict(a ) model.eval() # verify our conversion _snake_case : Optional[int] = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" _snake_case : Optional[Any] = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=a ) _snake_case : Dict = Image.open(a ).convert("RGB" ) _snake_case : Union[str, Any] = normalize(resize(a , a ) ).unsqueeze(0 ) _snake_case : str = model(a ) if "detection" in checkpoint_url: _snake_case : int = (1, 15, 3) _snake_case : List[str] = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) _snake_case : List[str] = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: _snake_case : Union[str, Any] = (1, 125, 7) _snake_case : str = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) _snake_case : Optional[Any] = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , a , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , a , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) _snake_case : int = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(a ) image_processor.push_to_hub(a ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _a : Any = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
87
0
"""simple docstring""" import math from collections.abc import Callable def a__ ( a : Callable[[float], float] , a : float , a : float ): """simple docstring""" _snake_case : float = xa _snake_case : float = xa while True: if x_n == x_na or function(a ) == function(a ): raise ZeroDivisionError("float division by zero, could not find root" ) _snake_case : float = x_na - ( function(a ) / ((function(a ) - function(a )) / (x_na - x_n)) ) if abs(x_na - x_na ) < 10**-5: return x_na _snake_case : Optional[Any] = x_na _snake_case : List[Any] = x_na def a__ ( a : float ): """simple docstring""" return math.pow(a , 3 ) - (2 * x) - 5 if __name__ == "__main__": print(intersection(f, 3, 3.5))
711
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
87
0
import unittest import numpy as np from transformers.testing_utils import require_torch, require_vision from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ViTImageProcessor class _UpperCAmelCase ( unittest.TestCase): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=3 , snake_case_=2_24 , snake_case_=30 , snake_case_=4_00 , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=[0.5, 0.5, 0.5] , snake_case_=[0.5, 0.5, 0.5] , ): _snake_case : List[str] = size if size is not None else {"height": 18, "width": 18} _snake_case : str = parent _snake_case : str = batch_size _snake_case : Dict = num_channels _snake_case : Optional[int] = image_size _snake_case : List[Any] = min_resolution _snake_case : int = max_resolution _snake_case : Any = do_resize _snake_case : Dict = size _snake_case : Dict = do_normalize _snake_case : List[str] = image_mean _snake_case : Optional[int] = image_std def lowerCamelCase__ ( self ): return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Tuple = ViTImageProcessor if is_vision_available() else None def lowerCamelCase__ ( self ): _snake_case : List[Any] = EfficientFormerImageProcessorTester(self ) @property def lowerCamelCase__ ( self ): return self.image_proc_tester.prepare_image_processor_dict() def lowerCamelCase__ ( self ): _snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , "image_mean" ) ) self.assertTrue(hasattr(snake_case_ , "image_std" ) ) self.assertTrue(hasattr(snake_case_ , "do_normalize" ) ) self.assertTrue(hasattr(snake_case_ , "do_resize" ) ) self.assertTrue(hasattr(snake_case_ , "size" ) ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): # Initialize image_processor _snake_case : Any = self.image_processing_class(**self.image_processor_dict ) # create random PIL images _snake_case : Tuple = prepare_image_inputs(self.image_proc_tester , equal_resolution=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , Image.Image ) # Test not batched input _snake_case : Dict = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _snake_case : Optional[int] = image_processor(snake_case_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def lowerCamelCase__ ( self ): # Initialize image_processor _snake_case : Union[str, Any] = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors _snake_case : Union[str, Any] = prepare_image_inputs(self.image_proc_tester , equal_resolution=snake_case_ , numpify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , np.ndarray ) # Test not batched input _snake_case : Dict = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _snake_case : Tuple = image_processor(snake_case_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) def lowerCamelCase__ ( self ): # Initialize image_processor _snake_case : List[str] = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors _snake_case : Optional[int] = prepare_image_inputs(self.image_proc_tester , equal_resolution=snake_case_ , torchify=snake_case_ ) for image in image_inputs: self.assertIsInstance(snake_case_ , torch.Tensor ) # Test not batched input _snake_case : Optional[int] = image_processor(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , ) # Test batched _snake_case : Dict = image_processor(snake_case_ , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_proc_tester.batch_size, self.image_proc_tester.num_channels, self.image_proc_tester.size["height"], self.image_proc_tester.size["width"], ) , )
712
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _UpperCAmelCase ( _snake_case , _snake_case): __lowercase : List[Any] = """convnextv2""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=4 , snake_case_=None , snake_case_=None , snake_case_="gelu" , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=0.0 , snake_case_=2_24 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Tuple = num_channels _snake_case : Optional[int] = patch_size _snake_case : Tuple = num_stages _snake_case : int = [96, 1_92, 3_84, 7_68] if hidden_sizes is None else hidden_sizes _snake_case : str = [3, 3, 9, 3] if depths is None else depths _snake_case : int = hidden_act _snake_case : Tuple = initializer_range _snake_case : Union[str, Any] = layer_norm_eps _snake_case : Optional[int] = drop_path_rate _snake_case : Union[str, Any] = image_size _snake_case : List[Any] = ["stem"] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] _snake_case , _snake_case : Dict = get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names )
87
0
"""simple docstring""" from itertools import count def a__ ( a : int = 50 ): """simple docstring""" _snake_case : Dict = [1] * min_block_length for n in count(a ): fill_count_functions.append(1 ) for block_length in range(a , n + 1 ): for block_start in range(n - block_length ): fill_count_functions[n] += fill_count_functions[ n - block_start - block_length - 1 ] fill_count_functions[n] += 1 if fill_count_functions[n] > 1_000_000: break return n if __name__ == "__main__": print(f'{solution() = }')
713
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def a__ ( a : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) _a : int = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : Dict = parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=snake_case_ , required=snake_case_ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=snake_case_ , required=snake_case_ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=snake_case_ , required=snake_case_ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=snake_case_ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=snake_case_ , default=snake_case_ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : str = logging.get_logger("transformers-cli/converting" ) self._logger.info(F'Loading model {model_type}' ) _snake_case : Optional[int] = model_type _snake_case : Any = tf_checkpoint _snake_case : Optional[int] = pytorch_dump_output _snake_case : Tuple = config _snake_case : Tuple = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) if "ckpt" in self._tf_checkpoint.lower(): _snake_case : int = self._tf_checkpoint _snake_case : Optional[Any] = "" else: _snake_case : Optional[int] = self._tf_checkpoint _snake_case : List[str] = "" convert_transfo_xl_checkpoint_to_pytorch( snake_case_ , self._config , self._pytorch_dump_output , snake_case_ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
87
0
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available _a : str = { """configuration_m2m_100""": ["""M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP""", """M2M100Config""", """M2M100OnnxConfig"""], """tokenization_m2m_100""": ["""M2M100Tokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : int = [ """M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST""", """M2M100ForConditionalGeneration""", """M2M100Model""", """M2M100PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_mam_aaa import M2M_100_PRETRAINED_CONFIG_ARCHIVE_MAP, MaMaaaConfig, MaMaaaOnnxConfig from .tokenization_mam_aaa import MaMaaaTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_mam_aaa import ( M2M_100_PRETRAINED_MODEL_ARCHIVE_LIST, MaMaaaForConditionalGeneration, MaMaaaModel, MaMaaaPreTrainedModel, ) else: import sys _a : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
714
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def a__ ( a : List[str] , a : Any ): """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _snake_case : Any = flax_key_tuple[:-1] + ("weight",) _snake_case : str = torch.permute(a , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(a ): # linear layer _snake_case : Optional[int] = flax_key_tuple[:-1] + ("weight",) _snake_case : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _snake_case : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) return flax_key_tuple, flax_tensor def a__ ( a : List[Any] , a : Union[str, Any] , a : List[str] ): """simple docstring""" if "metadata" in layer: _snake_case : Optional[int] = layer.split("metadata" ) _snake_case : Optional[int] = "".join(split_layer[0] )[:-1] _snake_case : int = [tuple(("metadata" + split_layer[1]).split("/" ) )] elif "kvstore" in layer: _snake_case : Any = layer.split("kvstore" ) _snake_case : str = "".join(split_layer[0] )[:-1] _snake_case : Any = [tuple(("kvstore" + split_layer[1]).split("/" ) )] else: _snake_case : List[Any] = layer.split("/" ) _snake_case : Tuple = "/".join(split_layer[:-1] ) _snake_case : int = (split_layer[-1],) if "kvstore/path" in layer: _snake_case : Optional[Any] = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _snake_case : Tuple = "file" else: _snake_case : Optional[int] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def a__ ( a : List[Any] , a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = rename_keys(a ) _snake_case : int = {} for k, v in current_block.items(): _snake_case : Optional[int] = v _snake_case : Optional[int] = new_current_block torch.save(a , a ) def a__ ( a : Dict , a : Tuple , a : List[str] , a : int , a : str = WEIGHTS_NAME ): """simple docstring""" _snake_case : Any = convert_file_size_to_int(a ) _snake_case : Tuple = [] _snake_case : Optional[int] = {} _snake_case : Tuple = 0 _snake_case : Optional[Any] = 0 os.makedirs(a , exist_ok=a ) with gfile.GFile(switch_checkpoint_path + "/checkpoint" , "rb" ) as fp: _snake_case : Any = serialization.msgpack_restore(fp.read() )["optimizer"]["target"] _snake_case : Optional[Any] = flatten_dict(a , sep="/" ) _snake_case : Optional[Any] = {} for layer in checkpoint_info.keys(): _snake_case , _snake_case , _snake_case : int = get_key_and_tensorstore_dict( a , a , a ) if curr_real_layer_name in all_layers: _snake_case : Dict = content else: _snake_case : Tuple = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _snake_case : List[str] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _snake_case : Dict = torch.tensor(a ) _snake_case : Dict = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _snake_case , _snake_case : Optional[int] = rename_base_flax_keys(tuple(key.split("/" ) ) , a ) _snake_case : Optional[Any] = "/".join(a ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _snake_case : Any = os.path.join( a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) del current_block _snake_case : List[Any] = {} _snake_case : str = 0 _snake_case : List[str] = raw_weights.to(getattr(a , a ) ) current_block_size += weight_size total_size += weight_size # Add the last block _snake_case : int = os.path.join(a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(a ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _snake_case : str = {} _snake_case : Any = {} for idx, shard in enumerate(a ): _snake_case : Optional[int] = weights_name.replace( ".bin" , f'-{idx+1:05d}-of-{len(a ):05d}.bin' ) # len(sharded_state_dicts):05d} _snake_case : Dict = os.path.join(a , weights_name.replace(".bin" , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(a , os.path.join(a , a ) ) _snake_case : Dict = shard for key in shard: _snake_case : int = shard_file # Add the metadata _snake_case : List[Any] = {"total_size": total_size} _snake_case : Any = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(a , a ) , "w" , encoding="utf-8" ) as f: _snake_case : Union[str, Any] = json.dumps(a , indent=2 , sort_keys=a ) + "\n" f.write(a ) return metadata, index if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) _a : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def a__ ( ): """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _snake_case : List[str] = SwitchTransformersConfig.from_pretrained("google/switch-base-8" ) config.save_pretrained("/home/arthur_huggingface_co/transformers/switch_converted" ) _snake_case : str = SwitchTransformersForConditionalGeneration.from_pretrained( "/home/arthur_huggingface_co/transformers/switch_converted" , device_map="auto" ) _snake_case : List[Any] = TaTokenizer.from_pretrained("t5-small" ) _snake_case : Optional[Any] = "A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>." _snake_case : Dict = tokenizer(a , return_tensors="pt" ).input_ids _snake_case : List[Any] = model.generate(a , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
87
0
"""simple docstring""" import os from math import logaa def a__ ( a : str = "base_exp.txt" ): """simple docstring""" _snake_case : float = 0 _snake_case : int = 0 for i, line in enumerate(open(os.path.join(os.path.dirname(a ) , a ) ) ): _snake_case : List[str] = list(map(a , line.split("," ) ) ) if x * logaa(a ) > largest: _snake_case : List[Any] = x * logaa(a ) _snake_case : Union[str, Any] = i + 1 return result if __name__ == "__main__": print(solution())
715
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available _a : str = { """configuration_swinv2""": ["""SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Swinv2Config"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = [ """SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST""", """Swinv2ForImageClassification""", """Swinv2ForMaskedImageModeling""", """Swinv2Model""", """Swinv2PreTrainedModel""", ] if TYPE_CHECKING: from .configuration_swinva import SWINV2_PRETRAINED_CONFIG_ARCHIVE_MAP, SwinvaConfig try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_swinva import ( SWINV2_PRETRAINED_MODEL_ARCHIVE_LIST, SwinvaForImageClassification, SwinvaForMaskedImageModeling, SwinvaModel, SwinvaPreTrainedModel, ) else: import sys _a : int = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
716
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _UpperCAmelCase ( _snake_case , _snake_case): __lowercase : List[Any] = """convnextv2""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=4 , snake_case_=None , snake_case_=None , snake_case_="gelu" , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=0.0 , snake_case_=2_24 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Tuple = num_channels _snake_case : Optional[int] = patch_size _snake_case : Tuple = num_stages _snake_case : int = [96, 1_92, 3_84, 7_68] if hidden_sizes is None else hidden_sizes _snake_case : str = [3, 3, 9, 3] if depths is None else depths _snake_case : int = hidden_act _snake_case : Tuple = initializer_range _snake_case : Union[str, Any] = layer_norm_eps _snake_case : Optional[int] = drop_path_rate _snake_case : Union[str, Any] = image_size _snake_case : List[Any] = ["stem"] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] _snake_case : Dict = get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names )
717
"""simple docstring""" def a__ ( a : list , a : int , a : int = 0 , a : int = 0 ): """simple docstring""" _snake_case : Optional[int] = right or len(a ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(a , a , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import argparse import torch from transformers import ( EncodecConfig, EncodecFeatureExtractor, EncodecModel, logging, ) # checkpoints downloaded from: # https://dl.fbaipublicfiles.com/encodec/v0/encodec_24khz-d7cc33bc.th # https://huggingface.co/facebook/musicgen-small/resolve/main/compression_state_dict.bin # https://dl.fbaipublicfiles.com/encodec/v0/encodec_48khz-7e698e3e.th logging.set_verbosity_info() _a : Dict = logging.get_logger("""transformers.models.encodec""") _a : Optional[Any] = { """quantizer.vq.layers.*._codebook.inited""": """quantizer.layers.*.codebook.inited""", """quantizer.vq.layers.*._codebook.cluster_size""": """quantizer.layers.*.codebook.cluster_size""", """quantizer.vq.layers.*._codebook.embed""": """quantizer.layers.*.codebook.embed""", """quantizer.vq.layers.*._codebook.embed_avg""": """quantizer.layers.*.codebook.embed_avg""", } _a : Tuple = { """encoder.model.0.conv.conv""": """encoder.layers.0.conv""", """encoder.model.1.block.1.conv.conv""": """encoder.layers.1.block.1.conv""", """encoder.model.1.block.3.conv.conv""": """encoder.layers.1.block.3.conv""", """encoder.model.1.shortcut.conv.conv""": """encoder.layers.1.shortcut.conv""", """encoder.model.3.conv.conv""": """encoder.layers.3.conv""", """encoder.model.4.block.1.conv.conv""": """encoder.layers.4.block.1.conv""", """encoder.model.4.block.3.conv.conv""": """encoder.layers.4.block.3.conv""", """encoder.model.4.shortcut.conv.conv""": """encoder.layers.4.shortcut.conv""", """encoder.model.6.conv.conv""": """encoder.layers.6.conv""", """encoder.model.7.block.1.conv.conv""": """encoder.layers.7.block.1.conv""", """encoder.model.7.block.3.conv.conv""": """encoder.layers.7.block.3.conv""", """encoder.model.7.shortcut.conv.conv""": """encoder.layers.7.shortcut.conv""", """encoder.model.9.conv.conv""": """encoder.layers.9.conv""", """encoder.model.10.block.1.conv.conv""": """encoder.layers.10.block.1.conv""", """encoder.model.10.block.3.conv.conv""": """encoder.layers.10.block.3.conv""", """encoder.model.10.shortcut.conv.conv""": """encoder.layers.10.shortcut.conv""", """encoder.model.12.conv.conv""": """encoder.layers.12.conv""", """encoder.model.13.lstm""": """encoder.layers.13.lstm""", """encoder.model.15.conv.conv""": """encoder.layers.15.conv""", } _a : Dict = { """encoder.model.0.conv.norm""": """encoder.layers.0.norm""", """encoder.model.1.block.1.conv.norm""": """encoder.layers.1.block.1.norm""", """encoder.model.1.block.3.conv.norm""": """encoder.layers.1.block.3.norm""", """encoder.model.1.shortcut.conv.norm""": """encoder.layers.1.shortcut.norm""", """encoder.model.3.conv.norm""": """encoder.layers.3.norm""", """encoder.model.4.block.1.conv.norm""": """encoder.layers.4.block.1.norm""", """encoder.model.4.block.3.conv.norm""": """encoder.layers.4.block.3.norm""", """encoder.model.4.shortcut.conv.norm""": """encoder.layers.4.shortcut.norm""", """encoder.model.6.conv.norm""": """encoder.layers.6.norm""", """encoder.model.7.block.1.conv.norm""": """encoder.layers.7.block.1.norm""", """encoder.model.7.block.3.conv.norm""": """encoder.layers.7.block.3.norm""", """encoder.model.7.shortcut.conv.norm""": """encoder.layers.7.shortcut.norm""", """encoder.model.9.conv.norm""": """encoder.layers.9.norm""", """encoder.model.10.block.1.conv.norm""": """encoder.layers.10.block.1.norm""", """encoder.model.10.block.3.conv.norm""": """encoder.layers.10.block.3.norm""", """encoder.model.10.shortcut.conv.norm""": """encoder.layers.10.shortcut.norm""", """encoder.model.12.conv.norm""": """encoder.layers.12.norm""", """encoder.model.15.conv.norm""": """encoder.layers.15.norm""", } _a : Tuple = { """decoder.model.0.conv.conv""": """decoder.layers.0.conv""", """decoder.model.1.lstm""": """decoder.layers.1.lstm""", """decoder.model.3.convtr.convtr""": """decoder.layers.3.conv""", """decoder.model.4.block.1.conv.conv""": """decoder.layers.4.block.1.conv""", """decoder.model.4.block.3.conv.conv""": """decoder.layers.4.block.3.conv""", """decoder.model.4.shortcut.conv.conv""": """decoder.layers.4.shortcut.conv""", """decoder.model.6.convtr.convtr""": """decoder.layers.6.conv""", """decoder.model.7.block.1.conv.conv""": """decoder.layers.7.block.1.conv""", """decoder.model.7.block.3.conv.conv""": """decoder.layers.7.block.3.conv""", """decoder.model.7.shortcut.conv.conv""": """decoder.layers.7.shortcut.conv""", """decoder.model.9.convtr.convtr""": """decoder.layers.9.conv""", """decoder.model.10.block.1.conv.conv""": """decoder.layers.10.block.1.conv""", """decoder.model.10.block.3.conv.conv""": """decoder.layers.10.block.3.conv""", """decoder.model.10.shortcut.conv.conv""": """decoder.layers.10.shortcut.conv""", """decoder.model.12.convtr.convtr""": """decoder.layers.12.conv""", """decoder.model.13.block.1.conv.conv""": """decoder.layers.13.block.1.conv""", """decoder.model.13.block.3.conv.conv""": """decoder.layers.13.block.3.conv""", """decoder.model.13.shortcut.conv.conv""": """decoder.layers.13.shortcut.conv""", """decoder.model.15.conv.conv""": """decoder.layers.15.conv""", } _a : List[Any] = { """decoder.model.0.conv.norm""": """decoder.layers.0.norm""", """decoder.model.3.convtr.norm""": """decoder.layers.3.norm""", """decoder.model.4.block.1.conv.norm""": """decoder.layers.4.block.1.norm""", """decoder.model.4.block.3.conv.norm""": """decoder.layers.4.block.3.norm""", """decoder.model.4.shortcut.conv.norm""": """decoder.layers.4.shortcut.norm""", """decoder.model.6.convtr.norm""": """decoder.layers.6.norm""", """decoder.model.7.block.1.conv.norm""": """decoder.layers.7.block.1.norm""", """decoder.model.7.block.3.conv.norm""": """decoder.layers.7.block.3.norm""", """decoder.model.7.shortcut.conv.norm""": """decoder.layers.7.shortcut.norm""", """decoder.model.9.convtr.norm""": """decoder.layers.9.norm""", """decoder.model.10.block.1.conv.norm""": """decoder.layers.10.block.1.norm""", """decoder.model.10.block.3.conv.norm""": """decoder.layers.10.block.3.norm""", """decoder.model.10.shortcut.conv.norm""": """decoder.layers.10.shortcut.norm""", """decoder.model.12.convtr.norm""": """decoder.layers.12.norm""", """decoder.model.13.block.1.conv.norm""": """decoder.layers.13.block.1.norm""", """decoder.model.13.block.3.conv.norm""": """decoder.layers.13.block.3.norm""", """decoder.model.13.shortcut.conv.norm""": """decoder.layers.13.shortcut.norm""", """decoder.model.15.conv.norm""": """decoder.layers.15.norm""", } _a : List[Any] = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_DECODER, } _a : Dict = { **MAPPING_QUANTIZER, **MAPPING_ENCODER, **MAPPING_ENCODER_48K, **MAPPING_DECODER, **MAPPING_DECODER_48K, } _a : Any = [] _a : Dict = [] def a__ ( a : Union[str, Any] , a : Dict , a : Tuple , a : List[Any] , a : Optional[int] ): """simple docstring""" for attribute in key.split("." ): _snake_case : int = getattr(a , a ) if weight_type is not None: _snake_case : Any = getattr(a , a ).shape else: _snake_case : List[str] = hf_pointer.shape if hf_shape != value.shape: raise ValueError( f'Shape of hf {key + "." + weight_type if weight_type is not None else ""} is {hf_shape}, but should be' f' {value.shape} for {full_name}' ) if weight_type == "weight": _snake_case : Any = value elif weight_type == "weight_g": _snake_case : str = value elif weight_type == "weight_v": _snake_case : List[Any] = value elif weight_type == "bias": _snake_case : int = value elif weight_type == "running_mean": _snake_case : Any = value elif weight_type == "running_var": _snake_case : str = value elif weight_type == "num_batches_tracked": _snake_case : int = value elif weight_type == "weight_ih_l0": _snake_case : int = value elif weight_type == "weight_hh_l0": _snake_case : Optional[Any] = value elif weight_type == "bias_ih_l0": _snake_case : Optional[Any] = value elif weight_type == "bias_hh_l0": _snake_case : List[str] = value elif weight_type == "weight_ih_l1": _snake_case : Tuple = value elif weight_type == "weight_hh_l1": _snake_case : Tuple = value elif weight_type == "bias_ih_l1": _snake_case : Optional[Any] = value elif weight_type == "bias_hh_l1": _snake_case : Dict = value else: _snake_case : str = value logger.info(f'{key + ("." + weight_type if weight_type is not None else "")} was initialized from {full_name}.' ) def a__ ( a : int , a : str ): """simple docstring""" for key in ignore_keys: if key.endswith(".*" ): if name.startswith(key[:-1] ): return True elif ".*." in key: _snake_case : List[str] = key.split(".*." ) if prefix in name and suffix in name: return True elif key in name: return True return False def a__ ( a : str , a : Union[str, Any] , a : Any ): """simple docstring""" _snake_case : int = [] if model_name == "encodec_24khz" or "encodec_32khz": _snake_case : Optional[Any] = MAPPING_24K elif model_name == "encodec_48khz": _snake_case : str = MAPPING_48K else: raise ValueError(f'Unsupported model: {model_name}' ) for name, value in orig_dict.items(): if should_ignore(a , a ): logger.info(f'{name} was ignored' ) continue _snake_case : Tuple = False for key, mapped_key in MAPPING.items(): if "*" in key: _snake_case : List[str] = key.split(".*." ) if prefix in name and suffix in name: _snake_case : str = suffix if key in name: # HACK otherwise .embed gets initialized with .embed_avg too if key.endswith("embed" ) and name.endswith("embed_avg" ): continue _snake_case : Optional[int] = True if "*" in mapped_key: _snake_case : int = name.split(a )[0].split("." )[-2] _snake_case : Optional[int] = mapped_key.replace("*" , a ) if "weight_g" in name: _snake_case : Optional[int] = "weight_g" elif "weight_v" in name: _snake_case : Any = "weight_v" elif "weight_ih_l0" in name: _snake_case : Dict = "weight_ih_l0" elif "weight_hh_l0" in name: _snake_case : Dict = "weight_hh_l0" elif "bias_ih_l0" in name: _snake_case : Any = "bias_ih_l0" elif "bias_hh_l0" in name: _snake_case : Union[str, Any] = "bias_hh_l0" elif "weight_ih_l1" in name: _snake_case : Optional[Any] = "weight_ih_l1" elif "weight_hh_l1" in name: _snake_case : str = "weight_hh_l1" elif "bias_ih_l1" in name: _snake_case : Optional[Any] = "bias_ih_l1" elif "bias_hh_l1" in name: _snake_case : int = "bias_hh_l1" elif "bias" in name: _snake_case : Union[str, Any] = "bias" elif "weight" in name: _snake_case : Optional[int] = "weight" elif "running_mean" in name: _snake_case : List[str] = "running_mean" elif "running_var" in name: _snake_case : int = "running_var" elif "num_batches_tracked" in name: _snake_case : Union[str, Any] = "num_batches_tracked" else: _snake_case : Union[str, Any] = None set_recursively(a , a , a , a , a ) continue if not is_used: unused_weights.append(a ) logger.warning(f'Unused weights: {unused_weights}' ) @torch.no_grad() def a__ ( a : Optional[Any] , a : Tuple , a : Union[str, Any] , a : List[Any]=None , a : Optional[int]=None , ): """simple docstring""" if config_path is not None: _snake_case : int = EncodecConfig.from_pretrained(a ) else: _snake_case : Optional[Any] = EncodecConfig() if model_name == "encodec_24khz": pass # config is already correct elif model_name == "encodec_32khz": _snake_case : Any = [8, 5, 4, 4] _snake_case : Union[str, Any] = [2.2] _snake_case : Union[str, Any] = 64 _snake_case : Optional[int] = 32_000 _snake_case : int = 2_048 _snake_case : str = False _snake_case : List[str] = False _snake_case : List[str] = False elif model_name == "encodec_48khz": _snake_case : Any = [8, 5, 4, 2] _snake_case : List[Any] = [3.0, 6.0, 12.0, 24.0] _snake_case : Optional[int] = 48_000 _snake_case : Any = 2 _snake_case : Tuple = False _snake_case : int = "time_group_norm" _snake_case : Union[str, Any] = True _snake_case : Any = 1.0 _snake_case : Dict = 0.01 else: raise ValueError(f'Unknown model name: {model_name}' ) _snake_case : Optional[int] = EncodecModel(a ) _snake_case : Union[str, Any] = EncodecFeatureExtractor( feature_size=config.audio_channels , sampling_rate=config.sampling_rate , chunk_length_s=config.chunk_length_s , overlap=config.overlap , ) feature_extractor.save_pretrained(a ) _snake_case : Union[str, Any] = torch.load(a ) if "best_state" in original_checkpoint: # we might have a training state saved, in which case discard the yaml results and just retain the weights _snake_case : int = original_checkpoint["best_state"] recursively_load_weights(a , a , a ) model.save_pretrained(a ) if repo_id: print("Pushing to the hub..." ) feature_extractor.push_to_hub(a ) model.push_to_hub(a ) if __name__ == "__main__": _a : List[Any] = argparse.ArgumentParser() parser.add_argument( """--model""", default="""encodec_24khz""", type=str, help="""The model to convert. Should be one of 'encodec_24khz', 'encodec_32khz', 'encodec_48khz'.""", ) parser.add_argument("""--checkpoint_path""", required=True, default=None, type=str, help="""Path to original checkpoint""") parser.add_argument("""--config_path""", default=None, type=str, help="""Path to hf config.json of model to convert""") parser.add_argument( """--pytorch_dump_folder_path""", required=True, default=None, type=str, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", default=None, type=str, help="""Where to upload the converted model on the 🤗 hub.""" ) _a : Tuple = parser.parse_args() convert_checkpoint( args.model, args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path, args.push_to_hub, )
718
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ ): _snake_case , _snake_case : Dict = text, pattern _snake_case , _snake_case : int = len(snake_case_ ), len(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCamelCase__ ( self ): # searches pattern in text and returns index positions _snake_case : List[str] = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case : Union[str, Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: _snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) _snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _a : List[Any] = """ABAABA""" _a : str = """AB""" _a : List[Any] = BoyerMooreSearch(text, pattern) _a : Any = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
87
0
"""simple docstring""" import json import multiprocessing as mp import re from collections import defaultdict from functools import partial from typing import Dict, List, Optional, Set, Tuple, Type from datasets import Dataset from datasketch import MinHash, MinHashLSH from dpu_utils.utils.iterators import ThreadedIterator from tqdm import tqdm _a : str = re.compile("""[^A-Za-z_0-9]""") # parameters used in DuplicationIndex _a : Any = 10 _a : str = 256 def a__ ( a : List[str] ): """simple docstring""" if len(a ) < MIN_NUM_TOKENS: return None _snake_case : Tuple = MinHash(num_perm=a ) for token in set(a ): min_hash.update(token.encode() ) return min_hash def a__ ( a : str ): """simple docstring""" return {t for t in NON_ALPHA.split(a ) if len(t.strip() ) > 0} class _UpperCAmelCase : def __init__( self , *, snake_case_ = 0.85 , ): _snake_case : Optional[Any] = duplication_jaccard_threshold _snake_case : List[Any] = NUM_PERM _snake_case : List[Any] = MinHashLSH(threshold=self._duplication_jaccard_threshold , num_perm=self._num_perm ) _snake_case : str = defaultdict(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ ): _snake_case : Optional[int] = self._index.query(snake_case_ ) if code_key in self._index.keys: print(F'Duplicate key {code_key}' ) return self._index.insert(snake_case_ , snake_case_ ) if len(snake_case_ ) > 0: for base_duplicate in close_duplicates: if base_duplicate in self._duplicate_clusters: self._duplicate_clusters[base_duplicate].add(snake_case_ ) break else: self._duplicate_clusters[close_duplicates[0]].add(snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = [] for base, duplicates in self._duplicate_clusters.items(): _snake_case : Tuple = [base] + list(snake_case_ ) # reformat the cluster to be a list of dict _snake_case : Optional[int] = [{"base_index": el[0], "repo_name": el[1], "path": el[2]} for el in cluster] duplicate_clusters.append(snake_case_ ) return duplicate_clusters def lowerCamelCase__ ( self , snake_case_ ): _snake_case : int = self.get_duplicate_clusters() with open(snake_case_ , "w" ) as f: json.dump(snake_case_ , snake_case_ ) def a__ ( a : str ): """simple docstring""" _snake_case : Union[str, Any] = element _snake_case : Union[str, Any] = get_min_hash([t for t in NON_ALPHA.split(data["content"] ) if len(t.strip() ) > 0] ) if min_hash is not None: return (index, data["repo_name"], data["path"]), min_hash def a__ ( a : Type[Dataset] ): """simple docstring""" with mp.Pool() as pool: for data in pool.imap_unordered( _compute_min_hash , ThreadedIterator(a , max_queue_size=10_000 ) , chunksize=100 , ): if data is not None: yield data def a__ ( a : Type[Dataset] , a : float ): """simple docstring""" _snake_case : Union[str, Any] = DuplicationIndex(duplication_jaccard_threshold=a ) for filename, min_hash in tqdm(ThreadedIterator(minhash_iter(enumerate(a ) ) , max_queue_size=100 ) ): di.add(a , a ) # Returns a List[Cluster] where Cluster is List[str] with the filenames. return di.get_duplicate_clusters() def a__ ( a : str , a : str ): """simple docstring""" _snake_case : Tuple = get_tokens(a ) _snake_case : List[Any] = get_tokens(a ) return len(tokensa & tokensa ) / len(tokensa | tokensa ) _a : List[Any] = None def a__ ( a : Optional[int] , a : Union[str, Any] ): """simple docstring""" _snake_case : Optional[Any] = [] for elementa in cluster: _snake_case : Any = _shared_dataset[elementa["base_index"]]["content"] for elementa in extremes: _snake_case : Dict = _shared_dataset[elementa["base_index"]]["content"] if jaccard_similarity(a , a ) >= jaccard_threshold: elementa["copies"] += 1 break else: _snake_case : str = 1 extremes.append(a ) return extremes def a__ ( a : Union[str, Any] , a : Any , a : Optional[int] ): """simple docstring""" global _shared_dataset _snake_case : int = dataset _snake_case : Dict = [] _snake_case : Union[str, Any] = partial(_find_cluster_extremes_shared , jaccard_threshold=a ) with mp.Pool() as pool: for extremes in tqdm( pool.imap_unordered( a , a , ) , total=len(a ) , ): extremes_list.append(a ) return extremes_list def a__ ( a : Type[Dataset] , a : float = 0.85 ): """simple docstring""" _snake_case : Tuple = make_duplicate_clusters(a , a ) _snake_case : Tuple = {x["base_index"] for cluster in duplicate_clusters for x in cluster} _snake_case : Union[str, Any] = {} _snake_case : Dict = find_extremes(a , a , a ) for extremes in extremes_clusters: for element in extremes: _snake_case : Optional[int] = element _snake_case : List[Any] = duplicate_indices - set(extreme_dict.keys() ) _snake_case : Tuple = dataset.filter(lambda a , a : idx not in remove_indices , with_indices=a ) # update duplicate_clusters for cluster in duplicate_clusters: for element in cluster: _snake_case : Tuple = element["base_index"] in extreme_dict if element["is_extreme"]: _snake_case : Any = extreme_dict[element["base_index"]]["copies"] print(f'Original dataset size: {len(a )}' ) print(f'Number of duplicate clusters: {len(a )}' ) print(f'Files in duplicate cluster: {len(a )}' ) print(f'Unique files in duplicate cluster: {len(a )}' ) print(f'Filtered dataset size: {len(a )}' ) return ds_filter, duplicate_clusters
719
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...file_utils import _LazyModule, is_torch_available from ...utils import OptionalDependencyNotAvailable _a : Tuple = { """configuration_gpt_neox_japanese""": ["""GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP""", """GPTNeoXJapaneseConfig"""], """tokenization_gpt_neox_japanese""": ["""GPTNeoXJapaneseTokenizer"""], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[int] = [ """GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST""", """GPTNeoXJapaneseForCausalLM""", """GPTNeoXJapaneseLayer""", """GPTNeoXJapaneseModel""", """GPTNeoXJapanesePreTrainedModel""", ] if TYPE_CHECKING: from .configuration_gpt_neox_japanese import GPT_NEOX_JAPANESE_PRETRAINED_CONFIG_ARCHIVE_MAP, GPTNeoXJapaneseConfig from .tokenization_gpt_neox_japanese import GPTNeoXJapaneseTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_gpt_neox_japanese import ( GPT_NEOX_JAPANESE_PRETRAINED_MODEL_ARCHIVE_LIST, GPTNeoXJapaneseForCausalLM, GPTNeoXJapaneseLayer, GPTNeoXJapaneseModel, GPTNeoXJapanesePreTrainedModel, ) else: import sys _a : Any = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
720
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from typing import Optional, Tuple, Union import flax import flax.linen as nn import jax import jax.numpy as jnp from flax.core.frozen_dict import FrozenDict from ..configuration_utils import ConfigMixin, flax_register_to_config from ..utils import BaseOutput from .embeddings_flax import FlaxTimestepEmbedding, FlaxTimesteps from .modeling_flax_utils import FlaxModelMixin from .unet_ad_blocks_flax import ( FlaxCrossAttnDownBlockaD, FlaxDownBlockaD, FlaxUNetMidBlockaDCrossAttn, ) @flax.struct.dataclass class _UpperCAmelCase ( _snake_case): __lowercase : jnp.ndarray __lowercase : jnp.ndarray class _UpperCAmelCase ( nn.Module): __lowercase : int __lowercase : Tuple[int] = (1_6, 3_2, 9_6, 2_5_6) __lowercase : jnp.dtype = jnp.floataa def lowerCamelCase__ ( self ): _snake_case : Any = nn.Conv( self.block_out_channels[0] , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) _snake_case : Any = [] for i in range(len(self.block_out_channels ) - 1 ): _snake_case : Dict = self.block_out_channels[i] _snake_case : int = self.block_out_channels[i + 1] _snake_case : str = nn.Conv( snake_case_ , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(snake_case_ ) _snake_case : Any = nn.Conv( snake_case_ , kernel_size=(3, 3) , strides=(2, 2) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) blocks.append(snake_case_ ) _snake_case : Any = blocks _snake_case : Dict = nn.Conv( self.conditioning_embedding_channels , kernel_size=(3, 3) , padding=((1, 1), (1, 1)) , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , snake_case_ ): _snake_case : List[Any] = self.conv_in(snake_case_ ) _snake_case : List[Any] = nn.silu(snake_case_ ) for block in self.blocks: _snake_case : Tuple = block(snake_case_ ) _snake_case : str = nn.silu(snake_case_ ) _snake_case : Dict = self.conv_out(snake_case_ ) return embedding @flax_register_to_config class _UpperCAmelCase ( nn.Module , _snake_case , _snake_case): __lowercase : int = 3_2 __lowercase : int = 4 __lowercase : Tuple[str] = ( "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "CrossAttnDownBlock2D", "DownBlock2D", ) __lowercase : Union[bool, Tuple[bool]] = False __lowercase : Tuple[int] = (3_2_0, 6_4_0, 1_2_8_0, 1_2_8_0) __lowercase : int = 2 __lowercase : Union[int, Tuple[int]] = 8 __lowercase : Optional[Union[int, Tuple[int]]] = None __lowercase : int = 1_2_8_0 __lowercase : float = 0.0 __lowercase : bool = False __lowercase : jnp.dtype = jnp.floataa __lowercase : bool = True __lowercase : int = 0 __lowercase : str = "rgb" __lowercase : Tuple[int] = (1_6, 3_2, 9_6, 2_5_6) def lowerCamelCase__ ( self , snake_case_ ): # init input tensors _snake_case : int = (1, self.in_channels, self.sample_size, self.sample_size) _snake_case : Dict = jnp.zeros(snake_case_ , dtype=jnp.floataa ) _snake_case : Any = jnp.ones((1,) , dtype=jnp.intaa ) _snake_case : Any = jnp.zeros((1, 1, self.cross_attention_dim) , dtype=jnp.floataa ) _snake_case : int = (1, 3, self.sample_size * 8, self.sample_size * 8) _snake_case : Dict = jnp.zeros(snake_case_ , dtype=jnp.floataa ) _snake_case : List[Any] = jax.random.split(snake_case_ ) _snake_case : Any = {"params": params_rng, "dropout": dropout_rng} return self.init(snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ )["params"] def lowerCamelCase__ ( self ): _snake_case : int = self.block_out_channels _snake_case : Tuple = block_out_channels[0] * 4 # If `num_attention_heads` is not defined (which is the case for most models) # it will default to `attention_head_dim`. This looks weird upon first reading it and it is. # The reason for this behavior is to correct for incorrectly named variables that were introduced # when this library was created. The incorrect naming was only discovered much later in https://github.com/huggingface/diffusers/issues/2011#issuecomment-1547958131 # Changing `attention_head_dim` to `num_attention_heads` for 40,000+ configurations is too backwards breaking # which is why we correct for the naming here. _snake_case : int = self.num_attention_heads or self.attention_head_dim # input _snake_case : Optional[Any] = nn.Conv( block_out_channels[0] , kernel_size=(3, 3) , strides=(1, 1) , padding=((1, 1), (1, 1)) , dtype=self.dtype , ) # time _snake_case : Tuple = FlaxTimesteps( block_out_channels[0] , flip_sin_to_cos=self.flip_sin_to_cos , freq_shift=self.config.freq_shift ) _snake_case : Tuple = FlaxTimestepEmbedding(snake_case_ , dtype=self.dtype ) _snake_case : Any = FlaxControlNetConditioningEmbedding( conditioning_embedding_channels=block_out_channels[0] , block_out_channels=self.conditioning_embedding_out_channels , ) _snake_case : int = self.only_cross_attention if isinstance(snake_case_ , snake_case_ ): _snake_case : Optional[int] = (only_cross_attention,) * len(self.down_block_types ) if isinstance(snake_case_ , snake_case_ ): _snake_case : Optional[int] = (num_attention_heads,) * len(self.down_block_types ) # down _snake_case : Optional[Any] = [] _snake_case : Optional[int] = [] _snake_case : List[Any] = block_out_channels[0] _snake_case : Union[str, Any] = nn.Conv( snake_case_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case_ ) for i, down_block_type in enumerate(self.down_block_types ): _snake_case : Optional[Any] = output_channel _snake_case : Dict = block_out_channels[i] _snake_case : int = i == len(snake_case_ ) - 1 if down_block_type == "CrossAttnDownBlock2D": _snake_case : Tuple = FlaxCrossAttnDownBlockaD( in_channels=snake_case_ , out_channels=snake_case_ , dropout=self.dropout , num_layers=self.layers_per_block , num_attention_heads=num_attention_heads[i] , add_downsample=not is_final_block , use_linear_projection=self.use_linear_projection , only_cross_attention=only_cross_attention[i] , dtype=self.dtype , ) else: _snake_case : Optional[Any] = FlaxDownBlockaD( in_channels=snake_case_ , out_channels=snake_case_ , dropout=self.dropout , num_layers=self.layers_per_block , add_downsample=not is_final_block , dtype=self.dtype , ) down_blocks.append(snake_case_ ) for _ in range(self.layers_per_block ): _snake_case : Tuple = nn.Conv( snake_case_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case_ ) if not is_final_block: _snake_case : Dict = nn.Conv( snake_case_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) controlnet_down_blocks.append(snake_case_ ) _snake_case : int = down_blocks _snake_case : Dict = controlnet_down_blocks # mid _snake_case : Union[str, Any] = block_out_channels[-1] _snake_case : int = FlaxUNetMidBlockaDCrossAttn( in_channels=snake_case_ , dropout=self.dropout , num_attention_heads=num_attention_heads[-1] , use_linear_projection=self.use_linear_projection , dtype=self.dtype , ) _snake_case : Union[str, Any] = nn.Conv( snake_case_ , kernel_size=(1, 1) , padding="VALID" , kernel_init=nn.initializers.zeros_init() , bias_init=nn.initializers.zeros_init() , dtype=self.dtype , ) def __call__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ = 1.0 , snake_case_ = True , snake_case_ = False , ): _snake_case : Union[str, Any] = self.controlnet_conditioning_channel_order if channel_order == "bgr": _snake_case : List[str] = jnp.flip(snake_case_ , axis=1 ) # 1. time if not isinstance(snake_case_ , jnp.ndarray ): _snake_case : int = jnp.array([timesteps] , dtype=jnp.intaa ) elif isinstance(snake_case_ , jnp.ndarray ) and len(timesteps.shape ) == 0: _snake_case : Optional[int] = timesteps.astype(dtype=jnp.floataa ) _snake_case : Union[str, Any] = jnp.expand_dims(snake_case_ , 0 ) _snake_case : Any = self.time_proj(snake_case_ ) _snake_case : Optional[int] = self.time_embedding(snake_case_ ) # 2. pre-process _snake_case : Union[str, Any] = jnp.transpose(snake_case_ , (0, 2, 3, 1) ) _snake_case : str = self.conv_in(snake_case_ ) _snake_case : str = jnp.transpose(snake_case_ , (0, 2, 3, 1) ) _snake_case : Dict = self.controlnet_cond_embedding(snake_case_ ) sample += controlnet_cond # 3. down _snake_case : Any = (sample,) for down_block in self.down_blocks: if isinstance(snake_case_ , snake_case_ ): _snake_case : List[Any] = down_block(snake_case_ , snake_case_ , snake_case_ , deterministic=not train ) else: _snake_case : Dict = down_block(snake_case_ , snake_case_ , deterministic=not train ) down_block_res_samples += res_samples # 4. mid _snake_case : Any = self.mid_block(snake_case_ , snake_case_ , snake_case_ , deterministic=not train ) # 5. contronet blocks _snake_case : Tuple = () for down_block_res_sample, controlnet_block in zip(snake_case_ , self.controlnet_down_blocks ): _snake_case : Dict = controlnet_block(snake_case_ ) controlnet_down_block_res_samples += (down_block_res_sample,) _snake_case : List[str] = controlnet_down_block_res_samples _snake_case : Tuple = self.controlnet_mid_block(snake_case_ ) # 6. scaling _snake_case : int = [sample * conditioning_scale for sample in down_block_res_samples] mid_block_res_sample *= conditioning_scale if not return_dict: return (down_block_res_samples, mid_block_res_sample) return FlaxControlNetOutput( down_block_res_samples=snake_case_ , mid_block_res_sample=snake_case_ )
721
"""simple docstring""" import argparse import json import subprocess def a__ ( a : Optional[Any] , a : Optional[int] ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[Any] = ( f'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _snake_case : Dict = subprocess.run(a , shell=a , stdout=subprocess.PIPE ) _snake_case : Tuple = output.stdout.decode("utf-8" ) _snake_case : List[str] = json.loads(a ) _snake_case : Any = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(a ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(a ) ) if len(a ) > 0: _snake_case : Any = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def a__ ( a : Optional[int] ): """simple docstring""" return values.split("," ) _a : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _a : List[str] = parser.parse_args() get_runner_status(args.target_runners, args.token)
87
0
from ...configuration_utils import PretrainedConfig from ...utils import logging _a : int = logging.get_logger(__name__) _a : str = { """vinvino02/glpn-kitti""": """https://huggingface.co/vinvino02/glpn-kitti/resolve/main/config.json""", # See all GLPN models at https://huggingface.co/models?filter=glpn } class _UpperCAmelCase ( _snake_case): __lowercase : List[str] = """glpn""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=[2, 2, 2, 2] , snake_case_=[8, 4, 2, 1] , snake_case_=[32, 64, 1_60, 2_56] , snake_case_=[7, 3, 3, 3] , snake_case_=[4, 2, 2, 2] , snake_case_=[1, 2, 5, 8] , snake_case_=[4, 4, 4, 4] , snake_case_="gelu" , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.02 , snake_case_=0.1 , snake_case_=1E-6 , snake_case_=64 , snake_case_=10 , snake_case_=-1 , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Union[str, Any] = num_channels _snake_case : List[str] = num_encoder_blocks _snake_case : int = depths _snake_case : Optional[Any] = sr_ratios _snake_case : Any = hidden_sizes _snake_case : Union[str, Any] = patch_sizes _snake_case : Tuple = strides _snake_case : List[Any] = mlp_ratios _snake_case : List[Any] = num_attention_heads _snake_case : int = hidden_act _snake_case : List[Any] = hidden_dropout_prob _snake_case : List[Any] = attention_probs_dropout_prob _snake_case : Any = initializer_range _snake_case : Optional[Any] = drop_path_rate _snake_case : Optional[int] = layer_norm_eps _snake_case : Union[str, Any] = decoder_hidden_size _snake_case : int = max_depth _snake_case : Any = head_in_index
700
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
87
0
from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging _a : Any = logging.get_logger(__name__) _a : Dict = { """google/mobilenet_v1_1.0_224""": """https://huggingface.co/google/mobilenet_v1_1.0_224/resolve/main/config.json""", """google/mobilenet_v1_0.75_192""": """https://huggingface.co/google/mobilenet_v1_0.75_192/resolve/main/config.json""", # See all MobileNetV1 models at https://huggingface.co/models?filter=mobilenet_v1 } class _UpperCAmelCase ( _snake_case): __lowercase : int = """mobilenet_v1""" def __init__( self , snake_case_=3 , snake_case_=2_24 , snake_case_=1.0 , snake_case_=8 , snake_case_="relu6" , snake_case_=True , snake_case_=0.999 , snake_case_=0.02 , snake_case_=0.001 , **snake_case_ , ): super().__init__(**snake_case_ ) if depth_multiplier <= 0: raise ValueError("depth_multiplier must be greater than zero." ) _snake_case : List[str] = num_channels _snake_case : str = image_size _snake_case : Optional[Any] = depth_multiplier _snake_case : List[Any] = min_depth _snake_case : Any = hidden_act _snake_case : Optional[int] = tf_padding _snake_case : Tuple = classifier_dropout_prob _snake_case : List[Any] = initializer_range _snake_case : Optional[int] = layer_norm_eps class _UpperCAmelCase ( _snake_case): __lowercase : Optional[Any] = version.parse("""1.11""") @property def lowerCamelCase__ ( self ): return OrderedDict([("pixel_values", {0: "batch"})] ) @property def lowerCamelCase__ ( self ): if self.task == "image-classification": return OrderedDict([("logits", {0: "batch"})] ) else: return OrderedDict([("last_hidden_state", {0: "batch"}), ("pooler_output", {0: "batch"})] ) @property def lowerCamelCase__ ( self ): return 1E-4
701
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import os import unittest from transformers import LayoutLMTokenizer, LayoutLMTokenizerFast from transformers.models.layoutlm.tokenization_layoutlm import VOCAB_FILES_NAMES from transformers.testing_utils import require_tokenizers from ...test_tokenization_common import TokenizerTesterMixin @require_tokenizers class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Optional[Any] = LayoutLMTokenizer __lowercase : List[Any] = LayoutLMTokenizerFast __lowercase : str = True __lowercase : List[str] = True def lowerCamelCase__ ( self ): super().setUp() _snake_case : Dict = [ "[UNK]", "[CLS]", "[SEP]", "want", "##want", "##ed", "wa", "un", "runn", "##ing", ",", "low", "lowest", ] _snake_case : List[str] = os.path.join(self.tmpdirname , VOCAB_FILES_NAMES["vocab_file"] ) with open(self.vocab_file , "w" , encoding="utf-8" ) as vocab_writer: vocab_writer.write("".join([x + "\n" for x in vocab_tokens] ) ) def lowerCamelCase__ ( self , **snake_case_ ): return LayoutLMTokenizer.from_pretrained(self.tmpdirname , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): _snake_case : Dict = "UNwant\u00E9d,running" _snake_case : List[Any] = "unwanted, running" return input_text, output_text def lowerCamelCase__ ( self ): _snake_case : Tuple = self.tokenizer_class(self.vocab_file ) _snake_case : Tuple = tokenizer.tokenize("UNwant\u00E9d,running" ) self.assertListEqual(snake_case_ , ["un", "##want", "##ed", ",", "runn", "##ing"] ) self.assertListEqual(tokenizer.convert_tokens_to_ids(snake_case_ ) , [7, 4, 5, 10, 8, 9] ) def lowerCamelCase__ ( self ): pass
702
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Any = TextToVideoSDPipeline __lowercase : str = TEXT_TO_IMAGE_PARAMS __lowercase : int = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __lowercase : Optional[int] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) _snake_case : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : Tuple = CLIPTextModel(snake_case_ ) _snake_case : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : str = torch.manual_seed(snake_case_ ) else: _snake_case : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def lowerCamelCase__ ( self ): _snake_case : int = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : Optional[Any] = self.get_dummy_components() _snake_case : Tuple = TextToVideoSDPipeline(**snake_case_ ) _snake_case : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ ) _snake_case : Union[str, Any] = "np" _snake_case : Dict = sd_pipe(**snake_case_ ).frames _snake_case : Any = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _snake_case : Dict = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=1E-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) _snake_case : int = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _snake_case : Tuple = pipe.to("cuda" ) _snake_case : List[Any] = "Spiderman is surfing" _snake_case : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : int = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="pt" ).frames _snake_case : int = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def lowerCamelCase__ ( self ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) _snake_case : str = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : int = pipe.to("cuda" ) _snake_case : Any = "Spiderman is surfing" _snake_case : str = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Any = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type="pt" ).frames _snake_case : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
87
0
"""simple docstring""" import coval # From: git+https://github.com/ns-moosavi/coval.git # noqa: F401 from coval.conll import reader, util from coval.eval import evaluator import datasets _a : Optional[int] = datasets.logging.get_logger(__name__) _a : List[Any] = """\ @InProceedings{moosavi2019minimum, author = { Nafise Sadat Moosavi, Leo Born, Massimo Poesio and Michael Strube}, title = {Using Automatically Extracted Minimum Spans to Disentangle Coreference Evaluation from Boundary Detection}, year = {2019}, booktitle = {Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)}, publisher = {Association for Computational Linguistics}, address = {Florence, Italy}, } @inproceedings{10.3115/1072399.1072405, author = {Vilain, Marc and Burger, John and Aberdeen, John and Connolly, Dennis and Hirschman, Lynette}, title = {A Model-Theoretic Coreference Scoring Scheme}, year = {1995}, isbn = {1558604022}, publisher = {Association for Computational Linguistics}, address = {USA}, url = {https://doi.org/10.3115/1072399.1072405}, doi = {10.3115/1072399.1072405}, booktitle = {Proceedings of the 6th Conference on Message Understanding}, pages = {45–52}, numpages = {8}, location = {Columbia, Maryland}, series = {MUC6 ’95} } @INPROCEEDINGS{Bagga98algorithmsfor, author = {Amit Bagga and Breck Baldwin}, title = {Algorithms for Scoring Coreference Chains}, booktitle = {In The First International Conference on Language Resources and Evaluation Workshop on Linguistics Coreference}, year = {1998}, pages = {563--566} } @INPROCEEDINGS{Luo05oncoreference, author = {Xiaoqiang Luo}, title = {On coreference resolution performance metrics}, booktitle = {In Proc. of HLT/EMNLP}, year = {2005}, pages = {25--32}, publisher = {URL} } @inproceedings{moosavi-strube-2016-coreference, title = \"Which Coreference Evaluation Metric Do You Trust? A Proposal for a Link-based Entity Aware Metric\", author = \"Moosavi, Nafise Sadat and Strube, Michael\", booktitle = \"Proceedings of the 54th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)\", month = aug, year = \"2016\", address = \"Berlin, Germany\", publisher = \"Association for Computational Linguistics\", url = \"https://www.aclweb.org/anthology/P16-1060\", doi = \"10.18653/v1/P16-1060\", pages = \"632--642\", } """ _a : Union[str, Any] = """\ CoVal is a coreference evaluation tool for the CoNLL and ARRAU datasets which implements of the common evaluation metrics including MUC [Vilain et al, 1995], B-cubed [Bagga and Baldwin, 1998], CEAFe [Luo et al., 2005], LEA [Moosavi and Strube, 2016] and the averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) [Denis and Baldridge, 2009a; Pradhan et al., 2011]. This wrapper of CoVal currently only work with CoNLL line format: The CoNLL format has one word per line with all the annotation for this word in column separated by spaces: Column Type Description 1 Document ID This is a variation on the document filename 2 Part number Some files are divided into multiple parts numbered as 000, 001, 002, ... etc. 3 Word number 4 Word itself This is the token as segmented/tokenized in the Treebank. Initially the *_skel file contain the placeholder [WORD] which gets replaced by the actual token from the Treebank which is part of the OntoNotes release. 5 Part-of-Speech 6 Parse bit This is the bracketed structure broken before the first open parenthesis in the parse, and the word/part-of-speech leaf replaced with a *. The full parse can be created by substituting the asterix with the \"([pos] [word])\" string (or leaf) and concatenating the items in the rows of that column. 7 Predicate lemma The predicate lemma is mentioned for the rows for which we have semantic role information. All other rows are marked with a \"-\" 8 Predicate Frameset ID This is the PropBank frameset ID of the predicate in Column 7. 9 Word sense This is the word sense of the word in Column 3. 10 Speaker/Author This is the speaker or author name where available. Mostly in Broadcast Conversation and Web Log data. 11 Named Entities These columns identifies the spans representing various named entities. 12:N Predicate Arguments There is one column each of predicate argument structure information for the predicate mentioned in Column 7. N Coreference Coreference chain information encoded in a parenthesis structure. More informations on the format can be found here (section \"*_conll File Format\"): http://www.conll.cemantix.org/2012/data.html Details on the evaluation on CoNLL can be found here: https://github.com/ns-moosavi/coval/blob/master/conll/README.md CoVal code was written by @ns-moosavi. Some parts are borrowed from https://github.com/clarkkev/deep-coref/blob/master/evaluation.py The test suite is taken from https://github.com/conll/reference-coreference-scorers/ Mention evaluation and the test suite are added by @andreasvc. Parsing CoNLL files is developed by Leo Born. """ _a : Optional[Any] = """ Calculates coreference evaluation metrics. Args: predictions: list of sentences. Each sentence is a list of word predictions to score in the CoNLL format. Each prediction is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. references: list of sentences. Each sentence is a list of word reference to score in the CoNLL format. Each reference is a word with its annotations as a string made of columns joined with spaces. Only columns 4, 5, 6 and the last column are used (word, POS, Pars and coreference annotation) See the details on the format in the description of the metric. keep_singletons: After extracting all mentions of key or system files, mentions whose corresponding coreference chain is of size one, are considered as singletons. The default evaluation mode will include singletons in evaluations if they are included in the key or the system files. By setting 'keep_singletons=False', all singletons in the key and system files will be excluded from the evaluation. NP_only: Most of the recent coreference resolvers only resolve NP mentions and leave out the resolution of VPs. By setting the 'NP_only' option, the scorer will only evaluate the resolution of NPs. min_span: By setting 'min_span', the scorer reports the results based on automatically detected minimum spans. Minimum spans are determined using the MINA algorithm. Returns: 'mentions': mentions 'muc': MUC metric [Vilain et al, 1995] 'bcub': B-cubed [Bagga and Baldwin, 1998] 'ceafe': CEAFe [Luo et al., 2005] 'lea': LEA [Moosavi and Strube, 2016] 'conll_score': averaged CoNLL score (the average of the F1 values of MUC, B-cubed and CEAFe) Examples: >>> coval = datasets.load_metric('coval') >>> words = ['bc/cctv/00/cctv_0005 0 0 Thank VBP (TOP(S(VP* thank 01 1 Xu_li * (V*) * -', ... 'bc/cctv/00/cctv_0005 0 1 you PRP (NP*) - - - Xu_li * (ARG1*) (ARG0*) (116)', ... 'bc/cctv/00/cctv_0005 0 2 everyone NN (NP*) - - - Xu_li * (ARGM-DIS*) * (116)', ... 'bc/cctv/00/cctv_0005 0 3 for IN (PP* - - - Xu_li * (ARG2* * -', ... 'bc/cctv/00/cctv_0005 0 4 watching VBG (S(VP*)))) watch 01 1 Xu_li * *) (V*) -', ... 'bc/cctv/00/cctv_0005 0 5 . . *)) - - - Xu_li * * * -'] >>> references = [words] >>> predictions = [words] >>> results = coval.compute(predictions=predictions, references=references) >>> print(results) # doctest:+ELLIPSIS {'mentions/recall': 1.0,[...] 'conll_score': 100.0} """ def a__ ( a : Optional[int] , a : str , a : int=False , a : Optional[int]=False , a : Optional[int]=True , a : Dict=False , a : List[str]="dummy_doc" ): """simple docstring""" _snake_case : str = {doc: key_lines} _snake_case : Union[str, Any] = {doc: sys_lines} _snake_case : int = {} _snake_case : Union[str, Any] = 0 _snake_case : str = 0 _snake_case : Optional[int] = 0 _snake_case : Optional[int] = 0 _snake_case : List[str] = 0 _snake_case : List[str] = 0 _snake_case : List[Any] = reader.get_doc_mentions(a , key_doc_lines[doc] , a ) key_singletons_num += singletons_num if NP_only or min_span: _snake_case : Union[str, Any] = reader.set_annotated_parse_trees(a , key_doc_lines[doc] , a , a ) _snake_case : str = reader.get_doc_mentions(a , sys_doc_lines[doc] , a ) sys_singletons_num += singletons_num if NP_only or min_span: _snake_case : Any = reader.set_annotated_parse_trees(a , key_doc_lines[doc] , a , a ) if remove_nested: _snake_case : List[Any] = reader.remove_nested_coref_mentions(a , a ) key_nested_coref_num += nested_mentions key_removed_nested_clusters += removed_clusters _snake_case : List[Any] = reader.remove_nested_coref_mentions(a , a ) sys_nested_coref_num += nested_mentions sys_removed_nested_clusters += removed_clusters _snake_case : Optional[int] = reader.get_mention_assignments(a , a ) _snake_case : Optional[int] = reader.get_mention_assignments(a , a ) _snake_case : Dict = (key_clusters, sys_clusters, key_mention_sys_cluster, sys_mention_key_cluster) if remove_nested: logger.info( "Number of removed nested coreferring mentions in the key " f'annotation: {key_nested_coref_num}; and system annotation: {sys_nested_coref_num}' ) logger.info( "Number of resulting singleton clusters in the key " f'annotation: {key_removed_nested_clusters}; and system annotation: {sys_removed_nested_clusters}' ) if not keep_singletons: logger.info( f'{key_singletons_num:d} and {sys_singletons_num:d} singletons are removed from the key and system ' "files, respectively" ) return doc_coref_infos def a__ ( a : int , a : Optional[Any] , a : Any , a : Optional[int] , a : Union[str, Any] , a : Optional[int] , a : Any ): """simple docstring""" _snake_case : str = get_coref_infos(a , a , a , a , a , a ) _snake_case : Any = {} _snake_case : List[Any] = 0 _snake_case : Optional[Any] = 0 for name, metric in metrics: _snake_case : str = evaluator.evaluate_documents(a , a , beta=1 ) if name in ["muc", "bcub", "ceafe"]: conll += fa conll_subparts_num += 1 output_scores.update({f'{name}/recall': recall, f'{name}/precision': precision, f'{name}/f1': fa} ) logger.info( name.ljust(10 ) , f'Recall: {recall * 100:.2f}' , f' Precision: {precision * 100:.2f}' , f' F1: {fa * 100:.2f}' , ) if conll_subparts_num == 3: _snake_case : Any = (conll / 3) * 100 logger.info(f'CoNLL score: {conll:.2f}' ) output_scores.update({"conll_score": conll} ) return output_scores def a__ ( a : List[Any] ): """simple docstring""" _snake_case : int = False for line in key_lines: if not line.startswith("#" ): if len(line.split() ) > 6: _snake_case : Optional[int] = line.split()[5] if not parse_col == "-": _snake_case : List[str] = True break else: break return has_gold_parse @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _UpperCAmelCase ( datasets.Metric): def lowerCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" ) ), "references": datasets.Sequence(datasets.Value("string" ) ), } ) , codebase_urls=["https://github.com/ns-moosavi/coval"] , reference_urls=[ "https://github.com/ns-moosavi/coval", "https://www.aclweb.org/anthology/P16-1060", "http://www.conll.cemantix.org/2012/data.html", ] , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=True , snake_case_=False , snake_case_=False , snake_case_=False ): _snake_case : Optional[int] = [ ("mentions", evaluator.mentions), ("muc", evaluator.muc), ("bcub", evaluator.b_cubed), ("ceafe", evaluator.ceafe), ("lea", evaluator.lea), ] if min_span: _snake_case : List[Any] = util.check_gold_parse_annotation(snake_case_ ) if not has_gold_parse: raise NotImplementedError("References should have gold parse annotation to use 'min_span'." ) # util.parse_key_file(key_file) # key_file = key_file + ".parsed" _snake_case : str = evaluate( key_lines=snake_case_ , sys_lines=snake_case_ , metrics=snake_case_ , NP_only=snake_case_ , remove_nested=snake_case_ , keep_singletons=snake_case_ , min_span=snake_case_ , ) return score
703
"""simple docstring""" from typing import List, Optional import numpy as np from ...processing_utils import ProcessorMixin from ...utils import to_numpy class _UpperCAmelCase ( _snake_case): __lowercase : int = """EncodecFeatureExtractor""" __lowercase : str = ("""T5Tokenizer""", """T5TokenizerFast""") def __init__( self , snake_case_ , snake_case_ ): super().__init__(snake_case_ , snake_case_ ) _snake_case : Dict = self.feature_extractor _snake_case : Any = False def lowerCamelCase__ ( self , snake_case_=None , snake_case_=None , snake_case_=True ): return self.tokenizer.get_decoder_prompt_ids(task=snake_case_ , language=snake_case_ , no_timestamps=snake_case_ ) def __call__( self , *snake_case_ , **snake_case_ ): # For backward compatibility if self._in_target_context_manager: return self.current_processor(*snake_case_ , **snake_case_ ) _snake_case : str = kwargs.pop("audio" , snake_case_ ) _snake_case : Optional[int] = kwargs.pop("sampling_rate" , snake_case_ ) _snake_case : Optional[Any] = kwargs.pop("text" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Any = args[0] _snake_case : Union[str, Any] = args[1:] if audio is None and text is None: raise ValueError("You need to specify either an `audio` or `text` input to process." ) if text is not None: _snake_case : Any = self.tokenizer(snake_case_ , **snake_case_ ) if audio is not None: _snake_case : Any = self.feature_extractor(snake_case_ , *snake_case_ , sampling_rate=snake_case_ , **snake_case_ ) if audio is None: return inputs elif text is None: return audio_inputs else: _snake_case : str = audio_inputs["input_values"] if "padding_mask" in audio_inputs: _snake_case : List[str] = audio_inputs["padding_mask"] return inputs def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): _snake_case : Tuple = kwargs.pop("audio" , snake_case_ ) _snake_case : List[str] = kwargs.pop("padding_mask" , snake_case_ ) if len(snake_case_ ) > 0: _snake_case : Tuple = args[0] _snake_case : Dict = args[1:] if audio_values is not None: return self._decode_audio(snake_case_ , padding_mask=snake_case_ ) else: return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ = None ): _snake_case : Optional[int] = to_numpy(snake_case_ ) _snake_case , _snake_case , _snake_case : Tuple = audio_values.shape if padding_mask is None: return list(snake_case_ ) _snake_case : Optional[int] = to_numpy(snake_case_ ) # match the sequence length of the padding mask to the generated audio arrays by padding with the **non-padding** # token (so that the generated audio values are **not** treated as padded tokens) _snake_case : Any = seq_len - padding_mask.shape[-1] _snake_case : Optional[Any] = 1 - self.feature_extractor.padding_value _snake_case : Optional[int] = np.pad(snake_case_ , ((0, 0), (0, difference)) , "constant" , constant_values=snake_case_ ) _snake_case : Any = audio_values.tolist() for i in range(snake_case_ ): _snake_case : Tuple = np.asarray(audio_values[i] )[ padding_mask[i][None, :] != self.feature_extractor.padding_value ] _snake_case : Tuple = sliced_audio.reshape(snake_case_ , -1 ) return audio_values
87
0
import argparse from transformers import BigBirdConfig, BigBirdForPreTraining, BigBirdForQuestionAnswering, load_tf_weights_in_big_bird from transformers.utils import logging logging.set_verbosity_info() def a__ ( a : Dict , a : Union[str, Any] , a : int , a : Any ): """simple docstring""" _snake_case : int = BigBirdConfig.from_json_file(a ) print(f'Building PyTorch model from configuration: {config}' ) if is_trivia_qa: _snake_case : Union[str, Any] = BigBirdForQuestionAnswering(a ) else: _snake_case : Dict = BigBirdForPreTraining(a ) # Load weights from tf checkpoint load_tf_weights_in_big_bird(a , a , is_trivia_qa=a ) # Save pytorch-model print(f'Save PyTorch model to {pytorch_dump_path}' ) model.save_pretrained(a ) if __name__ == "__main__": _a : List[Any] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--tf_checkpoint_path""", default=None, type=str, required=True, help="""Path to the TensorFlow checkpoint path.""" ) parser.add_argument( """--big_bird_config_file""", default=None, type=str, required=True, help=( """The config json file corresponding to the pre-trained BERT model. \n""" """This specifies the model architecture.""" ), ) parser.add_argument( """--pytorch_dump_path""", default=None, type=str, required=True, help="""Path to the output PyTorch model.""" ) parser.add_argument( """--is_trivia_qa""", action="""store_true""", help="""Whether to convert a model with a trivia_qa head.""" ) _a : List[Any] = parser.parse_args() convert_tf_checkpoint_to_pytorch( args.tf_checkpoint_path, args.big_bird_config_file, args.pytorch_dump_path, args.is_trivia_qa )
704
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : str = {"""configuration_yolos""": ["""YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP""", """YolosConfig""", """YolosOnnxConfig"""]} try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Optional[Any] = ["""YolosFeatureExtractor"""] _a : List[Any] = ["""YolosImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Union[str, Any] = [ """YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST""", """YolosForObjectDetection""", """YolosModel""", """YolosPreTrainedModel""", ] if TYPE_CHECKING: from .configuration_yolos import YOLOS_PRETRAINED_CONFIG_ARCHIVE_MAP, YolosConfig, YolosOnnxConfig try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .feature_extraction_yolos import YolosFeatureExtractor from .image_processing_yolos import YolosImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_yolos import ( YOLOS_PRETRAINED_MODEL_ARCHIVE_LIST, YolosForObjectDetection, YolosModel, YolosPreTrainedModel, ) else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" def a__ ( a : int ): """simple docstring""" if not isinstance(a , a ): raise TypeError("Input value must be an 'int' type" ) _snake_case : Union[str, Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
705
"""simple docstring""" import numpy as np import torch from torch.utils.data import Dataset, IterableDataset from ..utils.generic import ModelOutput class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = dataset _snake_case : str = process _snake_case : int = params def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): _snake_case : Union[str, Any] = self.dataset[i] _snake_case : Optional[Any] = self.process(snake_case_ , **self.params ) return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): _snake_case : Union[str, Any] = loader _snake_case : Tuple = infer _snake_case : List[Any] = params if loader_batch_size == 1: # Let's spare some time by deactivating altogether _snake_case : int = None _snake_case : int = loader_batch_size # Internal bookkeeping _snake_case : Any = None _snake_case : Dict = None def __len__( self ): return len(self.loader ) def __iter__( self ): _snake_case : int = iter(self.loader ) return self def lowerCamelCase__ ( self ): if isinstance(self._loader_batch_data , torch.Tensor ): # Batch data is simple tensor, just fetch the slice _snake_case : List[Any] = self._loader_batch_data[self._loader_batch_index] else: # Batch data is assumed to be BaseModelOutput (or dict) _snake_case : int = {} for k, element in self._loader_batch_data.items(): if isinstance(snake_case_ , snake_case_ ): # Convert ModelOutput to tuple first _snake_case : Tuple = element.to_tuple() if isinstance(element[0] , torch.Tensor ): _snake_case : Union[str, Any] = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : int = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if k in {"hidden_states", "past_key_values", "attentions"} and isinstance(snake_case_ , snake_case_ ): # Those are stored as lists of tensors so need specific unbatching. if isinstance(element[0] , torch.Tensor ): _snake_case : Tuple = tuple(el[self._loader_batch_index].unsqueeze(0 ) for el in element ) elif isinstance(element[0] , np.ndarray ): _snake_case : Tuple = tuple(np.expand_dims(el[self._loader_batch_index] , 0 ) for el in element ) continue if element is None: # This can happen for optional data that get passed around _snake_case : Tuple = None elif isinstance(element[self._loader_batch_index] , torch.Tensor ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : List[Any] = element[self._loader_batch_index].unsqueeze(0 ) elif isinstance(element[self._loader_batch_index] , np.ndarray ): # Take correct batch data, but make it looked like batch_size=1 # For compatibility with other methods within transformers _snake_case : Union[str, Any] = np.expand_dims(element[self._loader_batch_index] , 0 ) else: # This is typically a list, so no need to `unsqueeze`. _snake_case : List[Any] = element[self._loader_batch_index] # Recreate the element by reusing the original class to make it look # batch_size=1 _snake_case : int = self._loader_batch_data.__class__(snake_case_ ) self._loader_batch_index += 1 return result def lowerCamelCase__ ( self ): if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: # We are currently unrolling a batch so we just need to return # the current item within a batch return self.loader_batch_item() # We're out of items within a batch _snake_case : Tuple = next(self.iterator ) _snake_case : Any = self.infer(snake_case_ , **self.params ) # We now have a batch of "inferred things". if self.loader_batch_size is not None: # Try to infer the size of the batch if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Optional[int] = list(processed.keys() )[0] _snake_case : List[str] = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = len(snake_case_ ) else: _snake_case : Optional[int] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Union[str, Any] = observed_batch_size # Setting internal index to unwrap the batch _snake_case : str = processed _snake_case : List[Any] = 0 return self.loader_batch_item() else: # We're not unrolling batches return processed class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_=None ): super().__init__(snake_case_ , snake_case_ , snake_case_ ) def __iter__( self ): _snake_case : Tuple = iter(self.loader ) _snake_case : List[Any] = None return self def lowerCamelCase__ ( self ): if self.subiterator is None: _snake_case : Optional[Any] = self.infer(next(self.iterator ) , **self.params ) try: # Try to return next item _snake_case : Union[str, Any] = next(self.subiterator ) except StopIteration: # When a preprocess iterator ends, we can start lookig at the next item # ChunkIterator will keep feeding until ALL elements of iterator # all have created their subiterator and have been iterating against. # # Another way to look at it, is we're basically flattening lists of lists # into a single list, but with generators _snake_case : str = self.infer(next(self.iterator ) , **self.params ) _snake_case : Tuple = next(self.subiterator ) return processed class _UpperCAmelCase ( _snake_case): def __iter__( self ): _snake_case : Optional[Any] = iter(self.loader ) return self def lowerCamelCase__ ( self ): # Extremely similar to PipelineIterator in its unpacking mechanism # BUT, we have an extra required item which is the presence of `is_last` # That is because everything is flattened by `PipelineChunkIterator` we # need to keep track of how to regroup here in the original `process` # boundaries so that `process` and `postprocess` see the same data. # This iterator accumulates items (possibly while unbatching) until it # its a `is_last` and then just passes it on to the caller. _snake_case : Optional[Any] = False _snake_case : Tuple = [] if self._loader_batch_index is not None and self._loader_batch_index < self.loader_batch_size: while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : str = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator while not is_last: _snake_case : List[str] = self.infer(next(self.iterator ) , **self.params ) if self.loader_batch_size is not None: if isinstance(snake_case_ , torch.Tensor ): _snake_case : Union[str, Any] = processed else: _snake_case : Tuple = list(processed.keys() )[0] _snake_case : Tuple = processed[key] if isinstance(snake_case_ , snake_case_ ): _snake_case : Any = len(snake_case_ ) else: _snake_case : List[Any] = first_tensor.shape[0] if 0 < observed_batch_size < self.loader_batch_size: # could be last batch so we can't unroll as many # elements. _snake_case : Dict = observed_batch_size _snake_case : List[Any] = processed _snake_case : List[str] = 0 while self._loader_batch_index < self.loader_batch_size: _snake_case : Union[str, Any] = self.loader_batch_item() _snake_case : int = item.pop("is_last" ) accumulator.append(snake_case_ ) if is_last: return accumulator else: _snake_case : Dict = processed _snake_case : Dict = item.pop("is_last" ) accumulator.append(snake_case_ ) return accumulator class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ ): _snake_case : str = dataset _snake_case : Any = key def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return self.dataset[i][self.key] class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = dataset _snake_case : Any = keya _snake_case : int = keya def __len__( self ): return len(self.dataset ) def __getitem__( self , snake_case_ ): return {"text": self.dataset[i][self.keya], "text_pair": self.dataset[i][self.keya]}
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available lowerCamelCase_ : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: lowerCamelCase_ : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys lowerCamelCase_ : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
706
"""simple docstring""" def a__ ( a : int ): """simple docstring""" if not isinstance(a , a ): raise TypeError("Input value must be an 'int' type" ) _snake_case : Union[str, Any] = 0 while number: position += 1 number >>= 1 return position if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" import json import sys import tempfile import unittest from pathlib import Path import transformers from transformers import ( CONFIG_MAPPING, IMAGE_PROCESSOR_MAPPING, AutoConfig, AutoImageProcessor, CLIPConfig, CLIPImageProcessor, ) from transformers.testing_utils import DUMMY_UNKNOWN_IDENTIFIER sys.path.append(str(Path(__file__).parent.parent.parent.parent / """utils""")) from test_module.custom_configuration import CustomConfig # noqa E402 from test_module.custom_image_processing import CustomImageProcessor # noqa E402 class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : Optional[int] = 0 def lowerCamelCase__ ( self ): _snake_case : str = AutoImageProcessor.from_pretrained("openai/clip-vit-base-patch32" ) self.assertIsInstance(snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Dict = Path(snake_case_ ) / "preprocessor_config.json" _snake_case : List[str] = Path(snake_case_ ) / "config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(snake_case_ , "w" ) , ) json.dump({"model_type": "clip"} , open(snake_case_ , "w" ) ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): # Ensure we can load the image processor from the feature extractor config with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : str = Path(snake_case_ ) / "preprocessor_config.json" _snake_case : Dict = Path(snake_case_ ) / "config.json" json.dump( {"feature_extractor_type": "CLIPFeatureExtractor", "processor_class": "CLIPProcessor"} , open(snake_case_ , "w" ) , ) json.dump({"model_type": "clip"} , open(snake_case_ , "w" ) ) _snake_case : Dict = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Tuple = CLIPConfig() # Create a dummy config file with image_proceesor_type _snake_case : Any = Path(snake_case_ ) / "preprocessor_config.json" _snake_case : Dict = Path(snake_case_ ) / "config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(snake_case_ , "w" ) , ) json.dump({"model_type": "clip"} , open(snake_case_ , "w" ) ) # remove image_processor_type to make sure config.json alone is enough to load image processor locally _snake_case : int = AutoImageProcessor.from_pretrained(snake_case_ ).to_dict() config_dict.pop("image_processor_type" ) _snake_case : Dict = CLIPImageProcessor(**snake_case_ ) # save in new folder model_config.save_pretrained(snake_case_ ) config.save_pretrained(snake_case_ ) _snake_case : List[Any] = AutoImageProcessor.from_pretrained(snake_case_ ) # make sure private variable is not incorrectly saved _snake_case : Union[str, Any] = json.loads(config.to_json_string() ) self.assertTrue("_processor_class" not in dict_as_saved ) self.assertIsInstance(snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Tuple = Path(snake_case_ ) / "preprocessor_config.json" json.dump( {"image_processor_type": "CLIPImageProcessor", "processor_class": "CLIPProcessor"} , open(snake_case_ , "w" ) , ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): with self.assertRaisesRegex( snake_case_ , "clip-base is not a local folder and is not a valid model identifier" ): _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained("clip-base" ) def lowerCamelCase__ ( self ): with self.assertRaisesRegex( snake_case_ , r"aaaaaa is not a valid git identifier \(branch name, tag name or commit id\)" ): _snake_case : List[str] = AutoImageProcessor.from_pretrained(snake_case_ , revision="aaaaaa" ) def lowerCamelCase__ ( self ): with self.assertRaisesRegex( snake_case_ , "hf-internal-testing/config-no-model does not appear to have a file named preprocessor_config.json." , ): _snake_case : str = AutoImageProcessor.from_pretrained("hf-internal-testing/config-no-model" ) def lowerCamelCase__ ( self ): # If remote code is not set, we will time out when asking whether to load the model. with self.assertRaises(snake_case_ ): _snake_case : str = AutoImageProcessor.from_pretrained("hf-internal-testing/test_dynamic_image_processor" ) # If remote code is disabled, we can't load this config. with self.assertRaises(snake_case_ ): _snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=snake_case_ ) _snake_case : Dict = AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) # Test image processor can be reloaded. with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(snake_case_ ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained(snake_case_ , trust_remote_code=snake_case_ ) self.assertEqual(reloaded_image_processor.__class__.__name__ , "NewImageProcessor" ) def lowerCamelCase__ ( self ): try: AutoConfig.register("custom" , snake_case_ ) AutoImageProcessor.register(snake_case_ , snake_case_ ) # Trying to register something existing in the Transformers library will raise an error with self.assertRaises(snake_case_ ): AutoImageProcessor.register(snake_case_ , snake_case_ ) with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : Optional[Any] = Path(snake_case_ ) / "preprocessor_config.json" _snake_case : Dict = Path(snake_case_ ) / "config.json" json.dump( {"feature_extractor_type": "CLIPFeatureExtractor", "processor_class": "CLIPProcessor"} , open(snake_case_ , "w" ) , ) json.dump({"model_type": "clip"} , open(snake_case_ , "w" ) ) _snake_case : Tuple = CustomImageProcessor.from_pretrained(snake_case_ ) # Now that the config is registered, it can be used as any other config with the auto-API with tempfile.TemporaryDirectory() as tmp_dir: image_processor.save_pretrained(snake_case_ ) _snake_case : Optional[Any] = AutoImageProcessor.from_pretrained(snake_case_ ) self.assertIsInstance(snake_case_ , snake_case_ ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig] def lowerCamelCase__ ( self ): class _UpperCAmelCase ( _snake_case): __lowercase : str = True try: AutoConfig.register("custom" , snake_case_ ) AutoImageProcessor.register(snake_case_ , snake_case_ ) # If remote code is not set, the default is to use local _snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained("hf-internal-testing/test_dynamic_image_processor" ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(image_processor.is_local ) # If remote code is disabled, we load the local one. _snake_case : Union[str, Any] = AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(image_processor.is_local ) # If remote is enabled, we load from the Hub _snake_case : Dict = AutoImageProcessor.from_pretrained( "hf-internal-testing/test_dynamic_image_processor" , trust_remote_code=snake_case_ ) self.assertEqual(image_processor.__class__.__name__ , "NewImageProcessor" ) self.assertTrue(not hasattr(snake_case_ , "is_local" ) ) finally: if "custom" in CONFIG_MAPPING._extra_content: del CONFIG_MAPPING._extra_content["custom"] if CustomConfig in IMAGE_PROCESSOR_MAPPING._extra_content: del IMAGE_PROCESSOR_MAPPING._extra_content[CustomConfig]
707
"""simple docstring""" from __future__ import annotations import requests _a : List[str] = set( """approved_at_utc approved_by author_flair_background_color author_flair_css_class author_flair_richtext author_flair_template_id author_fullname author_premium can_mod_post category clicked content_categories created_utc downs edited gilded gildings hidden hide_score is_created_from_ads_ui is_meta is_original_content is_reddit_media_domain is_video link_flair_css_class link_flair_richtext link_flair_text link_flair_text_color media_embed mod_reason_title name permalink pwls quarantine saved score secure_media secure_media_embed selftext subreddit subreddit_name_prefixed subreddit_type thumbnail title top_awarded_type total_awards_received ups upvote_ratio url user_reports""".split() ) def a__ ( a : str , a : int = 1 , a : str = "new" , a : list | None = None ): """simple docstring""" _snake_case : Any = wanted_data or [] if invalid_search_terms := ", ".join(sorted(set(a ) - valid_terms ) ): _snake_case : Optional[int] = f'Invalid search term: {invalid_search_terms}' raise ValueError(a ) _snake_case : int = requests.get( f'https://reddit.com/r/{subreddit}/{age}.json?limit={limit}' , headers={"User-agent": "A random string"} , ) if response.status_code == 429: raise requests.HTTPError _snake_case : Optional[Any] = response.json() if not wanted_data: return {id_: data["data"]["children"][id_] for id_ in range(a )} _snake_case : Tuple = {} for id_ in range(a ): _snake_case : List[str] = { item: data["data"]["children"][id_]["data"][item] for item in wanted_data } return data_dict if __name__ == "__main__": # If you get Error 429, that means you are rate limited.Try after some time print(get_subreddit_data("""learnpython""", wanted_data=["""title""", """url""", """selftext"""]))
87
0
"""simple docstring""" def a__ ( a : list ): """simple docstring""" _snake_case : Optional[Any] = len(a ) for _ in range(a ): for i in range(_ % 2 , arr_size - 1 , 2 ): if arr[i + 1] < arr[i]: _snake_case : Dict = arr[i + 1], arr[i] return arr if __name__ == "__main__": _a : Union[str, Any] = list(range(10, 0, -1)) print(f'Original: {arr}. Sorted: {odd_even_transposition(arr)}')
708
"""simple docstring""" from __future__ import annotations from numpy import array, cos, cross, floataa, radians, sin from numpy.typing import NDArray def a__ ( a : float , a : float , a : bool = False ): """simple docstring""" if radian_mode: return [magnitude * cos(a ), magnitude * sin(a )] return [magnitude * cos(radians(a ) ), magnitude * sin(radians(a ) )] def a__ ( a : NDArray[floataa] , a : NDArray[floataa] , a : float = 10**-1 ): """simple docstring""" _snake_case : NDArray[floataa] = cross(a , a ) _snake_case : float = sum(a ) return abs(a ) < eps if __name__ == "__main__": # Test to check if it works _a : Tuple = array( [ polar_force(718.4, 180 - 30), polar_force(879.54, 45), polar_force(100, -90), ] ) _a : NDArray[floataa] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem 1 in image_data/2D_problems.jpg _a : List[Any] = array( [ polar_force(30 * 9.81, 15), polar_force(215, 180 - 45), polar_force(264, 90 - 30), ] ) _a : List[Any] = array([[0, 0], [0, 0], [0, 0]]) assert in_static_equilibrium(forces, location) # Problem in image_data/2D_problems_1.jpg _a : List[str] = array([[0, -2_000], [0, -1_200], [0, 15_600], [0, -12_400]]) _a : Optional[int] = array([[0, 0], [6, 0], [10, 0], [12, 0]]) assert in_static_equilibrium(forces, location) import doctest doctest.testmod()
87
0
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import _LazyModule _a : List[Any] = {"""tokenization_bertweet""": ["""BertweetTokenizer"""]} if TYPE_CHECKING: from .tokenization_bertweet import BertweetTokenizer else: import sys _a : Union[str, Any] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
709
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : Optional[int] = logging.get_logger(__name__) _a : str = {"""openai-gpt""": """https://huggingface.co/openai-gpt/resolve/main/config.json"""} class _UpperCAmelCase ( _snake_case): __lowercase : Optional[Any] = """openai-gpt""" __lowercase : Dict = { """max_position_embeddings""": """n_positions""", """hidden_size""": """n_embd""", """num_attention_heads""": """n_head""", """num_hidden_layers""": """n_layer""", } def __init__( self , snake_case_=4_04_78 , snake_case_=5_12 , snake_case_=7_68 , snake_case_=12 , snake_case_=12 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=0.1 , snake_case_=1E-5 , snake_case_=0.02 , snake_case_="cls_index" , snake_case_=True , snake_case_=None , snake_case_=True , snake_case_=0.1 , **snake_case_ , ): _snake_case : Tuple = vocab_size _snake_case : Dict = n_positions _snake_case : Any = n_embd _snake_case : Any = n_layer _snake_case : Optional[int] = n_head _snake_case : Union[str, Any] = afn _snake_case : Dict = resid_pdrop _snake_case : str = embd_pdrop _snake_case : Union[str, Any] = attn_pdrop _snake_case : str = layer_norm_epsilon _snake_case : Union[str, Any] = initializer_range _snake_case : Any = summary_type _snake_case : List[str] = summary_use_proj _snake_case : Optional[int] = summary_activation _snake_case : Union[str, Any] = summary_first_dropout _snake_case : Optional[int] = summary_proj_to_labels super().__init__(**snake_case_ )
87
0
"""simple docstring""" import os import time import pytest from datasets.utils.filelock import FileLock, Timeout def a__ ( a : int ): """simple docstring""" _snake_case : Optional[Any] = FileLock(str(tmpdir / "foo.lock" ) ) _snake_case : int = FileLock(str(tmpdir / "foo.lock" ) ) _snake_case : int = 0.01 with locka.acquire(): with pytest.raises(a ): _snake_case : List[Any] = time.time() locka.acquire(a ) assert time.time() - _start > timeout def a__ ( a : Optional[int] ): """simple docstring""" _snake_case : Optional[int] = "a" * 1_000 + ".lock" _snake_case : Tuple = FileLock(str(tmpdir / filename ) ) assert locka._lock_file.endswith(".lock" ) assert not locka._lock_file.endswith(a ) assert len(os.path.basename(locka._lock_file ) ) <= 255 _snake_case : int = FileLock(tmpdir / filename ) with locka.acquire(): with pytest.raises(a ): locka.acquire(0 )
710
"""simple docstring""" import argparse from collections import OrderedDict from pathlib import Path import torch from huggingface_hub import hf_hub_download from PIL import Image from torchvision.transforms import functional as F from transformers import DetrImageProcessor, TableTransformerConfig, TableTransformerForObjectDetection from transformers.utils import logging logging.set_verbosity_info() _a : Tuple = logging.get_logger(__name__) # here we list all keys to be renamed (original name on the left, our name on the right) _a : int = [] for i in range(6): # encoder layers: output projection, 2 feedforward neural networks and 2 layernorms rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.weight', f'encoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.encoder.layers.{i}.self_attn.out_proj.bias', f'encoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.weight', f'encoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear1.bias', f'encoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.weight', f'encoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.linear2.bias', f'encoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.encoder.layers.{i}.norm1.weight', f'encoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.encoder.layers.{i}.norm1.bias', f'encoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.weight', f'encoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.encoder.layers.{i}.norm2.bias', f'encoder.layers.{i}.final_layer_norm.bias')) # decoder layers: 2 times output projection, 2 feedforward neural networks and 3 layernorms rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.weight', f'decoder.layers.{i}.self_attn.out_proj.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.self_attn.out_proj.bias', f'decoder.layers.{i}.self_attn.out_proj.bias') ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.weight', f'decoder.layers.{i}.encoder_attn.out_proj.weight', ) ) rename_keys.append( ( f'transformer.decoder.layers.{i}.multihead_attn.out_proj.bias', f'decoder.layers.{i}.encoder_attn.out_proj.bias', ) ) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.weight', f'decoder.layers.{i}.fc1.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear1.bias', f'decoder.layers.{i}.fc1.bias')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.weight', f'decoder.layers.{i}.fc2.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.linear2.bias', f'decoder.layers.{i}.fc2.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm1.weight', f'decoder.layers.{i}.self_attn_layer_norm.weight') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm1.bias', f'decoder.layers.{i}.self_attn_layer_norm.bias')) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.weight', f'decoder.layers.{i}.encoder_attn_layer_norm.weight') ) rename_keys.append( (f'transformer.decoder.layers.{i}.norm2.bias', f'decoder.layers.{i}.encoder_attn_layer_norm.bias') ) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.weight', f'decoder.layers.{i}.final_layer_norm.weight')) rename_keys.append((f'transformer.decoder.layers.{i}.norm3.bias', f'decoder.layers.{i}.final_layer_norm.bias')) # convolutional projection + query embeddings + layernorm of encoder + layernorm of decoder + class and bounding box heads rename_keys.extend( [ ("""input_proj.weight""", """input_projection.weight"""), ("""input_proj.bias""", """input_projection.bias"""), ("""query_embed.weight""", """query_position_embeddings.weight"""), ("""transformer.encoder.norm.weight""", """encoder.layernorm.weight"""), ("""transformer.encoder.norm.bias""", """encoder.layernorm.bias"""), ("""transformer.decoder.norm.weight""", """decoder.layernorm.weight"""), ("""transformer.decoder.norm.bias""", """decoder.layernorm.bias"""), ("""class_embed.weight""", """class_labels_classifier.weight"""), ("""class_embed.bias""", """class_labels_classifier.bias"""), ("""bbox_embed.layers.0.weight""", """bbox_predictor.layers.0.weight"""), ("""bbox_embed.layers.0.bias""", """bbox_predictor.layers.0.bias"""), ("""bbox_embed.layers.1.weight""", """bbox_predictor.layers.1.weight"""), ("""bbox_embed.layers.1.bias""", """bbox_predictor.layers.1.bias"""), ("""bbox_embed.layers.2.weight""", """bbox_predictor.layers.2.weight"""), ("""bbox_embed.layers.2.bias""", """bbox_predictor.layers.2.bias"""), ] ) def a__ ( a : List[str] , a : int , a : int ): """simple docstring""" _snake_case : Union[str, Any] = state_dict.pop(a ) _snake_case : Union[str, Any] = val def a__ ( a : Tuple ): """simple docstring""" _snake_case : Tuple = OrderedDict() for key, value in state_dict.items(): if "backbone.0.body" in key: _snake_case : Dict = key.replace("backbone.0.body" , "backbone.conv_encoder.model" ) _snake_case : Tuple = value else: _snake_case : Dict = value return new_state_dict def a__ ( a : int ): """simple docstring""" _snake_case : Any = "" # first: transformer encoder for i in range(6 ): # read in weights + bias of input projection layer (in PyTorch's MultiHeadAttention, this is a single matrix + bias) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : str = state_dict.pop(f'{prefix}transformer.encoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : int = in_proj_weight[:256, :] _snake_case : List[str] = in_proj_bias[:256] _snake_case : Optional[Any] = in_proj_weight[256:512, :] _snake_case : List[str] = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : Dict = in_proj_bias[-256:] # next: transformer decoder (which is a bit more complex because it also includes cross-attention) for i in range(6 ): # read in weights + bias of input projection layer of self-attention _snake_case : List[Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_weight' ) _snake_case : Union[str, Any] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.self_attn.in_proj_bias' ) # next, add query, keys and values (in that order) to the state dict _snake_case : Union[str, Any] = in_proj_weight[:256, :] _snake_case : Tuple = in_proj_bias[:256] _snake_case : int = in_proj_weight[256:512, :] _snake_case : int = in_proj_bias[256:512] _snake_case : Dict = in_proj_weight[-256:, :] _snake_case : str = in_proj_bias[-256:] # read in weights + bias of input projection layer of cross-attention _snake_case : Dict = state_dict.pop( f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_weight' ) _snake_case : Optional[int] = state_dict.pop(f'{prefix}transformer.decoder.layers.{i}.multihead_attn.in_proj_bias' ) # next, add query, keys and values (in that order) of cross-attention to the state dict _snake_case : Dict = in_proj_weight_cross_attn[:256, :] _snake_case : Any = in_proj_bias_cross_attn[:256] _snake_case : Union[str, Any] = in_proj_weight_cross_attn[256:512, :] _snake_case : Optional[int] = in_proj_bias_cross_attn[256:512] _snake_case : Any = in_proj_weight_cross_attn[-256:, :] _snake_case : str = in_proj_bias_cross_attn[-256:] def a__ ( a : str , a : int ): """simple docstring""" _snake_case , _snake_case : List[str] = image.size _snake_case : Dict = max(a , a ) _snake_case : Union[str, Any] = 800 if "detection" in checkpoint_url else 1_000 _snake_case : Any = target_max_size / current_max_size _snake_case : str = image.resize((int(round(scale * width ) ), int(round(scale * height ) )) ) return resized_image def a__ ( a : str ): """simple docstring""" _snake_case : str = F.to_tensor(a ) _snake_case : Union[str, Any] = F.normalize(a , mean=[0.485, 0.456, 0.406] , std=[0.229, 0.224, 0.225] ) return image @torch.no_grad() def a__ ( a : Optional[Any] , a : Any , a : Union[str, Any] ): """simple docstring""" logger.info("Converting model..." ) # load original state dict _snake_case : Tuple = torch.hub.load_state_dict_from_url(a , map_location="cpu" ) # rename keys for src, dest in rename_keys: rename_key(a , a , a ) _snake_case : Union[str, Any] = rename_backbone_keys(a ) # query, key and value matrices need special treatment read_in_q_k_v(a ) # important: we need to prepend a prefix to each of the base model keys as the head models use different attributes for them _snake_case : int = "model." for key in state_dict.copy().keys(): if not key.startswith("class_labels_classifier" ) and not key.startswith("bbox_predictor" ): _snake_case : Optional[int] = state_dict.pop(a ) _snake_case : Any = val # create HuggingFace model and load state dict _snake_case : Tuple = TableTransformerConfig( backbone="resnet18" , mask_loss_coefficient=1 , dice_loss_coefficient=1 , ce_loss_coefficient=1 , bbox_loss_coefficient=5 , giou_loss_coefficient=2 , eos_coefficient=0.4 , class_cost=1 , bbox_cost=5 , giou_cost=2 , ) if "detection" in checkpoint_url: _snake_case : Any = 15 _snake_case : int = 2 _snake_case : Optional[Any] = {0: "table", 1: "table rotated"} _snake_case : Union[str, Any] = idalabel _snake_case : Union[str, Any] = {v: k for k, v in idalabel.items()} else: _snake_case : Any = 125 _snake_case : Union[str, Any] = 6 _snake_case : List[str] = { 0: "table", 1: "table column", 2: "table row", 3: "table column header", 4: "table projected row header", 5: "table spanning cell", } _snake_case : Any = idalabel _snake_case : Optional[int] = {v: k for k, v in idalabel.items()} _snake_case : Union[str, Any] = DetrImageProcessor( format="coco_detection" , max_size=800 if "detection" in checkpoint_url else 1_000 ) _snake_case : str = TableTransformerForObjectDetection(a ) model.load_state_dict(a ) model.eval() # verify our conversion _snake_case : Optional[int] = "example_pdf.png" if "detection" in checkpoint_url else "example_table.png" _snake_case : Optional[Any] = hf_hub_download(repo_id="nielsr/example-pdf" , repo_type="dataset" , filename=a ) _snake_case : Dict = Image.open(a ).convert("RGB" ) _snake_case : Union[str, Any] = normalize(resize(a , a ) ).unsqueeze(0 ) _snake_case : str = model(a ) if "detection" in checkpoint_url: _snake_case : int = (1, 15, 3) _snake_case : List[str] = torch.tensor( [[-6.7897, -16.9985, 6.7937], [-8.0186, -22.2192, 6.9677], [-7.3117, -21.0708, 7.4055]] ) _snake_case : List[str] = torch.tensor([[0.4867, 0.1767, 0.6732], [0.6718, 0.4479, 0.3830], [0.4716, 0.1760, 0.6364]] ) else: _snake_case : Union[str, Any] = (1, 125, 7) _snake_case : str = torch.tensor( [[-18.1430, -8.3214, 4.8274], [-18.4685, -7.1361, -4.2667], [-26.3693, -9.3429, -4.9962]] ) _snake_case : Optional[Any] = torch.tensor([[0.4983, 0.5595, 0.9440], [0.4916, 0.6315, 0.5954], [0.6108, 0.8637, 0.1135]] ) assert outputs.logits.shape == expected_shape assert torch.allclose(outputs.logits[0, :3, :3] , a , atol=1e-4 ) assert torch.allclose(outputs.pred_boxes[0, :3, :3] , a , atol=1e-4 ) print("Looks ok!" ) if pytorch_dump_folder_path is not None: # Save model and image processor logger.info(f'Saving PyTorch model and image processor to {pytorch_dump_folder_path}...' ) Path(a ).mkdir(exist_ok=a ) model.save_pretrained(a ) image_processor.save_pretrained(a ) if push_to_hub: # Push model to HF hub logger.info("Pushing model to the hub..." ) _snake_case : int = ( "microsoft/table-transformer-detection" if "detection" in checkpoint_url else "microsoft/table-transformer-structure-recognition" ) model.push_to_hub(a ) image_processor.push_to_hub(a ) if __name__ == "__main__": _a : Tuple = argparse.ArgumentParser() parser.add_argument( """--checkpoint_url""", default="""https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", type=str, choices=[ """https://pubtables1m.blob.core.windows.net/model/pubtables1m_detection_detr_r18.pth""", """https://pubtables1m.blob.core.windows.net/model/pubtables1m_structure_detr_r18.pth""", ], help="""URL of the Table Transformer checkpoint you'd like to convert.""", ) parser.add_argument( """--pytorch_dump_folder_path""", default=None, type=str, help="""Path to the folder to output PyTorch model.""" ) parser.add_argument( """--push_to_hub""", action="""store_true""", help="""Whether or not to push the converted model to the 🤗 hub.""" ) _a : Any = parser.parse_args() convert_table_transformer_checkpoint(args.checkpoint_url, args.pytorch_dump_folder_path, args.push_to_hub)
87
0
"""simple docstring""" def a__ ( a : int , a : int ): """simple docstring""" return int((input_a, input_a).count(1 ) != 0 ) def a__ ( ): """simple docstring""" assert or_gate(0 , 0 ) == 0 assert or_gate(0 , 1 ) == 1 assert or_gate(1 , 0 ) == 1 assert or_gate(1 , 1 ) == 1 if __name__ == "__main__": print(or_gate(0, 1)) print(or_gate(1, 0)) print(or_gate(0, 0)) print(or_gate(1, 1))
711
"""simple docstring""" from ...utils import ( OptionalDependencyNotAvailable, is_torch_available, is_transformers_available, is_transformers_version, ) try: if not (is_transformers_available() and is_torch_available() and is_transformers_version(""">=""", """4.25.0""")): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ...utils.dummy_torch_and_transformers_objects import UnCLIPImageVariationPipeline, UnCLIPPipeline else: from .pipeline_unclip import UnCLIPPipeline from .pipeline_unclip_image_variation import UnCLIPImageVariationPipeline from .text_proj import UnCLIPTextProjModel
87
0
import warnings from typing import List, Optional, Union from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding, PaddingStrategy, PreTokenizedInput, TextInput, TruncationStrategy from ...utils import TensorType class _UpperCAmelCase ( _snake_case): __lowercase : Union[str, Any] = ["""image_processor""", """tokenizer"""] __lowercase : Dict = """LayoutLMv2ImageProcessor""" __lowercase : List[Any] = ("""LayoutXLMTokenizer""", """LayoutXLMTokenizerFast""") def __init__( self , snake_case_=None , snake_case_=None , **snake_case_ ): if "feature_extractor" in kwargs: warnings.warn( "The `feature_extractor` argument is deprecated and will be removed in v5, use `image_processor`" " instead." , snake_case_ , ) _snake_case : List[str] = kwargs.pop("feature_extractor" ) _snake_case : Tuple = image_processor if image_processor is not None else feature_extractor if image_processor is None: raise ValueError("You need to specify an `image_processor`." ) if tokenizer is None: raise ValueError("You need to specify a `tokenizer`." ) super().__init__(snake_case_ , snake_case_ ) def __call__( self , snake_case_ , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = True , snake_case_ = False , snake_case_ = None , snake_case_ = None , snake_case_ = 0 , snake_case_ = None , snake_case_ = None , snake_case_ = None , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = False , snake_case_ = True , snake_case_ = None , **snake_case_ , ): # verify input if self.image_processor.apply_ocr and (boxes is not None): raise ValueError( "You cannot provide bounding boxes " "if you initialized the image processor with apply_ocr set to True." ) if self.image_processor.apply_ocr and (word_labels is not None): raise ValueError( "You cannot provide word labels if you initialized the image processor with apply_ocr set to True." ) if return_overflowing_tokens is True and return_offsets_mapping is False: raise ValueError("You cannot return overflowing tokens without returning the offsets mapping." ) # first, apply the image processor _snake_case : List[Any] = self.image_processor(images=snake_case_ , return_tensors=snake_case_ ) # second, apply the tokenizer if text is not None and self.image_processor.apply_ocr and text_pair is None: if isinstance(snake_case_ , snake_case_ ): _snake_case : Dict = [text] # add batch dimension (as the image processor always adds a batch dimension) _snake_case : Optional[Any] = features["words"] _snake_case : List[Any] = self.tokenizer( text=text if text is not None else features["words"] , text_pair=text_pair if text_pair is not None else None , boxes=boxes if boxes is not None else features["boxes"] , word_labels=snake_case_ , add_special_tokens=snake_case_ , padding=snake_case_ , truncation=snake_case_ , max_length=snake_case_ , stride=snake_case_ , pad_to_multiple_of=snake_case_ , return_token_type_ids=snake_case_ , return_attention_mask=snake_case_ , return_overflowing_tokens=snake_case_ , return_special_tokens_mask=snake_case_ , return_offsets_mapping=snake_case_ , return_length=snake_case_ , verbose=snake_case_ , return_tensors=snake_case_ , **snake_case_ , ) # add pixel values _snake_case : int = features.pop("pixel_values" ) if return_overflowing_tokens is True: _snake_case : int = self.get_overflowing_images(snake_case_ , encoded_inputs["overflow_to_sample_mapping"] ) _snake_case : Optional[int] = images return encoded_inputs def lowerCamelCase__ ( self , snake_case_ , snake_case_ ): # in case there's an overflow, ensure each `input_ids` sample is mapped to its corresponding image _snake_case : Tuple = [] for sample_idx in overflow_to_sample_mapping: images_with_overflow.append(images[sample_idx] ) if len(snake_case_ ) != len(snake_case_ ): raise ValueError( "Expected length of images to be the same as the length of `overflow_to_sample_mapping`, but got" F' {len(snake_case_ )} and {len(snake_case_ )}' ) return images_with_overflow def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.batch_decode(*snake_case_ , **snake_case_ ) def lowerCamelCase__ ( self , *snake_case_ , **snake_case_ ): return self.tokenizer.decode(*snake_case_ , **snake_case_ ) @property def lowerCamelCase__ ( self ): return ["input_ids", "bbox", "attention_mask", "image"] @property def lowerCamelCase__ ( self ): warnings.warn( "`feature_extractor_class` is deprecated and will be removed in v5. Use `image_processor_class` instead." , snake_case_ , ) return self.image_processor_class @property def lowerCamelCase__ ( self ): warnings.warn( "`feature_extractor` is deprecated and will be removed in v5. Use `image_processor` instead." , snake_case_ , ) return self.image_processor
712
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices _a : Optional[int] = logging.get_logger(__name__) _a : List[str] = { """facebook/convnextv2-tiny-1k-224""": """https://huggingface.co/facebook/convnextv2-tiny-1k-224/resolve/main/config.json""", } class _UpperCAmelCase ( _snake_case , _snake_case): __lowercase : List[Any] = """convnextv2""" def __init__( self , snake_case_=3 , snake_case_=4 , snake_case_=4 , snake_case_=None , snake_case_=None , snake_case_="gelu" , snake_case_=0.02 , snake_case_=1E-12 , snake_case_=0.0 , snake_case_=2_24 , snake_case_=None , snake_case_=None , **snake_case_ , ): super().__init__(**snake_case_ ) _snake_case : Tuple = num_channels _snake_case : Optional[int] = patch_size _snake_case : Tuple = num_stages _snake_case : int = [96, 1_92, 3_84, 7_68] if hidden_sizes is None else hidden_sizes _snake_case : str = [3, 3, 9, 3] if depths is None else depths _snake_case : int = hidden_act _snake_case : Tuple = initializer_range _snake_case : Union[str, Any] = layer_norm_eps _snake_case : Optional[int] = drop_path_rate _snake_case : Union[str, Any] = image_size _snake_case : List[Any] = ["stem"] + [F'stage{idx}' for idx in range(1 , len(self.depths ) + 1 )] _snake_case , _snake_case : Dict = get_aligned_output_features_output_indices( out_features=snake_case_ , out_indices=snake_case_ , stage_names=self.stage_names )
87
0
"""simple docstring""" import collections import gzip import os import urllib import numpy from tensorflow.python.framework import dtypes, random_seed from tensorflow.python.platform import gfile from tensorflow.python.util.deprecation import deprecated _a : List[Any] = collections.namedtuple("""_Datasets""", ["""train""", """validation""", """test"""]) # CVDF mirror of http://yann.lecun.com/exdb/mnist/ _a : Union[str, Any] = """https://storage.googleapis.com/cvdf-datasets/mnist/""" def a__ ( a : Dict ): """simple docstring""" _snake_case : Any = numpy.dtype(numpy.uintaa ).newbyteorder(">" ) return numpy.frombuffer(bytestream.read(4 ) , dtype=a )[0] @deprecated(a , "Please use tf.data to implement this functionality." ) def a__ ( a : Any ): """simple docstring""" print("Extracting" , f.name ) with gzip.GzipFile(fileobj=a ) as bytestream: _snake_case : Any = _readaa(a ) if magic != 2_051: raise ValueError( "Invalid magic number %d in MNIST image file: %s" % (magic, f.name) ) _snake_case : Optional[int] = _readaa(a ) _snake_case : Union[str, Any] = _readaa(a ) _snake_case : str = _readaa(a ) _snake_case : List[str] = bytestream.read(rows * cols * num_images ) _snake_case : List[str] = numpy.frombuffer(a , dtype=numpy.uinta ) _snake_case : Any = data.reshape(a , a , a , 1 ) return data @deprecated(a , "Please use tf.one_hot on tensors." ) def a__ ( a : Dict , a : Optional[Any] ): """simple docstring""" _snake_case : List[Any] = labels_dense.shape[0] _snake_case : Optional[Any] = numpy.arange(a ) * num_classes _snake_case : int = numpy.zeros((num_labels, num_classes) ) _snake_case : List[str] = 1 return labels_one_hot @deprecated(a , "Please use tf.data to implement this functionality." ) def a__ ( a : Optional[int] , a : Any=False , a : List[str]=10 ): """simple docstring""" print("Extracting" , f.name ) with gzip.GzipFile(fileobj=a ) as bytestream: _snake_case : Tuple = _readaa(a ) if magic != 2_049: raise ValueError( "Invalid magic number %d in MNIST label file: %s" % (magic, f.name) ) _snake_case : int = _readaa(a ) _snake_case : Tuple = bytestream.read(a ) _snake_case : Dict = numpy.frombuffer(a , dtype=numpy.uinta ) if one_hot: return _dense_to_one_hot(a , a ) return labels class _UpperCAmelCase : @deprecated( snake_case_ , "Please use alternatives such as official/mnist/_DataSet.py" " from tensorflow/models." , ) def __init__( self , snake_case_ , snake_case_ , snake_case_=False , snake_case_=False , snake_case_=dtypes.floataa , snake_case_=True , snake_case_=None , ): _snake_case : Tuple = random_seed.get_seed(snake_case_ ) # If op level seed is not set, use whatever graph level seed is returned numpy.random.seed(seeda if seed is None else seeda ) _snake_case : Optional[int] = dtypes.as_dtype(snake_case_ ).base_dtype if dtype not in (dtypes.uinta, dtypes.floataa): raise TypeError("Invalid image dtype %r, expected uint8 or float32" % dtype ) if fake_data: _snake_case : List[Any] = 1_00_00 _snake_case : Optional[Any] = one_hot else: assert ( images.shape[0] == labels.shape[0] ), F'images.shape: {images.shape} labels.shape: {labels.shape}' _snake_case : List[Any] = images.shape[0] # Convert shape from [num examples, rows, columns, depth] # to [num examples, rows*columns] (assuming depth == 1) if reshape: assert images.shape[3] == 1 _snake_case : Any = images.reshape( images.shape[0] , images.shape[1] * images.shape[2] ) if dtype == dtypes.floataa: # Convert from [0, 255] -> [0.0, 1.0]. _snake_case : Dict = images.astype(numpy.floataa ) _snake_case : Optional[Any] = numpy.multiply(snake_case_ , 1.0 / 255.0 ) _snake_case : Optional[int] = images _snake_case : List[str] = labels _snake_case : Any = 0 _snake_case : Tuple = 0 @property def lowerCamelCase__ ( self ): return self._images @property def lowerCamelCase__ ( self ): return self._labels @property def lowerCamelCase__ ( self ): return self._num_examples @property def lowerCamelCase__ ( self ): return self._epochs_completed def lowerCamelCase__ ( self , snake_case_ , snake_case_=False , snake_case_=True ): if fake_data: _snake_case : Union[str, Any] = [1] * 7_84 _snake_case : Dict = [1] + [0] * 9 if self.one_hot else 0 return ( [fake_image for _ in range(snake_case_ )], [fake_label for _ in range(snake_case_ )], ) _snake_case : Optional[Any] = self._index_in_epoch # Shuffle for the first epoch if self._epochs_completed == 0 and start == 0 and shuffle: _snake_case : List[Any] = numpy.arange(self._num_examples ) numpy.random.shuffle(snake_case_ ) _snake_case : Dict = self.images[perma] _snake_case : Union[str, Any] = self.labels[perma] # Go to the next epoch if start + batch_size > self._num_examples: # Finished epoch self._epochs_completed += 1 # Get the rest examples in this epoch _snake_case : str = self._num_examples - start _snake_case : Optional[Any] = self._images[start : self._num_examples] _snake_case : Union[str, Any] = self._labels[start : self._num_examples] # Shuffle the data if shuffle: _snake_case : Optional[Any] = numpy.arange(self._num_examples ) numpy.random.shuffle(snake_case_ ) _snake_case : Union[str, Any] = self.images[perm] _snake_case : Dict = self.labels[perm] # Start next epoch _snake_case : Union[str, Any] = 0 _snake_case : int = batch_size - rest_num_examples _snake_case : Dict = self._index_in_epoch _snake_case : List[Any] = self._images[start:end] _snake_case : Dict = self._labels[start:end] return ( numpy.concatenate((images_rest_part, images_new_part) , axis=0 ), numpy.concatenate((labels_rest_part, labels_new_part) , axis=0 ), ) else: self._index_in_epoch += batch_size _snake_case : Optional[int] = self._index_in_epoch return self._images[start:end], self._labels[start:end] @deprecated(a , "Please write your own downloading logic." ) def a__ ( a : List[Any] , a : int , a : Any ): """simple docstring""" if not gfile.Exists(a ): gfile.MakeDirs(a ) _snake_case : Optional[int] = os.path.join(a , a ) if not gfile.Exists(a ): urllib.request.urlretrieve(a , a ) # noqa: S310 with gfile.GFile(a ) as f: _snake_case : Any = f.size() print("Successfully downloaded" , a , a , "bytes." ) return filepath @deprecated( a , "Please use alternatives such as:" " tensorflow_datasets.load('mnist')" ) def a__ ( a : Optional[Any] , a : Optional[Any]=False , a : Tuple=False , a : str=dtypes.floataa , a : Union[str, Any]=True , a : Union[str, Any]=5_000 , a : Union[str, Any]=None , a : Optional[int]=DEFAULT_SOURCE_URL , ): """simple docstring""" if fake_data: def fake(): return _DataSet( [] , [] , fake_data=a , one_hot=a , dtype=a , seed=a ) _snake_case : int = fake() _snake_case : Any = fake() _snake_case : Optional[int] = fake() return _Datasets(train=a , validation=a , test=a ) if not source_url: # empty string check _snake_case : Optional[int] = DEFAULT_SOURCE_URL _snake_case : Tuple = "train-images-idx3-ubyte.gz" _snake_case : int = "train-labels-idx1-ubyte.gz" _snake_case : str = "t10k-images-idx3-ubyte.gz" _snake_case : int = "t10k-labels-idx1-ubyte.gz" _snake_case : Optional[int] = _maybe_download( a , a , source_url + train_images_file ) with gfile.Open(a , "rb" ) as f: _snake_case : List[str] = _extract_images(a ) _snake_case : Union[str, Any] = _maybe_download( a , a , source_url + train_labels_file ) with gfile.Open(a , "rb" ) as f: _snake_case : str = _extract_labels(a , one_hot=a ) _snake_case : List[Any] = _maybe_download( a , a , source_url + test_images_file ) with gfile.Open(a , "rb" ) as f: _snake_case : Optional[Any] = _extract_images(a ) _snake_case : str = _maybe_download( a , a , source_url + test_labels_file ) with gfile.Open(a , "rb" ) as f: _snake_case : Union[str, Any] = _extract_labels(a , one_hot=a ) if not 0 <= validation_size <= len(a ): _snake_case : Optional[int] = ( "Validation size should be between 0 and " f'{len(a )}. Received: {validation_size}.' ) raise ValueError(a ) _snake_case : Any = train_images[:validation_size] _snake_case : List[Any] = train_labels[:validation_size] _snake_case : Dict = train_images[validation_size:] _snake_case : int = train_labels[validation_size:] _snake_case : List[Any] = {"dtype": dtype, "reshape": reshape, "seed": seed} _snake_case : Tuple = _DataSet(a , a , **a ) _snake_case : Tuple = _DataSet(a , a , **a ) _snake_case : Tuple = _DataSet(a , a , **a ) return _Datasets(train=a , validation=a , test=a )
713
"""simple docstring""" from argparse import ArgumentParser, Namespace from ..utils import logging from . import BaseTransformersCLICommand def a__ ( a : Namespace ): """simple docstring""" return ConvertCommand( args.model_type , args.tf_checkpoint , args.pytorch_dump_output , args.config , args.finetuning_task_name ) _a : int = """ transformers can only be used from the commandline to convert TensorFlow models in PyTorch, In that case, it requires TensorFlow to be installed. Please see https://www.tensorflow.org/install/ for installation instructions. """ class _UpperCAmelCase ( _snake_case): @staticmethod def lowerCamelCase__ ( snake_case_ ): _snake_case : Dict = parser.add_parser( "convert" , help="CLI tool to run convert model from original author checkpoints to Transformers PyTorch checkpoints." , ) train_parser.add_argument("--model_type" , type=snake_case_ , required=snake_case_ , help="Model's type." ) train_parser.add_argument( "--tf_checkpoint" , type=snake_case_ , required=snake_case_ , help="TensorFlow checkpoint path or folder." ) train_parser.add_argument( "--pytorch_dump_output" , type=snake_case_ , required=snake_case_ , help="Path to the PyTorch saved model output." ) train_parser.add_argument("--config" , type=snake_case_ , default="" , help="Configuration file path or folder." ) train_parser.add_argument( "--finetuning_task_name" , type=snake_case_ , default=snake_case_ , help="Optional fine-tuning task name if the TF model was a finetuned model." , ) train_parser.set_defaults(func=snake_case_ ) def __init__( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , *snake_case_ , ): _snake_case : str = logging.get_logger("transformers-cli/converting" ) self._logger.info(F'Loading model {model_type}' ) _snake_case : Optional[int] = model_type _snake_case : Any = tf_checkpoint _snake_case : Optional[int] = pytorch_dump_output _snake_case : Tuple = config _snake_case : Tuple = finetuning_task_name def lowerCamelCase__ ( self ): if self._model_type == "albert": try: from ..models.albert.convert_albert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "bert": try: from ..models.bert.convert_bert_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "funnel": try: from ..models.funnel.convert_funnel_original_tf_checkpoint_to_pytorch import ( convert_tf_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "t5": try: from ..models.ta.convert_ta_original_tf_checkpoint_to_pytorch import convert_tf_checkpoint_to_pytorch except ImportError: raise ImportError(snake_case_ ) convert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "gpt": from ..models.openai.convert_openai_original_tf_checkpoint_to_pytorch import ( convert_openai_checkpoint_to_pytorch, ) convert_openai_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "transfo_xl": try: from ..models.transfo_xl.convert_transfo_xl_original_tf_checkpoint_to_pytorch import ( convert_transfo_xl_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) if "ckpt" in self._tf_checkpoint.lower(): _snake_case : int = self._tf_checkpoint _snake_case : Optional[Any] = "" else: _snake_case : Optional[int] = self._tf_checkpoint _snake_case : List[str] = "" convert_transfo_xl_checkpoint_to_pytorch( snake_case_ , self._config , self._pytorch_dump_output , snake_case_ ) elif self._model_type == "gpt2": try: from ..models.gpta.convert_gpta_original_tf_checkpoint_to_pytorch import ( convert_gpta_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_gpta_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) elif self._model_type == "xlnet": try: from ..models.xlnet.convert_xlnet_original_tf_checkpoint_to_pytorch import ( convert_xlnet_checkpoint_to_pytorch, ) except ImportError: raise ImportError(snake_case_ ) convert_xlnet_checkpoint_to_pytorch( self._tf_checkpoint , self._config , self._pytorch_dump_output , self._finetuning_task_name ) elif self._model_type == "xlm": from ..models.xlm.convert_xlm_original_pytorch_checkpoint_to_pytorch import ( convert_xlm_checkpoint_to_pytorch, ) convert_xlm_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "lxmert": from ..models.lxmert.convert_lxmert_original_tf_checkpoint_to_pytorch import ( convert_lxmert_checkpoint_to_pytorch, ) convert_lxmert_checkpoint_to_pytorch(self._tf_checkpoint , self._pytorch_dump_output ) elif self._model_type == "rembert": from ..models.rembert.convert_rembert_tf_checkpoint_to_pytorch import ( convert_rembert_tf_checkpoint_to_pytorch, ) convert_rembert_tf_checkpoint_to_pytorch(self._tf_checkpoint , self._config , self._pytorch_dump_output ) else: raise ValueError( "--model_type should be selected in the list [bert, gpt, gpt2, t5, transfo_xl, xlnet, xlm, lxmert]" )
87
0
import json import os import tempfile import unittest import numpy as np from datasets import load_dataset from transformers.testing_utils import require_torch, require_vision, slow from transformers.utils import is_torch_available, is_vision_available from ...test_image_processing_common import ImageProcessingSavingTestMixin if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import ImageGPTImageProcessor class _UpperCAmelCase ( unittest.TestCase): def __init__( self , snake_case_ , snake_case_=7 , snake_case_=3 , snake_case_=18 , snake_case_=30 , snake_case_=4_00 , snake_case_=True , snake_case_=None , snake_case_=True , ): _snake_case : Optional[Any] = size if size is not None else {"height": 18, "width": 18} _snake_case : Optional[int] = parent _snake_case : Dict = batch_size _snake_case : Optional[Any] = num_channels _snake_case : List[str] = image_size _snake_case : Optional[int] = min_resolution _snake_case : str = max_resolution _snake_case : Dict = do_resize _snake_case : Optional[int] = size _snake_case : List[str] = do_normalize def lowerCamelCase__ ( self ): return { # here we create 2 clusters for the sake of simplicity "clusters": np.asarray( [ [0.8866443634033203, 0.6618829369544983, 0.3891746401786804], [-0.6042559146881104, -0.02295008860528469, 0.5423797369003296], ] ), "do_resize": self.do_resize, "size": self.size, "do_normalize": self.do_normalize, } @require_torch @require_vision class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : List[Any] = ImageGPTImageProcessor if is_vision_available() else None def lowerCamelCase__ ( self ): _snake_case : List[str] = ImageGPTImageProcessingTester(self ) @property def lowerCamelCase__ ( self ): return self.image_processor_tester.prepare_image_processor_dict() def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(snake_case_ , "clusters" ) ) self.assertTrue(hasattr(snake_case_ , "do_resize" ) ) self.assertTrue(hasattr(snake_case_ , "size" ) ) self.assertTrue(hasattr(snake_case_ , "do_normalize" ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 18, "width": 18} ) _snake_case : Union[str, Any] = self.image_processing_class.from_dict(self.image_processor_dict , size=42 ) self.assertEqual(image_processor.size , {"height": 42, "width": 42} ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.image_processing_class(**self.image_processor_dict ) _snake_case : Dict = json.loads(image_processor.to_json_string() ) for key, value in self.image_processor_dict.items(): if key == "clusters": self.assertTrue(np.array_equal(snake_case_ , obj[key] ) ) else: self.assertEqual(obj[key] , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: _snake_case : List[Any] = os.path.join(snake_case_ , "image_processor.json" ) image_processor_first.to_json_file(snake_case_ ) _snake_case : str = self.image_processing_class.from_json_file(snake_case_ ).to_dict() _snake_case : Optional[int] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(snake_case_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.image_processing_class(**self.image_processor_dict ) with tempfile.TemporaryDirectory() as tmpdirname: image_processor_first.save_pretrained(snake_case_ ) _snake_case : Any = self.image_processing_class.from_pretrained(snake_case_ ).to_dict() _snake_case : List[str] = image_processor_first.to_dict() for key, value in image_processor_first.items(): if key == "clusters": self.assertTrue(np.array_equal(snake_case_ , image_processor_second[key] ) ) else: self.assertEqual(image_processor_first[key] , snake_case_ ) @unittest.skip("ImageGPT requires clusters at initialization" ) def lowerCamelCase__ ( self ): pass def a__ ( ): """simple docstring""" _snake_case : Optional[int] = load_dataset("hf-internal-testing/fixtures_image_utils" , split="test" ) _snake_case : List[str] = Image.open(dataset[4]["file"] ) _snake_case : List[str] = Image.open(dataset[5]["file"] ) _snake_case : Union[str, Any] = [imagea, imagea] return images @require_vision @require_torch class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = ImageGPTImageProcessor.from_pretrained("openai/imagegpt-small" ) _snake_case : Any = prepare_images() # test non-batched _snake_case : int = image_processing(images[0] , return_tensors="pt" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (1, 10_24) ) _snake_case : Optional[Any] = [3_06, 1_91, 1_91] self.assertEqual(encoding.input_ids[0, :3].tolist() , snake_case_ ) # test batched _snake_case : Dict = image_processing(snake_case_ , return_tensors="pt" ) self.assertIsInstance(encoding.input_ids , torch.LongTensor ) self.assertEqual(encoding.input_ids.shape , (2, 10_24) ) _snake_case : Any = [3_03, 13, 13] self.assertEqual(encoding.input_ids[1, -3:].tolist() , snake_case_ )
714
"""simple docstring""" import argparse import json import os import tensorstore as ts import torch from flax import serialization from flax.traverse_util import flatten_dict, unflatten_dict from tensorflow.io import gfile from transformers.modeling_utils import dtype_byte_size from transformers.models.switch_transformers.convert_switch_transformers_original_flax_checkpoint_to_pytorch import ( rename_keys, ) from transformers.utils import WEIGHTS_INDEX_NAME, WEIGHTS_NAME from transformers.utils.hub import convert_file_size_to_int def a__ ( a : List[str] , a : Any ): """simple docstring""" if flax_key_tuple[-1] == "kernel" and flax_tensor.ndim == 3: # expert layer _snake_case : Any = flax_key_tuple[:-1] + ("weight",) _snake_case : str = torch.permute(a , (0, 2, 1) ) elif flax_key_tuple[-1] == "kernel" and ".".join(a ): # linear layer _snake_case : Optional[int] = flax_key_tuple[:-1] + ("weight",) _snake_case : Any = flax_tensor.T elif flax_key_tuple[-1] in ["scale", "embedding"]: _snake_case : Union[str, Any] = flax_key_tuple[:-1] + ("weight",) return flax_key_tuple, flax_tensor def a__ ( a : List[Any] , a : Union[str, Any] , a : List[str] ): """simple docstring""" if "metadata" in layer: _snake_case : Optional[int] = layer.split("metadata" ) _snake_case : Optional[int] = "".join(split_layer[0] )[:-1] _snake_case : int = [tuple(("metadata" + split_layer[1]).split("/" ) )] elif "kvstore" in layer: _snake_case : Any = layer.split("kvstore" ) _snake_case : str = "".join(split_layer[0] )[:-1] _snake_case : Any = [tuple(("kvstore" + split_layer[1]).split("/" ) )] else: _snake_case : List[Any] = layer.split("/" ) _snake_case : Tuple = "/".join(split_layer[:-1] ) _snake_case : int = (split_layer[-1],) if "kvstore/path" in layer: _snake_case : Optional[Any] = f'{switch_checkpoint_path}/{checkpoint_info[layer]}' elif "kvstore/driver" in layer: _snake_case : Tuple = "file" else: _snake_case : Optional[int] = checkpoint_info[layer] return curr_real_layer_name, split_layer, content def a__ ( a : List[Any] , a : List[Any] ): """simple docstring""" _snake_case : Union[str, Any] = rename_keys(a ) _snake_case : int = {} for k, v in current_block.items(): _snake_case : Optional[int] = v _snake_case : Optional[int] = new_current_block torch.save(a , a ) def a__ ( a : Dict , a : Tuple , a : List[str] , a : int , a : str = WEIGHTS_NAME ): """simple docstring""" _snake_case : Any = convert_file_size_to_int(a ) _snake_case : Tuple = [] _snake_case : Optional[int] = {} _snake_case : Tuple = 0 _snake_case : Optional[Any] = 0 os.makedirs(a , exist_ok=a ) with gfile.GFile(switch_checkpoint_path + "/checkpoint" , "rb" ) as fp: _snake_case : Any = serialization.msgpack_restore(fp.read() )["optimizer"]["target"] _snake_case : Optional[Any] = flatten_dict(a , sep="/" ) _snake_case : Optional[Any] = {} for layer in checkpoint_info.keys(): _snake_case , _snake_case , _snake_case : int = get_key_and_tensorstore_dict( a , a , a ) if curr_real_layer_name in all_layers: _snake_case : Dict = content else: _snake_case : Tuple = {split_layer[-1]: content} for key in all_layers.keys(): # open tensorstore file _snake_case : List[str] = ts.open(unflatten_dict(all_layers[key] ) ).result().read().result() _snake_case : Dict = torch.tensor(a ) _snake_case : Dict = raw_weights.numel() * dtype_byte_size(raw_weights.dtype ) # use the renaming pattern from the small conversion scripts _snake_case , _snake_case : Optional[int] = rename_base_flax_keys(tuple(key.split("/" ) ) , a ) _snake_case : Optional[Any] = "/".join(a ) # If this weight is going to tip up over the maximal size, we split. if current_block_size + weight_size > max_shard_size: _snake_case : Any = os.path.join( a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) del current_block _snake_case : List[Any] = {} _snake_case : str = 0 _snake_case : List[str] = raw_weights.to(getattr(a , a ) ) current_block_size += weight_size total_size += weight_size # Add the last block _snake_case : int = os.path.join(a , weights_name.replace(".bin" , f'-{len(a )+1:05d}-of-???.bin' ) ) rename_and_save_block(a , a ) sharded_state_dicts.append(current_block.keys() ) # If we only have one shard, we return it if len(a ) == 1: return {weights_name: sharded_state_dicts[0]}, None # Otherwise, let's build the index _snake_case : str = {} _snake_case : Any = {} for idx, shard in enumerate(a ): _snake_case : Optional[int] = weights_name.replace( ".bin" , f'-{idx+1:05d}-of-{len(a ):05d}.bin' ) # len(sharded_state_dicts):05d} _snake_case : Dict = os.path.join(a , weights_name.replace(".bin" , f'-{idx+1:05d}-of-???.bin' ) ) os.rename(a , os.path.join(a , a ) ) _snake_case : Dict = shard for key in shard: _snake_case : int = shard_file # Add the metadata _snake_case : List[Any] = {"total_size": total_size} _snake_case : Any = {"metadata": metadata, "weight_map": weight_map} with open(os.path.join(a , a ) , "w" , encoding="utf-8" ) as f: _snake_case : Union[str, Any] = json.dumps(a , indent=2 , sort_keys=a ) + "\n" f.write(a ) return metadata, index if __name__ == "__main__": _a : Dict = argparse.ArgumentParser() # Required parameters parser.add_argument( """--switch_t5x_checkpoint_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128/checkpoint_634600""", type=str, required=False, help="""Path to a directory containing a folder per layer. Follows the original Google format.""", ) parser.add_argument("""--max_shard_size""", default="""10GB""", required=False, help="""Max shard size""") parser.add_argument("""--dtype""", default="""bfloat16""", type=str, required=False, help="""dtype of the saved model""") parser.add_argument( """--pytorch_dump_folder_path""", default="""/mnt/disks/disk_switch/original_checkpoints/switch-xxl-128-converted""", type=str, required=False, help="""Path to the output pytorch model.""", ) _a : Optional[int] = parser.parse_args() shard_on_the_fly( args.switch_tax_checkpoint_path, args.pytorch_dump_folder_path, args.max_shard_size, args.dtype, ) def a__ ( ): """simple docstring""" from transformers import SwitchTransformersConfig, SwitchTransformersForConditionalGeneration, TaTokenizer _snake_case : List[str] = SwitchTransformersConfig.from_pretrained("google/switch-base-8" ) config.save_pretrained("/home/arthur_huggingface_co/transformers/switch_converted" ) _snake_case : str = SwitchTransformersForConditionalGeneration.from_pretrained( "/home/arthur_huggingface_co/transformers/switch_converted" , device_map="auto" ) _snake_case : List[Any] = TaTokenizer.from_pretrained("t5-small" ) _snake_case : Optional[Any] = "A <extra_id_0> walks into a bar a orders a <extra_id_1> with <extra_id_2> pinch of <extra_id_3>." _snake_case : Dict = tokenizer(a , return_tensors="pt" ).input_ids _snake_case : List[Any] = model.generate(a , decoder_start_token_id=0 ) print(tokenizer.decode(out[0] ) )
87
0
"""simple docstring""" from __future__ import annotations from math import pi from typing import Protocol import matplotlib.pyplot as plt import numpy as np class _UpperCAmelCase ( _snake_case): def lowerCamelCase__ ( self , snake_case_ ): return 0.0 def a__ ( a : np.ndarray , a : int ): """simple docstring""" _snake_case : Any = min([-20, np.min(fft_results[1 : samplerate // 2 - 1] )] ) _snake_case : Union[str, Any] = max([20, np.max(fft_results[1 : samplerate // 2 - 1] )] ) return lowest, highest def a__ ( a : FilterType , a : int ): """simple docstring""" _snake_case : List[Any] = 512 _snake_case : Dict = [1] + [0] * (size - 1) _snake_case : Union[str, Any] = [filter_type.process(a ) for item in inputs] _snake_case : List[str] = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case : List[Any] = np.abs(np.fft.fft(a ) ) _snake_case : List[Any] = 20 * np.logaa(a ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("Frequency (Hz)" ) plt.xscale("log" ) # Display within reasonable bounds _snake_case : Tuple = get_bounds(a , a ) plt.ylim(max([-80, bounds[0]] ) , min([80, bounds[1]] ) ) plt.ylabel("Gain (dB)" ) plt.plot(a ) plt.show() def a__ ( a : FilterType , a : int ): """simple docstring""" _snake_case : str = 512 _snake_case : Optional[Any] = [1] + [0] * (size - 1) _snake_case : Optional[Any] = [filter_type.process(a ) for item in inputs] _snake_case : Union[str, Any] = [0] * (samplerate - size) # zero-padding outputs += filler _snake_case : Dict = np.angle(np.fft.fft(a ) ) # Frequencies on log scale from 24 to nyquist frequency plt.xlim(24 , samplerate / 2 - 1 ) plt.xlabel("Frequency (Hz)" ) plt.xscale("log" ) plt.ylim(-2 * pi , 2 * pi ) plt.ylabel("Phase shift (Radians)" ) plt.plot(np.unwrap(a , -2 * pi ) ) plt.show()
715
"""simple docstring""" from __future__ import annotations import unittest from transformers import MobileBertConfig, is_tf_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_MODEL_FOR_PRETRAINING_MAPPING, TFMobileBertForMaskedLM, TFMobileBertForMultipleChoice, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertModel, ) @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( TFMobileBertModel, TFMobileBertForMaskedLM, TFMobileBertForNextSentencePrediction, TFMobileBertForPreTraining, TFMobileBertForQuestionAnswering, TFMobileBertForSequenceClassification, TFMobileBertForTokenClassification, TFMobileBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : Optional[Any] = ( { """feature-extraction""": TFMobileBertModel, """fill-mask""": TFMobileBertForMaskedLM, """question-answering""": TFMobileBertForQuestionAnswering, """text-classification""": TFMobileBertForSequenceClassification, """token-classification""": TFMobileBertForTokenClassification, """zero-shot""": TFMobileBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_=False ): _snake_case : Union[str, Any] = super()._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) if return_labels: if model_class in get_values(snake_case_ ): _snake_case : Any = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) return inputs_dict class _UpperCAmelCase ( _snake_case): def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : Optional[Any] = parent _snake_case : List[Any] = batch_size _snake_case : Optional[int] = seq_length _snake_case : Dict = is_training _snake_case : Union[str, Any] = use_input_mask _snake_case : List[Any] = use_token_type_ids _snake_case : int = use_labels _snake_case : Dict = vocab_size _snake_case : Tuple = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Optional[Any] = intermediate_size _snake_case : Dict = hidden_act _snake_case : Tuple = hidden_dropout_prob _snake_case : Union[str, Any] = attention_probs_dropout_prob _snake_case : str = max_position_embeddings _snake_case : str = type_vocab_size _snake_case : Any = type_sequence_label_size _snake_case : Optional[int] = initializer_range _snake_case : List[Any] = num_labels _snake_case : Optional[int] = num_choices _snake_case : Optional[int] = scope _snake_case : Any = embedding_size def lowerCamelCase__ ( self ): _snake_case : Optional[int] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : Optional[Any] = None if self.use_input_mask: _snake_case : Optional[Any] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : List[str] = None if self.use_token_type_ids: _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : Dict = None _snake_case : Tuple = None _snake_case : str = None if self.use_labels: _snake_case : Union[str, Any] = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Any = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[int] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Tuple = MobileBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , embedding_size=self.embedding_size , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFMobileBertModel(config=snake_case_ ) _snake_case : Optional[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Optional[Any] = model(snake_case_ ) _snake_case : Dict = model(snake_case_ ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) self.parent.assertEqual(result.pooler_output.shape , (self.batch_size, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFMobileBertForMaskedLM(config=snake_case_ ) _snake_case : Tuple = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = TFMobileBertForNextSentencePrediction(config=snake_case_ ) _snake_case : Dict = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Tuple = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = TFMobileBertForPreTraining(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual( result.prediction_logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) self.parent.assertEqual(result.seq_relationship_logits.shape , (self.batch_size, 2) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : str = self.num_labels _snake_case : str = TFMobileBertForSequenceClassification(config=snake_case_ ) _snake_case : Union[str, Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Optional[int] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Any = self.num_choices _snake_case : Tuple = TFMobileBertForMultipleChoice(config=snake_case_ ) _snake_case : List[Any] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Tuple = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = self.num_labels _snake_case : Optional[int] = TFMobileBertForTokenClassification(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : List[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : int = TFMobileBertForQuestionAnswering(config=snake_case_ ) _snake_case : str = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.prepare_config_and_inputs() ( ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ( _snake_case ) , ) : Tuple = config_and_inputs _snake_case : Tuple = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : int = TFMobileBertModelTest.TFMobileBertModelTester(self ) _snake_case : Optional[Any] = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : int = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_next_sequence_prediction(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : str = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_pretraining(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_mobilebert_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): # for model_name in TF_MOBILEBERT_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: for model_name in ["google/mobilebert-uncased"]: _snake_case : str = TFMobileBertModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFMobileBertForPreTraining.from_pretrained("google/mobilebert-uncased" ) _snake_case : str = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : Union[str, Any] = model(snake_case_ )[0] _snake_case : int = [1, 6, 3_05_22] self.assertEqual(output.shape , snake_case_ ) _snake_case : Optional[Any] = tf.constant( [ [ [-4.5919547, -9.248295, -9.645256], [-6.7306175, -6.440284, -6.6052837], [-7.2743506, -6.7847915, -6.024673], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
87
0
"""simple docstring""" from ...configuration_utils import PretrainedConfig from ...utils import logging _a : str = logging.get_logger(__name__) _a : Tuple = { """EleutherAI/gpt-neox-20b""": """https://huggingface.co/EleutherAI/gpt-neox-20b/resolve/main/config.json""", # See all GPTNeoX models at https://huggingface.co/models?filter=gpt_neox } class _UpperCAmelCase ( _snake_case): __lowercase : int = """gpt_neox""" def __init__( self , snake_case_=5_04_32 , snake_case_=61_44 , snake_case_=44 , snake_case_=64 , snake_case_=2_45_76 , snake_case_="gelu" , snake_case_=0.25 , snake_case_=1_00_00 , snake_case_=0.0 , snake_case_=0.0 , snake_case_=0.1 , snake_case_=20_48 , snake_case_=0.02 , snake_case_=1E-5 , snake_case_=True , snake_case_=0 , snake_case_=2 , snake_case_=False , snake_case_=True , snake_case_=None , **snake_case_ , ): super().__init__(bos_token_id=snake_case_ , eos_token_id=snake_case_ , **snake_case_ ) _snake_case : str = vocab_size _snake_case : Tuple = max_position_embeddings _snake_case : str = hidden_size _snake_case : Optional[int] = num_hidden_layers _snake_case : List[str] = num_attention_heads _snake_case : Tuple = intermediate_size _snake_case : Tuple = hidden_act _snake_case : Dict = rotary_pct _snake_case : List[Any] = rotary_emb_base _snake_case : Union[str, Any] = attention_dropout _snake_case : int = hidden_dropout _snake_case : List[Any] = classifier_dropout _snake_case : Any = initializer_range _snake_case : int = layer_norm_eps _snake_case : List[Any] = use_cache _snake_case : int = tie_word_embeddings _snake_case : Any = use_parallel_residual _snake_case : str = rope_scaling self._rope_scaling_validation() if self.hidden_size % self.num_attention_heads != 0: raise ValueError( "The hidden size is not divisble by the number of attention heads! Make sure to update them!" ) def lowerCamelCase__ ( self ): if self.rope_scaling is None: return if not isinstance(self.rope_scaling , snake_case_ ) or len(self.rope_scaling ) != 2: raise ValueError( "`rope_scaling` must be a dictionary with with two fields, `name` and `factor`, " F'got {self.rope_scaling}' ) _snake_case : List[Any] = self.rope_scaling.get("type" , snake_case_ ) _snake_case : List[str] = self.rope_scaling.get("factor" , snake_case_ ) if rope_scaling_type is None or rope_scaling_type not in ["linear", "dynamic"]: raise ValueError( F'`rope_scaling`\'s name field must be one of [\'linear\', \'dynamic\'], got {rope_scaling_type}' ) if rope_scaling_factor is None or not isinstance(snake_case_ , snake_case_ ) or rope_scaling_factor <= 1.0: raise ValueError(F'`rope_scaling`\'s factor field must be an float > 1, got {rope_scaling_factor}' )
716
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available _a : List[Any] = {} try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""BartphoTokenizer"""] if TYPE_CHECKING: try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_bartpho import BartphoTokenizer else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=7 , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=True , snake_case_=99 , snake_case_=32 , snake_case_=2 , snake_case_=4 , snake_case_=37 , snake_case_="gelu" , snake_case_=0.1 , snake_case_=0.1 , snake_case_=5_12 , snake_case_=16 , snake_case_=2 , snake_case_=0.02 , snake_case_=3 , snake_case_=4 , snake_case_=None , ): _snake_case : str = parent _snake_case : List[Any] = 13 _snake_case : Any = 7 _snake_case : Union[str, Any] = True _snake_case : str = True _snake_case : Optional[int] = True _snake_case : str = True _snake_case : List[str] = 99 _snake_case : int = 3_84 _snake_case : Optional[int] = 2 _snake_case : Optional[Any] = 4 _snake_case : Optional[int] = 37 _snake_case : int = "gelu" _snake_case : Any = 0.1 _snake_case : Any = 0.1 _snake_case : Tuple = 5_12 _snake_case : Optional[int] = 16 _snake_case : Tuple = 2 _snake_case : Tuple = 0.02 _snake_case : str = 3 _snake_case : Any = 4 _snake_case : str = 1_28 _snake_case : Union[str, Any] = 2 _snake_case : Optional[int] = 9 _snake_case : List[str] = 1 _snake_case : List[str] = None def lowerCamelCase__ ( self ): _snake_case : List[Any] = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) _snake_case : int = None if self.use_input_mask: _snake_case : List[str] = random_attention_mask([self.batch_size, self.seq_length] ) _snake_case : Union[str, Any] = None if self.use_token_type_ids: _snake_case : Dict = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) _snake_case : List[str] = None _snake_case : Any = None _snake_case : Optional[int] = None if self.use_labels: _snake_case : Dict = ids_tensor([self.batch_size] , self.type_sequence_label_size ) _snake_case : Union[str, Any] = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) _snake_case : Optional[Any] = ids_tensor([self.batch_size] , self.num_choices ) _snake_case : Optional[int] = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=snake_case_ , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = TFConvBertModel(config=snake_case_ ) _snake_case : List[Any] = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} _snake_case : Union[str, Any] = [input_ids, input_mask] _snake_case : Dict = model(snake_case_ ) _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[Any] = TFConvBertForMaskedLM(config=snake_case_ ) _snake_case : str = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _snake_case : int = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[Any] = self.num_labels _snake_case : Any = TFConvBertForSequenceClassification(config=snake_case_ ) _snake_case : int = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _snake_case : int = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Dict = self.num_choices _snake_case : int = TFConvBertForMultipleChoice(config=snake_case_ ) _snake_case : Any = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : int = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : List[str] = tf.tile(tf.expand_dims(snake_case_ , 1 ) , (1, self.num_choices, 1) ) _snake_case : Union[str, Any] = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } _snake_case : Optional[Any] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[Any] = self.num_labels _snake_case : Any = TFConvBertForTokenClassification(config=snake_case_ ) _snake_case : List[str] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _snake_case : List[str] = model(snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ , snake_case_ ): _snake_case : List[str] = TFConvBertForQuestionAnswering(config=snake_case_ ) _snake_case : Optional[Any] = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } _snake_case : int = model(snake_case_ ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = self.prepare_config_and_inputs() ( _snake_case ) : Tuple = config_and_inputs _snake_case : Optional[int] = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Optional[int] = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __lowercase : List[Any] = ( { """feature-extraction""": TFConvBertModel, """fill-mask""": TFConvBertForMaskedLM, """question-answering""": TFConvBertForQuestionAnswering, """text-classification""": TFConvBertForSequenceClassification, """token-classification""": TFConvBertForTokenClassification, """zero-shot""": TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __lowercase : Optional[int] = False __lowercase : Tuple = False __lowercase : Optional[Any] = False def lowerCamelCase__ ( self ): _snake_case : List[Any] = TFConvBertModelTester(self ) _snake_case : Dict = ConfigTester(self , config_class=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.config_tester.run_common_tests() def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Dict = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : List[str] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Any = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): _snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : List[Any] = True _snake_case : Dict = True if hasattr(snake_case_ , "use_cache" ): _snake_case : Dict = True _snake_case : str = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _snake_case : List[Any] = getattr(self.model_tester , "key_length" , snake_case_ ) for model_class in self.all_model_classes: _snake_case : Optional[Any] = self._prepare_for_class(snake_case_ , snake_case_ ) _snake_case : str = model_class(snake_case_ ) _snake_case : List[str] = len(model(snake_case_ ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(snake_case_ , saved_model=snake_case_ ) _snake_case : List[str] = os.path.join(snake_case_ , "saved_model" , "1" ) _snake_case : Dict = tf.keras.models.load_model(snake_case_ ) _snake_case : str = model(snake_case_ ) if self.is_encoder_decoder: _snake_case : str = outputs["encoder_hidden_states"] _snake_case : List[Any] = outputs["encoder_attentions"] else: _snake_case : int = outputs["hidden_states"] _snake_case : List[Any] = outputs["attentions"] self.assertEqual(len(snake_case_ ) , snake_case_ ) _snake_case : Dict = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(snake_case_ ) , snake_case_ ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowerCamelCase__ ( self ): _snake_case : Any = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs_for_common() _snake_case : Optional[Any] = True _snake_case : Tuple = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) _snake_case : Any = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) _snake_case : List[str] = getattr(self.model_tester , "key_length" , snake_case_ ) _snake_case : List[str] = getattr(self.model_tester , "key_length" , snake_case_ ) def check_decoder_attentions_output(snake_case_ ): _snake_case : Optional[int] = len(snake_case_ ) self.assertEqual(out_len % 2 , 0 ) _snake_case : List[str] = outputs.decoder_attentions self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(snake_case_ ): _snake_case : Optional[Any] = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(snake_case_ ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: _snake_case : str = True _snake_case : Tuple = False _snake_case : List[Any] = model_class(snake_case_ ) _snake_case : List[str] = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) _snake_case : Optional[int] = len(snake_case_ ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) if self.is_encoder_decoder: _snake_case : Optional[Any] = model_class(snake_case_ ) _snake_case : str = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_decoder_attentions_output(snake_case_ ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] _snake_case : List[Any] = True _snake_case : Dict = model_class(snake_case_ ) _snake_case : Tuple = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) # Check attention is always last and order is fine _snake_case : str = True _snake_case : int = True _snake_case : Optional[int] = model_class(snake_case_ ) _snake_case : List[str] = model(self._prepare_for_class(snake_case_ , snake_case_ ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(snake_case_ ) ) self.assertEqual(model.config.output_hidden_states , snake_case_ ) check_encoder_attentions_output(snake_case_ ) @require_tf class _UpperCAmelCase ( unittest.TestCase): @slow def lowerCamelCase__ ( self ): _snake_case : str = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) _snake_case : Optional[int] = tf.constant([[0, 1, 2, 3, 4, 5]] ) _snake_case : int = model(snake_case_ )[0] _snake_case : List[str] = [1, 6, 7_68] self.assertEqual(output.shape , snake_case_ ) _snake_case : List[Any] = tf.constant( [ [ [-0.03475493, -0.4686034, -0.30638832], [0.22637248, -0.26988646, -0.7423424], [0.10324868, -0.45013508, -0.58280784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , snake_case_ , atol=1E-4 )
717
"""simple docstring""" def a__ ( a : list , a : int , a : int = 0 , a : int = 0 ): """simple docstring""" _snake_case : Optional[int] = right or len(a ) - 1 if left > right: return -1 elif list_data[left] == key: return left elif list_data[right] == key: return right else: return search(a , a , left + 1 , right - 1 ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" from dataclasses import dataclass from typing import Optional, Tuple, Union import torch import torch.nn as nn from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput from .embeddings import GaussianFourierProjection, TimestepEmbedding, Timesteps from .modeling_utils import ModelMixin from .unet_ad_blocks import get_down_block, get_mid_block, get_out_block, get_up_block @dataclass class _UpperCAmelCase ( _snake_case): __lowercase : torch.FloatTensor class _UpperCAmelCase ( _snake_case , _snake_case): @register_to_config def __init__( self , snake_case_ = 6_55_36 , snake_case_ = None , snake_case_ = 2 , snake_case_ = 2 , snake_case_ = 0 , snake_case_ = "fourier" , snake_case_ = True , snake_case_ = False , snake_case_ = 0.0 , snake_case_ = ("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , snake_case_ = ("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , snake_case_ = "UNetMidBlock1D" , snake_case_ = None , snake_case_ = (32, 32, 64) , snake_case_ = None , snake_case_ = 8 , snake_case_ = 1 , snake_case_ = False , ): super().__init__() _snake_case : Optional[Any] = sample_size # time if time_embedding_type == "fourier": _snake_case : List[Any] = GaussianFourierProjection( embedding_size=8 , set_W_to_weight=snake_case_ , log=snake_case_ , flip_sin_to_cos=snake_case_ ) _snake_case : int = 2 * block_out_channels[0] elif time_embedding_type == "positional": _snake_case : Optional[Any] = Timesteps( block_out_channels[0] , flip_sin_to_cos=snake_case_ , downscale_freq_shift=snake_case_ ) _snake_case : List[Any] = block_out_channels[0] if use_timestep_embedding: _snake_case : Dict = block_out_channels[0] * 4 _snake_case : str = TimestepEmbedding( in_channels=snake_case_ , time_embed_dim=snake_case_ , act_fn=snake_case_ , out_dim=block_out_channels[0] , ) _snake_case : int = nn.ModuleList([] ) _snake_case : Any = None _snake_case : str = nn.ModuleList([] ) _snake_case : Tuple = None # down _snake_case : str = in_channels for i, down_block_type in enumerate(snake_case_ ): _snake_case : Optional[int] = output_channel _snake_case : List[Any] = block_out_channels[i] if i == 0: input_channel += extra_in_channels _snake_case : Union[str, Any] = i == len(snake_case_ ) - 1 _snake_case : Optional[Any] = get_down_block( snake_case_ , num_layers=snake_case_ , in_channels=snake_case_ , out_channels=snake_case_ , temb_channels=block_out_channels[0] , add_downsample=not is_final_block or downsample_each_block , ) self.down_blocks.append(snake_case_ ) # mid _snake_case : Any = get_mid_block( snake_case_ , in_channels=block_out_channels[-1] , mid_channels=block_out_channels[-1] , out_channels=block_out_channels[-1] , embed_dim=block_out_channels[0] , num_layers=snake_case_ , add_downsample=snake_case_ , ) # up _snake_case : int = list(reversed(snake_case_ ) ) _snake_case : List[Any] = reversed_block_out_channels[0] if out_block_type is None: _snake_case : Optional[Any] = out_channels else: _snake_case : List[str] = block_out_channels[0] for i, up_block_type in enumerate(snake_case_ ): _snake_case : Any = output_channel _snake_case : Tuple = ( reversed_block_out_channels[i + 1] if i < len(snake_case_ ) - 1 else final_upsample_channels ) _snake_case : str = i == len(snake_case_ ) - 1 _snake_case : Union[str, Any] = get_up_block( snake_case_ , num_layers=snake_case_ , in_channels=snake_case_ , out_channels=snake_case_ , temb_channels=block_out_channels[0] , add_upsample=not is_final_block , ) self.up_blocks.append(snake_case_ ) _snake_case : Tuple = output_channel # out _snake_case : List[str] = norm_num_groups if norm_num_groups is not None else min(block_out_channels[0] // 4 , 32 ) _snake_case : Any = get_out_block( out_block_type=snake_case_ , num_groups_out=snake_case_ , embed_dim=block_out_channels[0] , out_channels=snake_case_ , act_fn=snake_case_ , fc_dim=block_out_channels[-1] // 4 , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ = True , ): _snake_case : Optional[int] = timestep if not torch.is_tensor(snake_case_ ): _snake_case : List[Any] = torch.tensor([timesteps] , dtype=torch.long , device=sample.device ) elif torch.is_tensor(snake_case_ ) and len(timesteps.shape ) == 0: _snake_case : str = timesteps[None].to(sample.device ) _snake_case : Tuple = self.time_proj(snake_case_ ) if self.config.use_timestep_embedding: _snake_case : Any = self.time_mlp(snake_case_ ) else: _snake_case : Tuple = timestep_embed[..., None] _snake_case : Any = timestep_embed.repeat([1, 1, sample.shape[2]] ).to(sample.dtype ) _snake_case : int = timestep_embed.broadcast_to((sample.shape[:1] + timestep_embed.shape[1:]) ) # 2. down _snake_case : Optional[Any] = () for downsample_block in self.down_blocks: _snake_case : str = downsample_block(hidden_states=snake_case_ , temb=snake_case_ ) down_block_res_samples += res_samples # 3. mid if self.mid_block: _snake_case : List[str] = self.mid_block(snake_case_ , snake_case_ ) # 4. up for i, upsample_block in enumerate(self.up_blocks ): _snake_case : Any = down_block_res_samples[-1:] _snake_case : Dict = down_block_res_samples[:-1] _snake_case : Dict = upsample_block(snake_case_ , res_hidden_states_tuple=snake_case_ , temb=snake_case_ ) # 5. post-process if self.out_block: _snake_case : List[Any] = self.out_block(snake_case_ , snake_case_ ) if not return_dict: return (sample,) return UNetaDOutput(sample=snake_case_ )
718
"""simple docstring""" from __future__ import annotations class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_ ): _snake_case , _snake_case : Dict = text, pattern _snake_case , _snake_case : int = len(snake_case_ ), len(snake_case_ ) def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if char == self.pattern[i]: return i return -1 def lowerCamelCase__ ( self , snake_case_ ): for i in range(self.patLen - 1 , -1 , -1 ): if self.pattern[i] != self.text[current_pos + i]: return current_pos + i return -1 def lowerCamelCase__ ( self ): # searches pattern in text and returns index positions _snake_case : List[str] = [] for i in range(self.textLen - self.patLen + 1 ): _snake_case : Union[str, Any] = self.mismatch_in_text(snake_case_ ) if mismatch_index == -1: positions.append(snake_case_ ) else: _snake_case : Tuple = self.match_in_pattern(self.text[mismatch_index] ) _snake_case : Tuple = ( mismatch_index - match_index ) # shifting index lgtm [py/multiple-definition] return positions _a : List[Any] = """ABAABA""" _a : str = """AB""" _a : List[Any] = BoyerMooreSearch(text, pattern) _a : Any = bms.bad_character_heuristic() if len(positions) == 0: print("""No match found""") else: print("""Pattern found in following positions: """) print(positions)
87
0
"""simple docstring""" import inspect import unittest from transformers import ConvNextVaConfig from transformers.models.auto import get_values from transformers.models.auto.modeling_auto import MODEL_FOR_BACKBONE_MAPPING_NAMES, MODEL_MAPPING_NAMES from transformers.testing_utils import require_torch, require_vision, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_vision_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from transformers import ConvNextVaBackbone, ConvNextVaForImageClassification, ConvNextVaModel from transformers.models.convnextva.modeling_convnextva import CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class _UpperCAmelCase : def __init__( self , snake_case_ , snake_case_=13 , snake_case_=32 , snake_case_=3 , snake_case_=4 , snake_case_=[10, 20, 30, 40] , snake_case_=[2, 2, 3, 2] , snake_case_=True , snake_case_=True , snake_case_=37 , snake_case_="gelu" , snake_case_=10 , snake_case_=0.02 , snake_case_=["stage2", "stage3", "stage4"] , snake_case_=[2, 3, 4] , snake_case_=None , ): _snake_case : Dict = parent _snake_case : Tuple = batch_size _snake_case : Optional[int] = image_size _snake_case : Dict = num_channels _snake_case : List[Any] = num_stages _snake_case : str = hidden_sizes _snake_case : Union[str, Any] = depths _snake_case : int = is_training _snake_case : Optional[Any] = use_labels _snake_case : Tuple = intermediate_size _snake_case : Tuple = hidden_act _snake_case : Tuple = num_labels _snake_case : Optional[Any] = initializer_range _snake_case : int = out_features _snake_case : Union[str, Any] = out_indices _snake_case : Union[str, Any] = scope def lowerCamelCase__ ( self ): _snake_case : Tuple = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) _snake_case : Optional[int] = None if self.use_labels: _snake_case : List[Any] = ids_tensor([self.batch_size] , self.num_labels ) _snake_case : Optional[Any] = self.get_config() return config, pixel_values, labels def lowerCamelCase__ ( self ): return ConvNextVaConfig( num_channels=self.num_channels , hidden_sizes=self.hidden_sizes , depths=self.depths , num_stages=self.num_stages , hidden_act=self.hidden_act , is_decoder=snake_case_ , initializer_range=self.initializer_range , out_features=self.out_features , out_indices=self.out_indices , num_labels=self.num_labels , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Union[str, Any] = ConvNextVaModel(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Optional[Any] = model(snake_case_ ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 32, self.image_size // 32) , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[int] = ConvNextVaForImageClassification(snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : str = model(snake_case_ , labels=snake_case_ ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ ): _snake_case : Tuple = ConvNextVaBackbone(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : Optional[Any] = model(snake_case_ ) # verify hidden states self.parent.assertEqual(len(result.feature_maps ) , len(config.out_features ) ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[1], 4, 4] ) # verify channels self.parent.assertEqual(len(model.channels ) , len(config.out_features ) ) self.parent.assertListEqual(model.channels , config.hidden_sizes[1:] ) # verify backbone works with out_features=None _snake_case : Any = None _snake_case : str = ConvNextVaBackbone(config=snake_case_ ) model.to(snake_case_ ) model.eval() _snake_case : str = model(snake_case_ ) # verify feature maps self.parent.assertEqual(len(result.feature_maps ) , 1 ) self.parent.assertListEqual(list(result.feature_maps[0].shape ) , [self.batch_size, self.hidden_sizes[-1], 1, 1] ) # verify channels self.parent.assertEqual(len(model.channels ) , 1 ) self.parent.assertListEqual(model.channels , [config.hidden_sizes[-1]] ) def lowerCamelCase__ ( self ): _snake_case : Tuple = self.prepare_config_and_inputs() _snake_case : str = config_and_inputs _snake_case : Dict = {"pixel_values": pixel_values} return config, inputs_dict def lowerCamelCase__ ( self ): _snake_case : List[str] = self.prepare_config_and_inputs() _snake_case : Optional[int] = config_and_inputs _snake_case : Union[str, Any] = {"pixel_values": pixel_values, "labels": labels} return config, inputs_dict @require_torch class _UpperCAmelCase ( _snake_case , _snake_case , unittest.TestCase): __lowercase : Dict = ( ( ConvNextVaModel, ConvNextVaForImageClassification, ConvNextVaBackbone, ) if is_torch_available() else () ) __lowercase : str = ( {"""feature-extraction""": ConvNextVaModel, """image-classification""": ConvNextVaForImageClassification} if is_torch_available() else {} ) __lowercase : Union[str, Any] = False __lowercase : str = False __lowercase : Tuple = False __lowercase : Union[str, Any] = False __lowercase : Optional[int] = False def lowerCamelCase__ ( self ): _snake_case : Any = ConvNextVaModelTester(self ) _snake_case : List[str] = ConfigTester(self , config_class=snake_case_ , has_text_modality=snake_case_ , hidden_size=37 ) def lowerCamelCase__ ( self ): self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowerCamelCase__ ( self ): return @unittest.skip(reason="ConvNextV2 does not use inputs_embeds" ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="ConvNextV2 does not support input and output embeddings" ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="ConvNextV2 does not use feedforward chunking" ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): if not self.model_tester.is_training: return for model_class in self.all_model_classes: _snake_case : List[Any] = self.model_tester.prepare_config_and_inputs_with_labels() _snake_case : Tuple = True if model_class.__name__ in [ *get_values(snake_case_ ), *get_values(snake_case_ ), ]: continue _snake_case : str = model_class(snake_case_ ) model.to(snake_case_ ) model.train() _snake_case : List[Any] = self._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) _snake_case : List[str] = model(**snake_case_ ).loss loss.backward() def lowerCamelCase__ ( self ): if not self.model_tester.is_training: return for model_class in self.all_model_classes: _snake_case : str = self.model_tester.prepare_config_and_inputs_with_labels() _snake_case : List[str] = False _snake_case : Optional[int] = True if ( model_class.__name__ in [*get_values(snake_case_ ), *get_values(snake_case_ )] or not model_class.supports_gradient_checkpointing ): continue _snake_case : Optional[int] = model_class(snake_case_ ) model.to(snake_case_ ) model.gradient_checkpointing_enable() model.train() _snake_case : Optional[int] = self._prepare_for_class(snake_case_ , snake_case_ , return_labels=snake_case_ ) _snake_case : List[Any] = model(**snake_case_ ).loss loss.backward() def lowerCamelCase__ ( self ): _snake_case : List[str] = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Dict = model_class(snake_case_ ) _snake_case : Dict = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic _snake_case : List[Any] = [*signature.parameters.keys()] _snake_case : Optional[Any] = ["pixel_values"] self.assertListEqual(arg_names[:1] , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[Any] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*snake_case_ ) def lowerCamelCase__ ( self ): def check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ): _snake_case : Optional[Any] = model_class(snake_case_ ) model.to(snake_case_ ) model.eval() with torch.no_grad(): _snake_case : Tuple = model(**self._prepare_for_class(snake_case_ , snake_case_ ) ) _snake_case : Tuple = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states _snake_case : Union[str, Any] = self.model_tester.num_stages self.assertEqual(len(snake_case_ ) , expected_num_stages + 1 ) # ConvNextV2's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 4, self.model_tester.image_size // 4] , ) _snake_case : Any = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: _snake_case : Optional[int] = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] _snake_case : Optional[Any] = True check_hidden_states_output(snake_case_ , snake_case_ , snake_case_ ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*snake_case_ ) @slow def lowerCamelCase__ ( self ): for model_name in CONVNEXTV2_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: _snake_case : Dict = ConvNextVaModel.from_pretrained(snake_case_ ) self.assertIsNotNone(snake_case_ ) def a__ ( ): """simple docstring""" _snake_case : Optional[Any] = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class _UpperCAmelCase ( unittest.TestCase): @cached_property def lowerCamelCase__ ( self ): return AutoImageProcessor.from_pretrained("facebook/convnextv2-tiny-1k-224" ) if is_vision_available() else None @slow def lowerCamelCase__ ( self ): _snake_case : List[Any] = ConvNextVaForImageClassification.from_pretrained("facebook/convnextv2-tiny-1k-224" ).to(snake_case_ ) _snake_case : Any = self.default_image_processor _snake_case : List[str] = prepare_img() _snake_case : Tuple = preprocessor(images=snake_case_ , return_tensors="pt" ).to(snake_case_ ) # forward pass with torch.no_grad(): _snake_case : Union[str, Any] = model(**snake_case_ ) # verify the logits _snake_case : str = torch.Size((1, 10_00) ) self.assertEqual(outputs.logits.shape , snake_case_ ) _snake_case : Any = torch.tensor([0.9996, 0.1966, -0.4386] ).to(snake_case_ ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , snake_case_ , atol=1E-4 ) )
719
"""simple docstring""" from datetime import datetime import requests from bsa import BeautifulSoup if __name__ == "__main__": _a : Dict = input("""Enter image url: """).strip() print(f'Downloading image from {url} ...') _a : str = BeautifulSoup(requests.get(url).content, """html.parser""") # The image URL is in the content field of the first meta tag with property og:image _a : str = soup.find("""meta""", {"""property""": """og:image"""})["""content"""] _a : Dict = requests.get(image_url).content _a : str = f'{datetime.now():%Y-%m-%d_%H:%M:%S}.jpg' with open(file_name, """wb""") as fp: fp.write(image_data) print(f'Done. Image saved to disk as {file_name}.')
87
0
"""simple docstring""" from math import atan, cos, radians, sin, tan from .haversine_distance import haversine_distance _a : List[Any] = 6_378_137.0 _a : int = 6_356_752.314_245 _a : int = 6_378_137 def a__ ( a : float , a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[int] = (AXIS_A - AXIS_B) / AXIS_A # Parametric latitudes # https://en.wikipedia.org/wiki/Latitude#Parametric_(or_reduced)_latitude _snake_case : int = atan((1 - flattening) * tan(radians(a ) ) ) _snake_case : Dict = atan((1 - flattening) * tan(radians(a ) ) ) # Compute central angle between two points # using haversine theta. sigma = haversine_distance / equatorial radius _snake_case : List[Any] = haversine_distance(a , a , a , a ) / EQUATORIAL_RADIUS # Intermediate P and Q values _snake_case : Tuple = (b_lata + b_lata) / 2 _snake_case : List[Any] = (b_lata - b_lata) / 2 # Intermediate X value # X = (sigma - sin(sigma)) * sin^2Pcos^2Q / cos^2(sigma/2) _snake_case : Optional[Any] = (sin(a ) ** 2) * (cos(a ) ** 2) _snake_case : Optional[int] = cos(sigma / 2 ) ** 2 _snake_case : Any = (sigma - sin(a )) * (x_numerator / x_demonimator) # Intermediate Y value # Y = (sigma + sin(sigma)) * cos^2Psin^2Q / sin^2(sigma/2) _snake_case : Optional[Any] = (cos(a ) ** 2) * (sin(a ) ** 2) _snake_case : str = sin(sigma / 2 ) ** 2 _snake_case : List[str] = (sigma + sin(a )) * (y_numerator / y_denominator) return EQUATORIAL_RADIUS * (sigma - ((flattening / 2) * (x_value + y_value))) if __name__ == "__main__": import doctest doctest.testmod()
720
"""simple docstring""" from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available, is_vision_available _a : Optional[int] = { """configuration_pix2struct""": [ """PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP""", """Pix2StructConfig""", """Pix2StructTextConfig""", """Pix2StructVisionConfig""", ], """processing_pix2struct""": ["""Pix2StructProcessor"""], } try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : List[Any] = ["""Pix2StructImageProcessor"""] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: _a : Dict = [ """PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST""", """Pix2StructPreTrainedModel""", """Pix2StructForConditionalGeneration""", """Pix2StructVisionModel""", """Pix2StructTextModel""", ] if TYPE_CHECKING: from .configuration_pixastruct import ( PIX2STRUCT_PRETRAINED_CONFIG_ARCHIVE_MAP, PixaStructConfig, PixaStructTextConfig, PixaStructVisionConfig, ) from .processing_pixastruct import PixaStructProcessor try: if not is_vision_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .image_processing_pixastruct import PixaStructImageProcessor try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_pixastruct import ( PIX2STRUCT_PRETRAINED_MODEL_ARCHIVE_LIST, PixaStructForConditionalGeneration, PixaStructPreTrainedModel, PixaStructTextModel, PixaStructVisionModel, ) else: import sys _a : Optional[int] = _LazyModule(__name__, globals()["""__file__"""], _import_structure, module_spec=__spec__)
87
0
"""simple docstring""" import math from collections.abc import Iterator from itertools import takewhile def a__ ( a : int ): """simple docstring""" if 1 < number < 4: # 2 and 3 are primes return True elif number < 2 or number % 2 == 0 or number % 3 == 0: # Negatives, 0, 1, all even numbers, all multiples of 3 are not primes return False # All primes number are in format of 6k +/- 1 for i in range(5 , int(math.sqrt(a ) + 1 ) , 6 ): if number % i == 0 or number % (i + 2) == 0: return False return True def a__ ( ): """simple docstring""" _snake_case : List[str] = 2 while True: if is_prime(a ): yield num num += 1 def a__ ( a : int = 2_000_000 ): """simple docstring""" return sum(takewhile(lambda a : x < n , prime_generator() ) ) if __name__ == "__main__": print(f'{solution() = }')
721
"""simple docstring""" import argparse import json import subprocess def a__ ( a : Optional[Any] , a : Optional[int] ): """simple docstring""" _snake_case : str = [] _snake_case : Optional[Any] = ( f'curl -H "Accept: application/vnd.github+json" -H "Authorization: Bearer {token}"' " https://api.github.com/repos/huggingface/transformers/actions/runners" ) _snake_case : Dict = subprocess.run(a , shell=a , stdout=subprocess.PIPE ) _snake_case : Tuple = output.stdout.decode("utf-8" ) _snake_case : List[str] = json.loads(a ) _snake_case : Any = status["runners"] for runner in runners: if runner["name"] in target_runners: if runner["status"] == "offline": offline_runners.append(a ) # save the result so we can report them on Slack with open("offline_runners.txt" , "w" ) as fp: fp.write(json.dumps(a ) ) if len(a ) > 0: _snake_case : Any = "\n".join([x["name"] for x in offline_runners] ) raise ValueError(f'The following runners are offline:\n{failed}' ) if __name__ == "__main__": def a__ ( a : Optional[int] ): """simple docstring""" return values.split("," ) _a : Optional[int] = argparse.ArgumentParser() # Required parameters parser.add_argument( """--target_runners""", default=None, type=list_str, required=True, help="""Comma-separated list of runners to check status.""", ) parser.add_argument( """--token""", default=None, type=str, required=True, help="""A token that has actions:read permission.""" ) _a : List[str] = parser.parse_args() get_runner_status(args.target_runners, args.token)
87
0
import os import sys import warnings from dataclasses import dataclass, field from io import BytesIO from typing import TYPE_CHECKING, Any, ClassVar, Dict, List, Optional, Union import numpy as np import pyarrow as pa from .. import config from ..download.streaming_download_manager import xopen from ..table import array_cast from ..utils.file_utils import is_local_path from ..utils.py_utils import first_non_null_value, no_op_if_value_is_null, string_to_dict if TYPE_CHECKING: import PIL.Image from .features import FeatureType _a : Optional[List[str]] = None _a : Optional[Any] = """<""" if sys.byteorder == """little""" else """>""" # Origin: https://github.com/python-pillow/Pillow/blob/698951e19e19972aeed56df686868f1329981c12/src/PIL/Image.py#L3126 minus "|i1" which values are not preserved correctly when saving and loading an image _a : str = [ np.dtype("""|b1"""), np.dtype("""|u1"""), np.dtype("""<u2"""), np.dtype(""">u2"""), np.dtype("""<i2"""), np.dtype(""">i2"""), np.dtype("""<u4"""), np.dtype(""">u4"""), np.dtype("""<i4"""), np.dtype(""">i4"""), np.dtype("""<f4"""), np.dtype(""">f4"""), np.dtype("""<f8"""), np.dtype(""">f8"""), ] @dataclass class _UpperCAmelCase : __lowercase : bool = True __lowercase : Optional[str] = None # Automatically constructed __lowercase : ClassVar[str] = "PIL.Image.Image" __lowercase : ClassVar[Any] = pa.struct({"""bytes""": pa.binary(), """path""": pa.string()}) __lowercase : str = field(default="""Image""" , init=_snake_case , repr=_snake_case) def __call__( self ): return self.pa_type def lowerCamelCase__ ( self , snake_case_ ): if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if isinstance(snake_case_ , snake_case_ ): _snake_case : int = np.array(snake_case_ ) if isinstance(snake_case_ , snake_case_ ): return {"path": value, "bytes": None} elif isinstance(snake_case_ , snake_case_ ): return {"path": None, "bytes": value} elif isinstance(snake_case_ , np.ndarray ): # convert the image array to PNG/TIFF bytes return encode_np_array(snake_case_ ) elif isinstance(snake_case_ , PIL.Image.Image ): # convert the PIL image to bytes (default format is PNG/TIFF) return encode_pil_image(snake_case_ ) elif value.get("path" ) is not None and os.path.isfile(value["path"] ): # we set "bytes": None to not duplicate the data if they're already available locally return {"bytes": None, "path": value.get("path" )} elif value.get("bytes" ) is not None or value.get("path" ) is not None: # store the image bytes, and path is used to infer the image format using the file extension return {"bytes": value.get("bytes" ), "path": value.get("path" )} else: raise ValueError( F'An image sample should have one of \'path\' or \'bytes\' but they are missing or None in {value}.' ) def lowerCamelCase__ ( self , snake_case_ , snake_case_=None ): if not self.decode: raise RuntimeError("Decoding is disabled for this feature. Please use Image(decode=True) instead." ) if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support decoding images, please install 'Pillow'." ) if token_per_repo_id is None: _snake_case : Dict = {} _snake_case : int = value["path"], value["bytes"] if bytes_ is None: if path is None: raise ValueError(F'An image should have one of \'path\' or \'bytes\' but both are None in {value}.' ) else: if is_local_path(snake_case_ ): _snake_case : Dict = PIL.Image.open(snake_case_ ) else: _snake_case : List[str] = path.split("::" )[-1] try: _snake_case : List[Any] = string_to_dict(snake_case_ , config.HUB_DATASETS_URL )["repo_id"] _snake_case : Union[str, Any] = token_per_repo_id.get(snake_case_ ) except ValueError: _snake_case : str = None with xopen(snake_case_ , "rb" , use_auth_token=snake_case_ ) as f: _snake_case : Tuple = BytesIO(f.read() ) _snake_case : Dict = PIL.Image.open(bytes_ ) else: _snake_case : int = PIL.Image.open(BytesIO(bytes_ ) ) image.load() # to avoid "Too many open files" errors return image def lowerCamelCase__ ( self ): from .features import Value return ( self if self.decode else { "bytes": Value("binary" ), "path": Value("string" ), } ) def lowerCamelCase__ ( self , snake_case_ ): if pa.types.is_string(storage.type ): _snake_case : int = pa.array([None] * len(snake_case_ ) , type=pa.binary() ) _snake_case : str = pa.StructArray.from_arrays([bytes_array, storage] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_binary(storage.type ): _snake_case : Dict = pa.array([None] * len(snake_case_ ) , type=pa.string() ) _snake_case : int = pa.StructArray.from_arrays([storage, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_struct(storage.type ): if storage.type.get_field_index("bytes" ) >= 0: _snake_case : Optional[int] = storage.field("bytes" ) else: _snake_case : str = pa.array([None] * len(snake_case_ ) , type=pa.binary() ) if storage.type.get_field_index("path" ) >= 0: _snake_case : Tuple = storage.field("path" ) else: _snake_case : str = pa.array([None] * len(snake_case_ ) , type=pa.string() ) _snake_case : int = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=storage.is_null() ) elif pa.types.is_list(storage.type ): _snake_case : List[Any] = pa.array( [encode_np_array(np.array(snake_case_ ) )["bytes"] if arr is not None else None for arr in storage.to_pylist()] , type=pa.binary() , ) _snake_case : int = pa.array([None] * len(snake_case_ ) , type=pa.string() ) _snake_case : str = pa.StructArray.from_arrays( [bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(snake_case_ , self.pa_type ) def lowerCamelCase__ ( self , snake_case_ ): @no_op_if_value_is_null def path_to_bytes(snake_case_ ): with xopen(snake_case_ , "rb" ) as f: _snake_case : Optional[Any] = f.read() return bytes_ _snake_case : Optional[Any] = pa.array( [ (path_to_bytes(x["path"] ) if x["bytes"] is None else x["bytes"]) if x is not None else None for x in storage.to_pylist() ] , type=pa.binary() , ) _snake_case : Tuple = pa.array( [os.path.basename(snake_case_ ) if path is not None else None for path in storage.field("path" ).to_pylist()] , type=pa.string() , ) _snake_case : Tuple = pa.StructArray.from_arrays([bytes_array, path_array] , ["bytes", "path"] , mask=bytes_array.is_null() ) return array_cast(snake_case_ , self.pa_type ) def a__ ( ): """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) global _IMAGE_COMPRESSION_FORMATS if _IMAGE_COMPRESSION_FORMATS is None: PIL.Image.init() _snake_case : Union[str, Any] = list(set(PIL.Image.OPEN.keys() ) & set(PIL.Image.SAVE.keys() ) ) return _IMAGE_COMPRESSION_FORMATS def a__ ( a : "PIL.Image.Image" ): """simple docstring""" _snake_case : Dict = BytesIO() if image.format in list_image_compression_formats(): _snake_case : Dict = image.format else: _snake_case : Optional[Any] = "PNG" if image.mode in ["1", "L", "LA", "RGB", "RGBA"] else "TIFF" image.save(a , format=a ) return buffer.getvalue() def a__ ( a : "PIL.Image.Image" ): """simple docstring""" if hasattr(a , "filename" ) and image.filename != "": return {"path": image.filename, "bytes": None} else: return {"path": None, "bytes": image_to_bytes(a )} def a__ ( a : np.ndarray ): """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) _snake_case : Tuple = array.dtype _snake_case : Optional[int] = dtype.byteorder if dtype.byteorder != "=" else _NATIVE_BYTEORDER _snake_case : Optional[Any] = dtype.kind _snake_case : Dict = dtype.itemsize _snake_case : Dict = None # Multi-channel array case (only np.dtype("|u1") is allowed) if array.shape[2:]: _snake_case : Dict = np.dtype("|u1" ) if dtype_kind not in ["u", "i"]: raise TypeError( f'Unsupported array dtype {dtype} for image encoding. Only {dest_dtype} is supported for multi-channel arrays.' ) if dtype is not dest_dtype: warnings.warn(f'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) # Exact match elif dtype in _VALID_IMAGE_ARRAY_DTPYES: _snake_case : Optional[Any] = dtype else: # Downcast the type within the kind (np.can_cast(from_type, to_type, casting="same_kind") doesn't behave as expected, so do it manually) while dtype_itemsize >= 1: _snake_case : str = dtype_byteorder + dtype_kind + str(a ) _snake_case : Optional[Any] = np.dtype(a ) if dest_dtype in _VALID_IMAGE_ARRAY_DTPYES: warnings.warn(f'Downcasting array dtype {dtype} to {dest_dtype} to be compatible with \'Pillow\'' ) break else: dtype_itemsize //= 2 if dest_dtype is None: raise TypeError( f'Cannot convert dtype {dtype} to a valid image dtype. Valid image dtypes: {_VALID_IMAGE_ARRAY_DTPYES}' ) _snake_case : Optional[int] = PIL.Image.fromarray(array.astype(a ) ) return {"path": None, "bytes": image_to_bytes(a )} def a__ ( a : Union[List[str], List[dict], List[np.ndarray], List["PIL.Image.Image"]] ): """simple docstring""" if config.PIL_AVAILABLE: import PIL.Image else: raise ImportError("To support encoding images, please install 'Pillow'." ) if objs: _snake_case : Tuple = first_non_null_value(a ) if isinstance(a , a ): return [{"path": obj, "bytes": None} if obj is not None else None for obj in objs] if isinstance(a , np.ndarray ): _snake_case : str = no_op_if_value_is_null(a ) return [obj_to_image_dict_func(a ) for obj in objs] elif isinstance(a , PIL.Image.Image ): _snake_case : Optional[Any] = no_op_if_value_is_null(a ) return [obj_to_image_dict_func(a ) for obj in objs] else: return objs else: return objs
700
"""simple docstring""" import unittest from .lib import ( Matrix, Vector, axpy, square_zero_matrix, unit_basis_vector, zero_vector, ) class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) self.assertEqual(x.component(0 ) , 1 ) self.assertEqual(x.component(2 ) , 3 ) _snake_case : List[Any] = Vector() def lowerCamelCase__ ( self ): _snake_case : Any = Vector([0, 0, 0, 0, 0, 1] ) self.assertEqual(str(snake_case_ ) , "(0,0,0,0,0,1)" ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 2, 3, 4] ) self.assertEqual(len(snake_case_ ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2] ) _snake_case : List[str] = Vector([1, 2, 3, 4, 5] ) _snake_case : List[Any] = Vector([0, 0, 0, 0, 0, 0, 0, 0, 0, 0] ) _snake_case : Any = Vector([1, -1, 1, -1, 2, -3, 4, -5] ) self.assertAlmostEqual(x.euclidean_length() , 2.236 , 3 ) self.assertAlmostEqual(y.euclidean_length() , 7.416 , 3 ) self.assertEqual(z.euclidean_length() , 0 ) self.assertAlmostEqual(w.euclidean_length() , 7.616 , 3 ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Vector([1, 2, 3] ) _snake_case : Any = Vector([1, 1, 1] ) self.assertEqual((x + y).component(0 ) , 2 ) self.assertEqual((x + y).component(1 ) , 3 ) self.assertEqual((x + y).component(2 ) , 4 ) def lowerCamelCase__ ( self ): _snake_case : str = Vector([1, 2, 3] ) _snake_case : Union[str, Any] = Vector([1, 1, 1] ) self.assertEqual((x - y).component(0 ) , 0 ) self.assertEqual((x - y).component(1 ) , 1 ) self.assertEqual((x - y).component(2 ) , 2 ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Vector([1, 2, 3] ) _snake_case : List[Any] = Vector([2, -1, 4] ) # for test of dot product _snake_case : Union[str, Any] = Vector([1, -2, -1] ) self.assertEqual(str(x * 3.0 ) , "(3.0,6.0,9.0)" ) self.assertEqual((a * b) , 0 ) def lowerCamelCase__ ( self ): self.assertEqual(str(zero_vector(10 ) ).count("0" ) , 10 ) def lowerCamelCase__ ( self ): self.assertEqual(str(unit_basis_vector(3 , 1 ) ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Vector([1, 2, 3] ) _snake_case : Optional[Any] = Vector([1, 0, 1] ) self.assertEqual(str(axpy(2 , snake_case_ , snake_case_ ) ) , "(3,4,7)" ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Vector([1, 0, 0, 0, 0, 0] ) _snake_case : Optional[int] = x.copy() self.assertEqual(str(snake_case_ ) , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Vector([1, 0, 0] ) x.change_component(0 , 0 ) x.change_component(1 , 1 ) self.assertEqual(str(snake_case_ ) , "(0,1,0)" ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual("|1,2,3|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Dict = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : str = [[-3, -14, -10], [-5, -10, -5], [-2, -1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(minors[x][y] , a.minor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : Union[str, Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = [[-3, 14, -10], [5, -10, 5], [-2, 1, 0]] for x in range(a.height() ): for y in range(a.width() ): self.assertEqual(cofactors[x][y] , a.cofactor(snake_case_ , snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(-5 , a.determinant() ) def lowerCamelCase__ ( self ): _snake_case : str = Matrix([[1, 2, 3], [4, 5, 6], [7, 8, 9]] , 3 , 3 ) _snake_case : List[str] = Vector([1, 2, 3] ) self.assertEqual("(14,32,50)" , str(a * x ) ) self.assertEqual("|2,4,6|\n|8,10,12|\n|14,16,18|\n" , str(a * 2 ) ) def lowerCamelCase__ ( self ): _snake_case : Optional[int] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) a.change_component(0 , 2 , 5 ) self.assertEqual("|1,2,5|\n|2,4,5|\n|6,7,8|\n" , str(snake_case_ ) ) def lowerCamelCase__ ( self ): _snake_case : List[Any] = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) self.assertEqual(7 , a.component(2 , 1 ) , 0.01 ) def lowerCamelCase__ ( self ): _snake_case : Tuple = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : int = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|2,4,10|\n|4,8,10|\n|12,14,18|\n" , str(a + b ) ) def lowerCamelCase__ ( self ): _snake_case : int = Matrix([[1, 2, 3], [2, 4, 5], [6, 7, 8]] , 3 , 3 ) _snake_case : Optional[Any] = Matrix([[1, 2, 7], [2, 4, 5], [6, 7, 10]] , 3 , 3 ) self.assertEqual("|0,0,-4|\n|0,0,0|\n|0,0,-2|\n" , str(a - b ) ) def lowerCamelCase__ ( self ): self.assertEqual( "|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n|0,0,0,0,0|\n" , str(square_zero_matrix(5 ) ) , ) if __name__ == "__main__": unittest.main()
87
0
from typing import Dict, List from nltk.translate import gleu_score import datasets from datasets import MetricInfo _a : Tuple = """\ @misc{wu2016googles, title={Google's Neural Machine Translation System: Bridging the Gap between Human and Machine Translation}, author={Yonghui Wu and Mike Schuster and Zhifeng Chen and Quoc V. Le and Mohammad Norouzi and Wolfgang Macherey and Maxim Krikun and Yuan Cao and Qin Gao and Klaus Macherey and Jeff Klingner and Apurva Shah and Melvin Johnson and Xiaobing Liu and Łukasz Kaiser and Stephan Gouws and Yoshikiyo Kato and Taku Kudo and Hideto Kazawa and Keith Stevens and George Kurian and Nishant Patil and Wei Wang and Cliff Young and Jason Smith and Jason Riesa and Alex Rudnick and Oriol Vinyals and Greg Corrado and Macduff Hughes and Jeffrey Dean}, year={2016}, eprint={1609.08144}, archivePrefix={arXiv}, primaryClass={cs.CL} } """ _a : str = """\ The BLEU score has some undesirable properties when used for single sentences, as it was designed to be a corpus measure. We therefore use a slightly different score for our RL experiments which we call the 'GLEU score'. For the GLEU score, we record all sub-sequences of 1, 2, 3 or 4 tokens in output and target sequence (n-grams). We then compute a recall, which is the ratio of the number of matching n-grams to the number of total n-grams in the target (ground truth) sequence, and a precision, which is the ratio of the number of matching n-grams to the number of total n-grams in the generated output sequence. Then GLEU score is simply the minimum of recall and precision. This GLEU score's range is always between 0 (no matches) and 1 (all match) and it is symmetrical when switching output and target. According to our experiments, GLEU score correlates quite well with the BLEU metric on a corpus level but does not have its drawbacks for our per sentence reward objective. """ _a : Optional[int] = """\ Computes corpus-level Google BLEU (GLEU) score of translated segments against one or more references. Instead of averaging the sentence level GLEU scores (i.e. macro-average precision), Wu et al. (2016) sum up the matching tokens and the max of hypothesis and reference tokens for each sentence, then compute using the aggregate values. Args: predictions (list of str): list of translations to score. Each translation should be tokenized into a list of tokens. references (list of list of str): list of lists of references for each translation. Each reference should be tokenized into a list of tokens. min_len (int): The minimum order of n-gram this function should extract. Defaults to 1. max_len (int): The maximum order of n-gram this function should extract. Defaults to 4. Returns: 'google_bleu': google_bleu score Examples: Example 1: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.44 Example 2: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references) >>> print(round(results[\"google_bleu\"], 2)) 0.61 Example 3: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses, references=list_of_references, min_len=2) >>> print(round(results[\"google_bleu\"], 2)) 0.53 Example 4: >>> hyp1 = ['It', 'is', 'a', 'guide', 'to', 'action', 'which', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'always', ... 'disobeys', 'the', 'commands', 'of', 'the', 'cat'] >>> ref1a = ['It', 'is', 'the', 'guiding', 'principle', 'which', ... 'guarantees', 'the', 'rubber', 'duck', 'forces', 'never', ... 'being', 'under', 'the', 'command', 'of', 'the', 'cat'] >>> ref1b = ['It', 'is', 'a', 'guide', 'to', 'action', 'that', ... 'ensures', 'that', 'the', 'rubber', 'duck', 'will', 'never', ... 'heed', 'the', 'cat', 'commands'] >>> ref1c = ['It', 'is', 'the', 'practical', 'guide', 'for', 'the', ... 'rubber', 'duck', 'army', 'never', 'to', 'heed', 'the', 'directions', ... 'of', 'the', 'cat'] >>> hyp2 = ['he', 'read', 'the', 'book', 'because', 'he', 'was', ... 'interested', 'in', 'world', 'history'] >>> ref2a = ['he', 'was', 'interested', 'in', 'world', 'history', ... 'because', 'he', 'read', 'the', 'book'] >>> list_of_references = [[ref1a, ref1b, ref1c], [ref2a]] >>> hypotheses = [hyp1, hyp2] >>> google_bleu = datasets.load_metric(\"google_bleu\") >>> results = google_bleu.compute(predictions=hypotheses,references=list_of_references, min_len=2, max_len=6) >>> print(round(results[\"google_bleu\"], 2)) 0.4 """ @datasets.utils.file_utils.add_start_docstrings(_DESCRIPTION , _KWARGS_DESCRIPTION) class _UpperCAmelCase ( datasets.Metric): def lowerCamelCase__ ( self ): return datasets.MetricInfo( description=_DESCRIPTION , citation=_CITATION , inputs_description=_KWARGS_DESCRIPTION , features=datasets.Features( { "predictions": datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ), "references": datasets.Sequence( datasets.Sequence(datasets.Value("string" , id="token" ) , id="sequence" ) , id="references" ), } ) , ) def lowerCamelCase__ ( self , snake_case_ , snake_case_ , snake_case_ = 1 , snake_case_ = 4 , ): return { "google_bleu": gleu_score.corpus_gleu( list_of_references=snake_case_ , hypotheses=snake_case_ , min_len=snake_case_ , max_len=snake_case_ ) }
701
"""simple docstring""" from __future__ import annotations from collections import namedtuple def a__ ( a : float , a : float , a : float ): """simple docstring""" _snake_case : Optional[Any] = namedtuple("result" , "name value" ) if (voltage, current, power).count(0 ) != 1: raise ValueError("Only one argument must be 0" ) elif power < 0: raise ValueError( "Power cannot be negative in any electrical/electronics system" ) elif voltage == 0: return result("voltage" , power / current ) elif current == 0: return result("current" , power / voltage ) elif power == 0: return result("power" , float(round(abs(voltage * current ) , 2 ) ) ) else: raise ValueError("Exactly one argument must be 0" ) if __name__ == "__main__": import doctest doctest.testmod()
87
0
"""simple docstring""" def a__ ( a : str ): """simple docstring""" return "".join(chr(ord(a ) - 32 ) if "a" <= char <= "z" else char for char in word ) if __name__ == "__main__": from doctest import testmod testmod()
702
"""simple docstring""" import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import ( AutoencoderKL, DDIMScheduler, DPMSolverMultistepScheduler, TextToVideoSDPipeline, UNetaDConditionModel, ) from diffusers.utils import is_xformers_available, load_numpy, skip_mps, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism from ..pipeline_params import TEXT_TO_IMAGE_BATCH_PARAMS, TEXT_TO_IMAGE_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() @skip_mps class _UpperCAmelCase ( _snake_case , unittest.TestCase): __lowercase : Any = TextToVideoSDPipeline __lowercase : str = TEXT_TO_IMAGE_PARAMS __lowercase : int = TEXT_TO_IMAGE_BATCH_PARAMS # No `output_type`. __lowercase : Optional[int] = frozenset( [ """num_inference_steps""", """generator""", """latents""", """return_dict""", """callback""", """callback_steps""", ]) def lowerCamelCase__ ( self ): torch.manual_seed(0 ) _snake_case : str = UNetaDConditionModel( block_out_channels=(32, 64, 64, 64) , layers_per_block=2 , sample_size=32 , in_channels=4 , out_channels=4 , down_block_types=("CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "CrossAttnDownBlock3D", "DownBlock3D") , up_block_types=("UpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D", "CrossAttnUpBlock3D") , cross_attention_dim=32 , attention_head_dim=4 , ) _snake_case : List[Any] = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , clip_sample=snake_case_ , set_alpha_to_one=snake_case_ , ) torch.manual_seed(0 ) _snake_case : Union[str, Any] = AutoencoderKL( block_out_channels=[32, 64] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , sample_size=1_28 , ) torch.manual_seed(0 ) _snake_case : Optional[Any] = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=32 , intermediate_size=37 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=10_00 , hidden_act="gelu" , projection_dim=5_12 , ) _snake_case : Tuple = CLIPTextModel(snake_case_ ) _snake_case : Optional[int] = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) _snake_case : Any = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, } return components def lowerCamelCase__ ( self , snake_case_ , snake_case_=0 ): if str(snake_case_ ).startswith("mps" ): _snake_case : str = torch.manual_seed(snake_case_ ) else: _snake_case : Union[str, Any] = torch.Generator(device=snake_case_ ).manual_seed(snake_case_ ) _snake_case : str = { "prompt": "A painting of a squirrel eating a burger", "generator": generator, "num_inference_steps": 2, "guidance_scale": 6.0, "output_type": "pt", } return inputs def lowerCamelCase__ ( self ): _snake_case : int = "cpu" # ensure determinism for the device-dependent torch.Generator _snake_case : Optional[Any] = self.get_dummy_components() _snake_case : Tuple = TextToVideoSDPipeline(**snake_case_ ) _snake_case : List[str] = sd_pipe.to(snake_case_ ) sd_pipe.set_progress_bar_config(disable=snake_case_ ) _snake_case : int = self.get_dummy_inputs(snake_case_ ) _snake_case : Union[str, Any] = "np" _snake_case : Dict = sd_pipe(**snake_case_ ).frames _snake_case : Any = frames[0][-3:, -3:, -1] assert frames[0].shape == (64, 64, 3) _snake_case : Dict = np.array([158.0, 160.0, 153.0, 125.0, 100.0, 121.0, 111.0, 93.0, 113.0] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 def lowerCamelCase__ ( self ): self._test_attention_slicing_forward_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=3E-3 ) @unittest.skipIf( torch_device != "cuda" or not is_xformers_available() , reason="XFormers attention is only available with CUDA and `xformers` installed" , ) def lowerCamelCase__ ( self ): self._test_xformers_attention_forwardGenerator_pass(test_mean_pixel_difference=snake_case_ , expected_max_diff=1E-2 ) @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="Batching needs to be properly figured out first for this pipeline." ) def lowerCamelCase__ ( self ): pass @unittest.skip(reason="`num_images_per_prompt` argument is not supported for this pipeline." ) def lowerCamelCase__ ( self ): pass def lowerCamelCase__ ( self ): return super().test_progress_bar() @slow @skip_mps class _UpperCAmelCase ( unittest.TestCase): def lowerCamelCase__ ( self ): _snake_case : List[Any] = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video.npy" ) _snake_case : int = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : str = DPMSolverMultistepScheduler.from_config(pipe.scheduler.config ) _snake_case : Tuple = pipe.to("cuda" ) _snake_case : List[Any] = "Spiderman is surfing" _snake_case : Optional[int] = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : int = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=25 , output_type="pt" ).frames _snake_case : int = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2 def lowerCamelCase__ ( self ): _snake_case : Any = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/text_to_video/video_2step.npy" ) _snake_case : str = TextToVideoSDPipeline.from_pretrained("damo-vilab/text-to-video-ms-1.7b" ) _snake_case : int = pipe.to("cuda" ) _snake_case : Any = "Spiderman is surfing" _snake_case : str = torch.Generator(device="cpu" ).manual_seed(0 ) _snake_case : Any = pipe(snake_case_ , generator=snake_case_ , num_inference_steps=2 , output_type="pt" ).frames _snake_case : Optional[int] = video_frames.cpu().numpy() assert np.abs(expected_video - video ).mean() < 5E-2
87
0