code
stringlengths
87
55.2k
code_codestyle
int64
0
349
style_context
stringlengths
135
49.1k
style_context_codestyle
int64
0
349
label
int64
0
1
from collections.abc import Iterator, MutableMapping from dataclasses import dataclass from typing import Generic, TypeVar UpperCamelCase__ = TypeVar('KEY') UpperCamelCase__ = TypeVar('VAL') @dataclass(frozen=UpperCAmelCase_ , slots=UpperCAmelCase_ ) class A ( Generic[KEY, VAL] ): __UpperCAmelCase : KEY __UpperCAmelCase : VAL class A ( _Item ): def __init__(self : str ) -> None: """simple docstring""" super().__init__(__UpperCAmelCase , __UpperCAmelCase ) def __bool__(self : List[Any] ) -> bool: """simple docstring""" return False UpperCamelCase__ = _DeletedItem() class A ( MutableMapping[KEY, VAL] ): def __init__(self : List[str] , __UpperCAmelCase : int = 8 , __UpperCAmelCase : float = 0.75 ) -> None: """simple docstring""" UpperCAmelCase__ = initial_block_size UpperCAmelCase__ = [None] * initial_block_size assert 0.0 < capacity_factor < 1.0 UpperCAmelCase__ = capacity_factor UpperCAmelCase__ = 0 def lowercase_ (self : str , __UpperCAmelCase : KEY ) -> int: """simple docstring""" return hash(__UpperCAmelCase ) % len(self._buckets ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int ) -> int: """simple docstring""" return (ind + 1) % len(self._buckets ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : KEY , __UpperCAmelCase : VAL ) -> bool: """simple docstring""" UpperCAmelCase__ = self._buckets[ind] if not stored: UpperCAmelCase__ = _Item(__UpperCAmelCase , __UpperCAmelCase ) self._len += 1 return True elif stored.key == key: UpperCAmelCase__ = _Item(__UpperCAmelCase , __UpperCAmelCase ) return True else: return False def lowercase_ (self : List[Any] ) -> bool: """simple docstring""" UpperCAmelCase__ = len(self._buckets ) * self._capacity_factor return len(self ) >= int(__UpperCAmelCase ) def lowercase_ (self : Dict ) -> bool: """simple docstring""" if len(self._buckets ) <= self._initial_block_size: return False UpperCAmelCase__ = len(self._buckets ) * self._capacity_factor / 2 return len(self ) < limit def lowercase_ (self : int , __UpperCAmelCase : int ) -> None: """simple docstring""" UpperCAmelCase__ = self._buckets UpperCAmelCase__ = [None] * new_size UpperCAmelCase__ = 0 for item in old_buckets: if item: self._add_item(item.key , item.val ) def lowercase_ (self : str ) -> None: """simple docstring""" self._resize(len(self._buckets ) * 2 ) def lowercase_ (self : List[str] ) -> None: """simple docstring""" self._resize(len(self._buckets ) // 2 ) def lowercase_ (self : Dict , __UpperCAmelCase : KEY ) -> Iterator[int]: """simple docstring""" UpperCAmelCase__ = self._get_bucket_index(__UpperCAmelCase ) for _ in range(len(self._buckets ) ): yield ind UpperCAmelCase__ = self._get_next_ind(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : KEY , __UpperCAmelCase : VAL ) -> None: """simple docstring""" for ind in self._iterate_buckets(__UpperCAmelCase ): if self._try_set(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ): break def __setitem__(self : List[Any] , __UpperCAmelCase : KEY , __UpperCAmelCase : VAL ) -> None: """simple docstring""" if self._is_full(): self._size_up() self._add_item(__UpperCAmelCase , __UpperCAmelCase ) def __delitem__(self : str , __UpperCAmelCase : KEY ) -> None: """simple docstring""" for ind in self._iterate_buckets(__UpperCAmelCase ): UpperCAmelCase__ = self._buckets[ind] if item is None: raise KeyError(__UpperCAmelCase ) if item is _deleted: continue if item.key == key: UpperCAmelCase__ = _deleted self._len -= 1 break if self._is_sparse(): self._size_down() def __getitem__(self : Dict , __UpperCAmelCase : KEY ) -> VAL: """simple docstring""" for ind in self._iterate_buckets(__UpperCAmelCase ): UpperCAmelCase__ = self._buckets[ind] if item is None: break if item is _deleted: continue if item.key == key: return item.val raise KeyError(__UpperCAmelCase ) def __len__(self : int ) -> int: """simple docstring""" return self._len def __iter__(self : Optional[Any] ) -> Iterator[KEY]: """simple docstring""" yield from (item.key for item in self._buckets if item) def __repr__(self : Optional[Any] ) -> str: """simple docstring""" UpperCAmelCase__ = " ,".join( f"""{item.key}: {item.val}""" for item in self._buckets if item ) return f"""HashMap({val_string})"""
65
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
1
from ...utils import is_torch_available, is_transformers_available if is_transformers_available() and is_torch_available(): from .pipeline_vq_diffusion import LearnedClassifierFreeSamplingEmbeddings, VQDiffusionPipeline
65
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
1
def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' while b: UpperCAmelCase__ , UpperCAmelCase__ = b, a % b return a def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' return a if b == 0 else euclidean_gcd_recursive(__A, a % b ) def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' print(f"""euclidean_gcd(3, 5) = {euclidean_gcd(3, 5 )}""" ) print(f"""euclidean_gcd(5, 3) = {euclidean_gcd(5, 3 )}""" ) print(f"""euclidean_gcd(1, 3) = {euclidean_gcd(1, 3 )}""" ) print(f"""euclidean_gcd(3, 6) = {euclidean_gcd(3, 6 )}""" ) print(f"""euclidean_gcd(6, 3) = {euclidean_gcd(6, 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 5) = {euclidean_gcd_recursive(3, 5 )}""" ) print(f"""euclidean_gcd_recursive(5, 3) = {euclidean_gcd_recursive(5, 3 )}""" ) print(f"""euclidean_gcd_recursive(1, 3) = {euclidean_gcd_recursive(1, 3 )}""" ) print(f"""euclidean_gcd_recursive(3, 6) = {euclidean_gcd_recursive(3, 6 )}""" ) print(f"""euclidean_gcd_recursive(6, 3) = {euclidean_gcd_recursive(6, 3 )}""" ) if __name__ == "__main__": main()
65
from __future__ import annotations from collections import deque class A : def __init__(self : Dict , __UpperCAmelCase : list[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__UpperCAmelCase ) self.set_fail_transitions() def lowercase_ (self : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : str ) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = 0 for character in keyword: UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , __UpperCAmelCase ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase__ = len(self.adlist ) - 1 else: UpperCAmelCase__ = next_state self.adlist[current_state]["output"].append(__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = deque() for node in self.adlist[0]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = 0 while q: UpperCAmelCase__ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = self.adlist[r]["fail_state"] while ( self.find_next_state(__UpperCAmelCase , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase__ = self.adlist[state]["fail_state"] UpperCAmelCase__ = self.find_next_state( __UpperCAmelCase , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase__ = 0 UpperCAmelCase__ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> dict[str, list[int]]: """simple docstring""" UpperCAmelCase__ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase__ = 0 for i in range(len(__UpperCAmelCase ) ): while ( self.find_next_state(__UpperCAmelCase , string[i] ) is None and current_state != 0 ): UpperCAmelCase__ = self.adlist[current_state]["fail_state"] UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , string[i] ) if next_state is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase__ = [] result[key].append(i - len(__UpperCAmelCase ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
65
1
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ['input_values', 'attention_mask'] def __init__(self : Any , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 1_6_0_0_0 , __UpperCAmelCase : float = 0.0 , __UpperCAmelCase : bool = False , __UpperCAmelCase : int = 8_0 , __UpperCAmelCase : int = 1_6 , __UpperCAmelCase : int = 6_4 , __UpperCAmelCase : str = "hann_window" , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : float = 8_0 , __UpperCAmelCase : float = 7_6_0_0 , __UpperCAmelCase : float = 1E-10 , __UpperCAmelCase : int = 2 , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Any , ) -> str: """simple docstring""" super().__init__(feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = do_normalize UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = frame_signal_scale UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = reduction_factor UpperCAmelCase__ = win_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = hop_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = optimal_fft_length(self.sample_size ) UpperCAmelCase__ = (self.n_fft // 2) + 1 UpperCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCAmelCase ) UpperCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowercase_ (__UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase , np.intaa ) UpperCAmelCase__ = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): UpperCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: UpperCAmelCase__ = padding_value normed_input_values.append(__UpperCAmelCase ) else: UpperCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowercase_ (self : Optional[int] , __UpperCAmelCase : np.ndarray , ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ = spectrogram( __UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__(self : Any , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[int] = None , **__UpperCAmelCase : str , ) -> BatchFeature: """simple docstring""" if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) else: UpperCAmelCase__ = None if audio_target is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) if inputs is None: return inputs_target else: UpperCAmelCase__ = inputs_target["input_values"] UpperCAmelCase__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: UpperCAmelCase__ = decoder_attention_mask return inputs def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCAmelCase : bool = False , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : Any , ) -> BatchFeature: """simple docstring""" UpperCAmelCase__ = isinstance(__UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): UpperCAmelCase__ = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs UpperCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: UpperCAmelCase__ = [self._extract_mel_features(__UpperCAmelCase ) for waveform in speech] UpperCAmelCase__ = BatchFeature({"input_values": features} ) UpperCAmelCase__ = self.num_mel_bins else: UpperCAmelCase__ = BatchFeature({"input_values": speech} ) UpperCAmelCase__ = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = feature_size_hack # convert input values to correct format UpperCAmelCase__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(__UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): UpperCAmelCase__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(__UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = input_values.astype(np.floataa ) # convert attention_mask to correct format UpperCAmelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: UpperCAmelCase__ = ( attention_mask if self._get_padding_strategies(__UpperCAmelCase , max_length=__UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=__UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def lowercase_ (self : Tuple ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. UpperCAmelCase__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
65
1
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import ScoreSdeVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A ( UpperCAmelCase_ ): __UpperCAmelCase : UNetaDModel __UpperCAmelCase : ScoreSdeVeScheduler def __init__(self : Tuple , __UpperCAmelCase : UNetaDModel , __UpperCAmelCase : ScoreSdeVeScheduler ) -> List[Any]: """simple docstring""" super().__init__() self.register_modules(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) @torch.no_grad() def __call__(self : List[Any] , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 2_0_0_0 , __UpperCAmelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __UpperCAmelCase : Optional[str] = "pil" , __UpperCAmelCase : bool = True , **__UpperCAmelCase : str , ) -> Union[ImagePipelineOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = self.unet.config.sample_size UpperCAmelCase__ = (batch_size, 3, img_size, img_size) UpperCAmelCase__ = self.unet UpperCAmelCase__ = randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase ) * self.scheduler.init_noise_sigma UpperCAmelCase__ = sample.to(self.device ) self.scheduler.set_timesteps(__UpperCAmelCase ) self.scheduler.set_sigmas(__UpperCAmelCase ) for i, t in enumerate(self.progress_bar(self.scheduler.timesteps ) ): UpperCAmelCase__ = self.scheduler.sigmas[i] * torch.ones(shape[0] , device=self.device ) # correction step for _ in range(self.scheduler.config.correct_steps ): UpperCAmelCase__ = self.unet(__UpperCAmelCase , __UpperCAmelCase ).sample UpperCAmelCase__ = self.scheduler.step_correct(__UpperCAmelCase , __UpperCAmelCase , generator=__UpperCAmelCase ).prev_sample # prediction step UpperCAmelCase__ = model(__UpperCAmelCase , __UpperCAmelCase ).sample UpperCAmelCase__ = self.scheduler.step_pred(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , generator=__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = output.prev_sample, output.prev_sample_mean UpperCAmelCase__ = sample_mean.clamp(0 , 1 ) UpperCAmelCase__ = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase__ = self.numpy_to_pil(__UpperCAmelCase ) if not return_dict: return (sample,) return ImagePipelineOutput(images=__UpperCAmelCase )
65
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[torch.FloatTensor] = None __UpperCAmelCase : torch.FloatTensor = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Union[str, Any]=5_1_2 , __UpperCAmelCase : List[str]="cls" , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=True , **__UpperCAmelCase : str , ) -> int: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = project_dim UpperCAmelCase__ = pooler_fn UpperCAmelCase__ = learn_encoder UpperCAmelCase__ = use_attention_mask class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = [r'pooler', r'logit_scale'] __UpperCAmelCase : int = [r'position_ids', r'predictions.decoder.bias'] __UpperCAmelCase : Any = 'roberta' __UpperCAmelCase : List[str] = RobertaSeriesConfig def __init__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = XLMRobertaModel(__UpperCAmelCase ) UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = getattr(__UpperCAmelCase , "has_pre_transformation" , __UpperCAmelCase ) if self.has_pre_transformation: UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.base_model( input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , position_ids=__UpperCAmelCase , head_mask=__UpperCAmelCase , inputs_embeds=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__UpperCAmelCase , ) if self.has_pre_transformation: UpperCAmelCase__ = outputs["hidden_states"][-2] UpperCAmelCase__ = self.pre_LN(__UpperCAmelCase ) UpperCAmelCase__ = self.transformation_pre(__UpperCAmelCase ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: UpperCAmelCase__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
65
1
from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'openai-gpt': 'https://huggingface.co/openai-gpt/resolve/main/config.json'} class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[str] = 'openai-gpt' __UpperCAmelCase : Optional[Any] = { 'max_position_embeddings': 'n_positions', 'hidden_size': 'n_embd', 'num_attention_heads': 'n_head', 'num_hidden_layers': 'n_layer', } def __init__(self : Optional[Any] , __UpperCAmelCase : str=4_0_4_7_8 , __UpperCAmelCase : int=5_1_2 , __UpperCAmelCase : int=7_6_8 , __UpperCAmelCase : Optional[int]=1_2 , __UpperCAmelCase : List[Any]=1_2 , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : Optional[int]=0.1 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : List[str]=1E-5 , __UpperCAmelCase : Dict=0.02 , __UpperCAmelCase : Union[str, Any]="cls_index" , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Any=True , __UpperCAmelCase : List[Any]=0.1 , **__UpperCAmelCase : int , ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = vocab_size UpperCAmelCase__ = n_positions UpperCAmelCase__ = n_embd UpperCAmelCase__ = n_layer UpperCAmelCase__ = n_head UpperCAmelCase__ = afn UpperCAmelCase__ = resid_pdrop UpperCAmelCase__ = embd_pdrop UpperCAmelCase__ = attn_pdrop UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = summary_type UpperCAmelCase__ = summary_use_proj UpperCAmelCase__ = summary_activation UpperCAmelCase__ = summary_first_dropout UpperCAmelCase__ = summary_proj_to_labels super().__init__(**__UpperCAmelCase )
65
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
1
import inspect import unittest from huggingface_hub import hf_hub_download from transformers import ASTConfig from transformers.testing_utils import require_torch, require_torchaudio, slow, torch_device from transformers.utils import cached_property, is_torch_available, is_torchaudio_available from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import ASTForAudioClassification, ASTModel from transformers.models.audio_spectrogram_transformer.modeling_audio_spectrogram_transformer import ( AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST, ) if is_torchaudio_available(): import torchaudio from transformers import ASTFeatureExtractor class A : def __init__(self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : Any=1_3 , __UpperCAmelCase : Dict=2 , __UpperCAmelCase : Tuple=2_4 , __UpperCAmelCase : List[str]=1_6 , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : Dict=3_2 , __UpperCAmelCase : str=5 , __UpperCAmelCase : Dict=4 , __UpperCAmelCase : Tuple=3_7 , __UpperCAmelCase : int="gelu" , __UpperCAmelCase : Dict=0.1 , __UpperCAmelCase : Dict=0.1 , __UpperCAmelCase : Optional[int]=1_0 , __UpperCAmelCase : Optional[Any]=0.02 , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Dict=2 , __UpperCAmelCase : Tuple=2 , ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = patch_size UpperCAmelCase__ = max_length UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = scope UpperCAmelCase__ = frequency_stride UpperCAmelCase__ = time_stride # in AST, the seq length equals the number of patches + 2 (we add 2 for the [CLS] and distillation tokens) UpperCAmelCase__ = (self.num_mel_bins - self.patch_size) // self.frequency_stride + 1 UpperCAmelCase__ = (self.max_length - self.patch_size) // self.time_stride + 1 UpperCAmelCase__ = frequency_out_dimension * time_out_dimension UpperCAmelCase__ = num_patches + 2 def lowercase_ (self : str ) -> str: """simple docstring""" UpperCAmelCase__ = floats_tensor([self.batch_size, self.max_length, self.num_mel_bins] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = self.get_config() return config, input_values, labels def lowercase_ (self : Optional[Any] ) -> Optional[Any]: """simple docstring""" return ASTConfig( patch_size=self.patch_size , max_length=self.max_length , num_mel_bins=self.num_mel_bins , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , frequency_stride=self.frequency_stride , time_stride=self.time_stride , ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = ASTModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ (self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = config_and_inputs UpperCAmelCase__ = {"input_values": input_values} return config, inputs_dict @require_torch class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : List[Any] = ( ( ASTModel, ASTForAudioClassification, ) if is_torch_available() else () ) __UpperCAmelCase : List[str] = ( {'audio-classification': ASTForAudioClassification, 'feature-extraction': ASTModel} if is_torch_available() else {} ) __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : Tuple = False __UpperCAmelCase : Any = False __UpperCAmelCase : Any = False def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Any , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple ) -> Any: """simple docstring""" if pipeline_test_casse_name == "AudioClassificationPipelineTests": return True return False def lowercase_ (self : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = ASTModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__UpperCAmelCase , has_text_modality=__UpperCAmelCase , hidden_size=3_7 ) def lowercase_ (self : Union[str, Any] ) -> Tuple: """simple docstring""" self.config_tester.run_common_tests() @unittest.skip(reason="AST does not use inputs_embeds" ) def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" pass def lowercase_ (self : int ) -> Any: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(__UpperCAmelCase ) self.assertIsInstance(model.get_input_embeddings() , (nn.Module) ) UpperCAmelCase__ = model.get_output_embeddings() self.assertTrue(x is None or isinstance(__UpperCAmelCase , nn.Linear ) ) def lowercase_ (self : str ) -> str: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ["input_values"] self.assertListEqual(arg_names[:1] , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) @slow def lowercase_ (self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" for model_name in AUDIO_SPECTROGRAM_TRANSFORMER_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = ASTModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' UpperCAmelCase__ = hf_hub_download( repo_id="nielsr/audio-spectogram-transformer-checkpoint", filename="sample_audio.flac", repo_type="dataset" ) UpperCAmelCase__ , UpperCAmelCase__ = torchaudio.load(__A ) return audio, sampling_rate @require_torch @require_torchaudio class A ( unittest.TestCase ): @cached_property def lowercase_ (self : List[Any] ) -> List[str]: """simple docstring""" return ( ASTFeatureExtractor.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ) if is_torchaudio_available() else None ) @slow def lowercase_ (self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase__ = self.default_feature_extractor UpperCAmelCase__ = ASTForAudioClassification.from_pretrained("MIT/ast-finetuned-audioset-10-10-0.4593" ).to(__UpperCAmelCase ) UpperCAmelCase__ = self.default_feature_extractor UpperCAmelCase__ , UpperCAmelCase__ = prepare_audio() UpperCAmelCase__ = audio.squeeze().numpy() UpperCAmelCase__ = feature_extractor(__UpperCAmelCase , sampling_rate=__UpperCAmelCase , return_tensors="pt" ).to(__UpperCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase__ = model(**__UpperCAmelCase ) # verify the logits UpperCAmelCase__ = torch.Size((1, 5_2_7) ) self.assertEqual(outputs.logits.shape , __UpperCAmelCase ) UpperCAmelCase__ = torch.tensor([-0.8760, -7.0042, -8.6602] ).to(__UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCAmelCase , atol=1E-4 ) )
65
import math import random def lowerCAmelCase_ ( __A, __A = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase__ = 0.0_2 def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' UpperCAmelCase__ = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__A ): # Forward propagation UpperCAmelCase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase__ = (expected / 100) - layer_a # Error delta UpperCAmelCase__ = layer_1_error * sigmoid_function(__A, __A ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input('Expected value: ')) UpperCamelCase__ = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
65
1
from collections.abc import Callable import numpy as np def lowerCAmelCase_ ( __A, __A, __A, __A, __A ) -> np.array: '''simple docstring''' UpperCAmelCase__ = int(np.ceil((x_end - xa) / step_size ) ) UpperCAmelCase__ = np.zeros((n + 1,) ) UpperCAmelCase__ = ya UpperCAmelCase__ = xa for k in range(__A ): UpperCAmelCase__ = y[k] + step_size * ode_func(__A, y[k] ) UpperCAmelCase__ = y[k] + ( (step_size / 2) * (ode_func(__A, y[k] ) + ode_func(x + step_size, __A )) ) x += step_size return y if __name__ == "__main__": import doctest doctest.testmod()
65
from __future__ import annotations class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : list[list[int]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed from a list of zero or more lists containing at " "least one and the same number of values, each of which must be of type " "int or float." ) if len(__UpperCAmelCase ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__UpperCAmelCase ) != cols: raise error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def lowercase_ (self : Any ) -> list[list[int]]: """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def lowercase_ (self : Any ) -> int: """simple docstring""" return len(self.rows ) @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" return len(self.rows[0] ) @property def lowercase_ (self : List[Any] ) -> tuple[int, int]: """simple docstring""" return (self.num_rows, self.num_columns) @property def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return self.order[0] == self.order[1] def lowercase_ (self : Any ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : int ) -> int: """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return bool(self.determinant() ) def lowercase_ (self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__UpperCAmelCase ).determinant() def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) return -1 * self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Matrix: """simple docstring""" return Matrix( [ [self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def lowercase_ (self : List[str] ) -> Matrix: """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def lowercase_ (self : Optional[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError("Only matrices with a non-zero determinant have an inverse" ) return self.adjugate() * (1 / determinant) def __repr__(self : Dict ) -> str: """simple docstring""" return str(self.rows ) def __str__(self : Optional[Any] ) -> str: """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ "[" + ". ".join([str(__UpperCAmelCase ) for value in row] ) + ".]" for row in self.rows ] ) + "]" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError("Row must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_columns: raise ValueError( "Row must be equal in length to the other rows in the matrix" ) if position is None: self.rows.append(__UpperCAmelCase ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError( "Column must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in column: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_rows: raise ValueError( "Column must be equal in length to the other columns in the matrix" ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__(self : Any , __UpperCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return NotImplemented return self.rows == other.rows def __ne__(self : int , __UpperCAmelCase : object ) -> bool: """simple docstring""" return not self == other def __neg__(self : Dict ) -> Matrix: """simple docstring""" return self * -1 def __add__(self : Dict , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Addition requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__(self : Optional[Any] , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Subtraction requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__(self : Tuple , __UpperCAmelCase : Matrix | int | float ) -> Matrix: """simple docstring""" if isinstance(__UpperCAmelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): if self.num_columns != other.num_rows: raise ValueError( "The number of columns in the first matrix must " "be equal to the number of rows in the second" ) return Matrix( [ [Matrix.dot_product(__UpperCAmelCase , __UpperCAmelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( "A Matrix can only be multiplied by an int, float, or another matrix" ) def __pow__(self : List[Any] , __UpperCAmelCase : int ) -> Matrix: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("A Matrix can only be raised to the power of an int" ) if not self.is_square: raise ValueError("Only square matrices can be raised to a power" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( "Only invertable matrices can be raised to a negative power" ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : list[int] , __UpperCAmelCase : list[int] ) -> int: """simple docstring""" return sum(row[i] * column[i] for i in range(len(__UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
1
def lowerCAmelCase_ ( __A, __A = False ) -> bool: '''simple docstring''' if n == 2: return True if not n % 2 or n < 2: return False if n > 5 and n % 10 not in (1, 3, 7, 9): # can quickly check last digit return False if n > 3_317_044_064_679_887_385_961_981 and not allow_probable: raise ValueError( "Warning: upper bound of deterministic test is exceeded. " "Pass allow_probable=True to allow probabilistic test. " "A return value of True indicates a probable prime." ) # array bounds provided by analysis UpperCAmelCase__ = [ 2_047, 1_373_653, 25_326_001, 3_215_031_751, 2_152_302_898_747, 3_474_749_660_383, 341_550_071_728_321, 1, 3_825_123_056_546_413_051, 1, 1, 318_665_857_834_031_151_167_461, 3_317_044_064_679_887_385_961_981, ] UpperCAmelCase__ = [2, 3, 5, 7, 11, 13, 17, 19, 23, 29, 31, 37, 41] for idx, _p in enumerate(__A, 1 ): if n < _p: # then we have our last prime to check UpperCAmelCase__ = primes[:idx] break UpperCAmelCase__ , UpperCAmelCase__ = n - 1, 0 # break up n -1 into a power of 2 (s) and # remaining odd component # essentially, solve for d * 2 ** s == n - 1 while d % 2 == 0: d //= 2 s += 1 for prime in plist: UpperCAmelCase__ = False for r in range(__A ): UpperCAmelCase__ = pow(__A, d * 2**r, __A ) # see article for analysis explanation for m if (r == 0 and m == 1) or ((m + 1) % n == 0): UpperCAmelCase__ = True # this loop will not determine compositeness break if pr: continue # if pr is False, then the above loop never evaluated to true, # and the n MUST be composite return False return True def lowerCAmelCase_ ( ) -> None: '''simple docstring''' assert not miller_rabin(561 ) assert miller_rabin(563 ) # 2047 assert not miller_rabin(838_201 ) assert miller_rabin(838_207 ) # 1_373_653 assert not miller_rabin(17_316_001 ) assert miller_rabin(17_316_017 ) # 25_326_001 assert not miller_rabin(3_078_386_641 ) assert miller_rabin(3_078_386_653 ) # 3_215_031_751 assert not miller_rabin(1_713_045_574_801 ) assert miller_rabin(1_713_045_574_819 ) # 2_152_302_898_747 assert not miller_rabin(2_779_799_728_307 ) assert miller_rabin(2_779_799_728_327 ) # 3_474_749_660_383 assert not miller_rabin(113_850_023_909_441 ) assert miller_rabin(113_850_023_909_527 ) # 341_550_071_728_321 assert not miller_rabin(1_275_041_018_848_804_351 ) assert miller_rabin(1_275_041_018_848_804_391 ) # 3_825_123_056_546_413_051 assert not miller_rabin(79_666_464_458_507_787_791_867 ) assert miller_rabin(79_666_464_458_507_787_791_951 ) # 318_665_857_834_031_151_167_461 assert not miller_rabin(552_840_677_446_647_897_660_333 ) assert miller_rabin(552_840_677_446_647_897_660_359 ) # 3_317_044_064_679_887_385_961_981 # upper limit for probabilistic test if __name__ == "__main__": test_miller_rabin()
65
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
1
import warnings from collections import OrderedDict from typing import Mapping from packaging import version from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'nvidia/segformer-b0-finetuned-ade-512-512': ( 'https://huggingface.co/nvidia/segformer-b0-finetuned-ade-512-512/resolve/main/config.json' ), # See all SegFormer models at https://huggingface.co/models?filter=segformer } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Dict = 'segformer' def __init__(self : Optional[int] , __UpperCAmelCase : List[Any]=3 , __UpperCAmelCase : List[Any]=4 , __UpperCAmelCase : Optional[int]=[2, 2, 2, 2] , __UpperCAmelCase : List[str]=[8, 4, 2, 1] , __UpperCAmelCase : Any=[3_2, 6_4, 1_6_0, 2_5_6] , __UpperCAmelCase : int=[7, 3, 3, 3] , __UpperCAmelCase : Optional[Any]=[4, 2, 2, 2] , __UpperCAmelCase : List[str]=[1, 2, 5, 8] , __UpperCAmelCase : Dict=[4, 4, 4, 4] , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : Union[str, Any]=0.0 , __UpperCAmelCase : Tuple=0.0 , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : Tuple=0.02 , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : List[str]=1E-6 , __UpperCAmelCase : Union[str, Any]=2_5_6 , __UpperCAmelCase : List[Any]=2_5_5 , **__UpperCAmelCase : Dict , ) -> Union[str, Any]: """simple docstring""" super().__init__(**__UpperCAmelCase ) if "reshape_last_stage" in kwargs and kwargs["reshape_last_stage"] is False: warnings.warn( "Reshape_last_stage is set to False in this config. This argument is deprecated and will soon be" " removed, as the behaviour will default to that of reshape_last_stage = True." , __UpperCAmelCase , ) UpperCAmelCase__ = num_channels UpperCAmelCase__ = num_encoder_blocks UpperCAmelCase__ = depths UpperCAmelCase__ = sr_ratios UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = patch_sizes UpperCAmelCase__ = strides UpperCAmelCase__ = mlp_ratios UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = classifier_dropout_prob UpperCAmelCase__ = initializer_range UpperCAmelCase__ = drop_path_rate UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = decoder_hidden_size UpperCAmelCase__ = kwargs.get("reshape_last_stage" , __UpperCAmelCase ) UpperCAmelCase__ = semantic_loss_ignore_index class A ( UpperCAmelCase_ ): __UpperCAmelCase : Any = version.parse('1.11' ) @property def lowercase_ (self : int ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" return OrderedDict( [ ("pixel_values", {0: "batch", 1: "num_channels", 2: "height", 3: "width"}), ] ) @property def lowercase_ (self : Tuple ) -> float: """simple docstring""" return 1E-4 @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return 1_2
65
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor class A ( nn.Module ): def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , __UpperCAmelCase : int=(6_4,) , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Any=3_2 , __UpperCAmelCase : str="silu" , __UpperCAmelCase : Any=True , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = torch.nn.Convad( __UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) # down UpperCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_down_block( __UpperCAmelCase , num_layers=self.layers_per_block , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) self.down_blocks.append(__UpperCAmelCase ) # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # out UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = 2 * out_channels if double_z else out_channels UpperCAmelCase__ = nn.Convad(block_out_channels[-1] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : List[Any] , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = x UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : int ): def custom_forward(*__UpperCAmelCase : Optional[Any] ): return module(*__UpperCAmelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __UpperCAmelCase ) else: # down for down_block in self.down_blocks: UpperCAmelCase__ = down_block(__UpperCAmelCase ) # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase ) # post-process UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : List[Any] , __UpperCAmelCase : str=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=("UpDecoderBlock2D",) , __UpperCAmelCase : str=(6_4,) , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : Any="silu" , __UpperCAmelCase : Any="group" , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) UpperCAmelCase__ = in_channels if norm_type == "spatial" else None # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # up UpperCAmelCase__ = list(reversed(__UpperCAmelCase ) ) UpperCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = reversed_block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_up_block( __UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , prev_output_channel=__UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , resnet_time_scale_shift=__UpperCAmelCase , ) self.up_blocks.append(__UpperCAmelCase ) UpperCAmelCase__ = output_channel # out if norm_type == "spatial": UpperCAmelCase__ = SpatialNorm(block_out_channels[0] , __UpperCAmelCase ) else: UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = nn.Convad(block_out_channels[0] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = z UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) UpperCAmelCase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : str ): def custom_forward(*__UpperCAmelCase : List[str] ): return module(*__UpperCAmelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase ) else: # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = up_block(__UpperCAmelCase , __UpperCAmelCase ) # post-process if latent_embeds is None: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) else: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Union[str, Any]="random" , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=True ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = n_e UpperCAmelCase__ = vq_embed_dim UpperCAmelCase__ = beta UpperCAmelCase__ = legacy UpperCAmelCase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) UpperCAmelCase__ = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) UpperCAmelCase__ = self.used.shape[0] UpperCAmelCase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": UpperCAmelCase__ = self.re_embed UpperCAmelCase__ = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: UpperCAmelCase__ = n_e UpperCAmelCase__ = sane_index_shape def lowercase_ (self : str , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) UpperCAmelCase__ = (inds[:, :, None] == used[None, None, ...]).long() UpperCAmelCase__ = match.argmax(-1 ) UpperCAmelCase__ = match.sum(2 ) < 1 if self.unknown_index == "random": UpperCAmelCase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: UpperCAmelCase__ = self.unknown_index return new.reshape(__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token UpperCAmelCase__ = 0 # simply set to zero UpperCAmelCase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __UpperCAmelCase ) return back.reshape(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() UpperCAmelCase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z UpperCAmelCase__ = torch.argmin(torch.cdist(__UpperCAmelCase , self.embedding.weight ) , dim=1 ) UpperCAmelCase__ = self.embedding(__UpperCAmelCase ).view(z.shape ) UpperCAmelCase__ = None UpperCAmelCase__ = None # compute loss for embedding if not self.legacy: UpperCAmelCase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: UpperCAmelCase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients UpperCAmelCase__ = z + (z_q - z).detach() # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: UpperCAmelCase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.remap_to_used(__UpperCAmelCase ) UpperCAmelCase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: UpperCAmelCase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Any: """simple docstring""" if self.remap is not None: UpperCAmelCase__ = indices.reshape(shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.unmap_to_all(__UpperCAmelCase ) UpperCAmelCase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors UpperCAmelCase__ = self.embedding(__UpperCAmelCase ) if shape is not None: UpperCAmelCase__ = z_q.view(__UpperCAmelCase ) # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parameters UpperCAmelCase__ , UpperCAmelCase__ = torch.chunk(__UpperCAmelCase , 2 , dim=1 ) UpperCAmelCase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) UpperCAmelCase__ = deterministic UpperCAmelCase__ = torch.exp(0.5 * self.logvar ) UpperCAmelCase__ = torch.exp(self.logvar ) if self.deterministic: UpperCAmelCase__ = UpperCAmelCase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = randn_tensor( self.mean.shape , generator=__UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) UpperCAmelCase__ = self.mean + self.std * sample return x def lowercase_ (self : str , __UpperCAmelCase : int=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase_ (self : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=[1, 2, 3] ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) UpperCAmelCase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" return self.mean
65
1
import argparse import pickle import numpy as np import torch from torch import nn from transformers import ReformerConfig, ReformerModelWithLMHead from transformers.utils import logging logging.set_verbosity_info() def lowerCAmelCase_ ( __A, __A, __A=None ) -> List[str]: '''simple docstring''' assert torch_layer.weight.shape == weight.shape, f"""{torch_layer} layer.weight does not match""" UpperCAmelCase__ = nn.Parameter(__A ) if bias is not None: assert torch_layer.bias.shape == bias.shape, f"""{torch_layer} layer.bias does not match""" UpperCAmelCase__ = nn.Parameter(__A ) def lowerCAmelCase_ ( __A, __A, __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = np.asarray(weights[0] ) UpperCAmelCase__ = np.asarray(weights[1] ) UpperCAmelCase__ = np.asarray(weights[2] ) set_param( torch_layer.self_attention.query_key, torch.tensor(__A ).transpose(1, 2 ).contiguous().view(-1, __A ), ) set_param( torch_layer.self_attention.value, torch.tensor(__A ).transpose(1, 2 ).contiguous().view(-1, __A ), ) set_param( torch_layer.output.dense, torch.tensor(__A ).view(-1, __A ).contiguous().transpose(0, 1 ), ) def lowerCAmelCase_ ( __A, __A, __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = np.asarray(weights[0] ) UpperCAmelCase__ = np.asarray(weights[1] ) UpperCAmelCase__ = np.asarray(weights[2] ) UpperCAmelCase__ = np.asarray(weights[3] ) set_param( torch_layer.self_attention.query, torch.tensor(__A ).transpose(1, 2 ).contiguous().view(-1, __A ), ) set_param( torch_layer.self_attention.key, torch.tensor(__A ).transpose(1, 2 ).contiguous().view(-1, __A ), ) set_param( torch_layer.self_attention.value, torch.tensor(__A ).transpose(1, 2 ).contiguous().view(-1, __A ), ) set_param( torch_layer.output.dense, torch.tensor(__A ).view(-1, __A ).contiguous().transpose(0, 1 ), ) def lowerCAmelCase_ ( __A, __A, __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = weights[0][0][0] UpperCAmelCase__ = np.asarray(layer_norm_a[0] ) UpperCAmelCase__ = np.asarray(layer_norm_a[1] ) set_param( torch_block.attention.layer_norm, torch.tensor(__A ), torch.tensor(__A ), ) # lsh weights + output UpperCAmelCase__ = weights[0][1] if len(__A ) < 4: set_layer_weights_in_torch_lsh(__A, torch_block.attention, __A ) else: set_layer_weights_in_torch_local(__A, torch_block.attention, __A ) # intermediate weighs UpperCAmelCase__ = weights[2][0][1][2] # Chunked Feed Forward if len(__A ) == 4: UpperCAmelCase__ = intermediate_weights[2] # layernorm 2 UpperCAmelCase__ = np.asarray(intermediate_weights[0][0] ) UpperCAmelCase__ = np.asarray(intermediate_weights[0][1] ) set_param( torch_block.feed_forward.layer_norm, torch.tensor(__A ), torch.tensor(__A ), ) # intermediate dense UpperCAmelCase__ = np.asarray(intermediate_weights[1][0] ) UpperCAmelCase__ = np.asarray(intermediate_weights[1][1] ) set_param( torch_block.feed_forward.dense.dense, torch.tensor(__A ).transpose(0, 1 ).contiguous(), torch.tensor(__A ), ) # intermediate out UpperCAmelCase__ = np.asarray(intermediate_weights[4][0] ) UpperCAmelCase__ = np.asarray(intermediate_weights[4][1] ) set_param( torch_block.feed_forward.output.dense, torch.tensor(__A ).transpose(0, 1 ).contiguous(), torch.tensor(__A ), ) def lowerCAmelCase_ ( __A, __A, __A ) -> str: '''simple docstring''' UpperCAmelCase__ = torch_model.reformer # word embeds UpperCAmelCase__ = np.asarray(weights[1] ) set_param( torch_model_reformer.embeddings.word_embeddings, torch.tensor(__A ), ) if isinstance(weights[3], __A ): UpperCAmelCase__ = torch_model_reformer.embeddings.position_embeddings for emb_idx in range(len(position_embeddings.weights ) ): UpperCAmelCase__ = np.asarray(weights[3][emb_idx][0] ) assert ( position_embeddings.weights[emb_idx].shape == emb_weights.shape ), f"""{position_embeddings[emb_idx]} emb does not match""" UpperCAmelCase__ = nn.Parameter(torch.tensor(__A ) ) UpperCAmelCase__ = weights[5] assert len(torch_model_reformer.encoder.layers ) * 4 == len( __A ), "HF and trax model do not have the same number of layers" for layer_idx, layer in enumerate(torch_model_reformer.encoder.layers ): UpperCAmelCase__ = trax_layer_weights[4 * layer_idx : 4 * (layer_idx + 1)] set_block_weights_in_torch(__A, __A, __A ) # output layer norm UpperCAmelCase__ = np.asarray(weights[7][0] ) UpperCAmelCase__ = np.asarray(weights[7][1] ) set_param( torch_model_reformer.encoder.layer_norm, torch.tensor(__A ), torch.tensor(__A ), ) # output embeddings UpperCAmelCase__ = np.asarray(weights[9][0] ) UpperCAmelCase__ = np.asarray(weights[9][1] ) set_param( torch_model.lm_head.decoder, torch.tensor(__A ).transpose(0, 1 ).contiguous(), torch.tensor(__A ), ) def lowerCAmelCase_ ( __A, __A, __A ) -> str: '''simple docstring''' UpperCAmelCase__ = ReformerConfig.from_json_file(__A ) print(f"""Building PyTorch model from configuration: {config}""" ) UpperCAmelCase__ = ReformerModelWithLMHead(__A ) with open(__A, "rb" ) as f: UpperCAmelCase__ = pickle.load(__A )["weights"] set_model_weights_in_torch(__A, __A, config.hidden_size ) # Save pytorch-model print(f"""Save PyTorch model to {pytorch_dump_path}""" ) torch.save(model.state_dict(), __A ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--trax_model_pkl_path', default=None, type=str, required=True, help='Path to the TensorFlow checkpoint path.' ) parser.add_argument( '--config_file', default=None, type=str, required=True, help=( 'The config json file corresponding to the pre-trained Reformer model. \n' 'This specifies the model architecture.' ), ) parser.add_argument( '--pytorch_dump_path', default=None, type=str, required=True, help='Path to the output PyTorch model.' ) UpperCamelCase__ = parser.parse_args() convert_trax_checkpoint_to_pytorch(args.trax_model_pkl_path, args.config_file, args.pytorch_dump_path)
65
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
1
from typing import Optional import torch import torch.utils.checkpoint from torch import Tensor, nn from torch.nn import BCEWithLogitsLoss, CrossEntropyLoss, MSELoss from ...activations import ACTaFN from ...modeling_outputs import ( BackboneOutput, BaseModelOutputWithNoAttention, BaseModelOutputWithPoolingAndNoAttention, ImageClassifierOutputWithNoAttention, ) from ...modeling_utils import PreTrainedModel from ...utils import ( add_code_sample_docstrings, add_start_docstrings, add_start_docstrings_to_model_forward, logging, replace_return_docstrings, ) from ...utils.backbone_utils import BackboneMixin from .configuration_resnet import ResNetConfig UpperCamelCase__ = logging.get_logger(__name__) # General docstring UpperCamelCase__ = 'ResNetConfig' # Base docstring UpperCamelCase__ = 'microsoft/resnet-50' UpperCamelCase__ = [1, 2_0_4_8, 7, 7] # Image classification docstring UpperCamelCase__ = 'microsoft/resnet-50' UpperCamelCase__ = 'tiger cat' UpperCamelCase__ = [ 'microsoft/resnet-50', # See all resnet models at https://huggingface.co/models?filter=resnet ] class A ( nn.Module ): def __init__(self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int = 3 , __UpperCAmelCase : int = 1 , __UpperCAmelCase : str = "relu" ) -> int: """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , __UpperCAmelCase , kernel_size=__UpperCAmelCase , stride=__UpperCAmelCase , padding=kernel_size // 2 , bias=__UpperCAmelCase ) UpperCAmelCase__ = nn.BatchNormad(__UpperCAmelCase ) UpperCAmelCase__ = ACTaFN[activation] if activation is not None else nn.Identity() def lowercase_ (self : List[Any] , __UpperCAmelCase : Tensor ) -> Tensor: """simple docstring""" UpperCAmelCase__ = self.convolution(__UpperCAmelCase ) UpperCAmelCase__ = self.normalization(__UpperCAmelCase ) UpperCAmelCase__ = self.activation(__UpperCAmelCase ) return hidden_state class A ( nn.Module ): def __init__(self : str , __UpperCAmelCase : ResNetConfig ) -> Any: """simple docstring""" super().__init__() UpperCAmelCase__ = ResNetConvLayer( config.num_channels , config.embedding_size , kernel_size=7 , stride=2 , activation=config.hidden_act ) UpperCAmelCase__ = nn.MaxPoolad(kernel_size=3 , stride=2 , padding=1 ) UpperCAmelCase__ = config.num_channels def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Tensor ) -> Tensor: """simple docstring""" UpperCAmelCase__ = pixel_values.shape[1] if num_channels != self.num_channels: raise ValueError( "Make sure that the channel dimension of the pixel values match with the one set in the configuration." ) UpperCAmelCase__ = self.embedder(__UpperCAmelCase ) UpperCAmelCase__ = self.pooler(__UpperCAmelCase ) return embedding class A ( nn.Module ): def __init__(self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int = 2 ) -> Union[str, Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = nn.Convad(__UpperCAmelCase , __UpperCAmelCase , kernel_size=1 , stride=__UpperCAmelCase , bias=__UpperCAmelCase ) UpperCAmelCase__ = nn.BatchNormad(__UpperCAmelCase ) def lowercase_ (self : List[str] , __UpperCAmelCase : Tensor ) -> Tensor: """simple docstring""" UpperCAmelCase__ = self.convolution(__UpperCAmelCase ) UpperCAmelCase__ = self.normalization(__UpperCAmelCase ) return hidden_state class A ( nn.Module ): def __init__(self : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int = 1 , __UpperCAmelCase : str = "relu" ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = in_channels != out_channels or stride != 1 UpperCAmelCase__ = ( ResNetShortCut(__UpperCAmelCase , __UpperCAmelCase , stride=__UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) UpperCAmelCase__ = nn.Sequential( ResNetConvLayer(__UpperCAmelCase , __UpperCAmelCase , stride=__UpperCAmelCase ) , ResNetConvLayer(__UpperCAmelCase , __UpperCAmelCase , activation=__UpperCAmelCase ) , ) UpperCAmelCase__ = ACTaFN[activation] def lowercase_ (self : List[Any] , __UpperCAmelCase : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase__ = hidden_state UpperCAmelCase__ = self.layer(__UpperCAmelCase ) UpperCAmelCase__ = self.shortcut(__UpperCAmelCase ) hidden_state += residual UpperCAmelCase__ = self.activation(__UpperCAmelCase ) return hidden_state class A ( nn.Module ): def __init__(self : List[str] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int = 1 , __UpperCAmelCase : str = "relu" , __UpperCAmelCase : int = 4 ) -> Optional[Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = in_channels != out_channels or stride != 1 UpperCAmelCase__ = out_channels // reduction UpperCAmelCase__ = ( ResNetShortCut(__UpperCAmelCase , __UpperCAmelCase , stride=__UpperCAmelCase ) if should_apply_shortcut else nn.Identity() ) UpperCAmelCase__ = nn.Sequential( ResNetConvLayer(__UpperCAmelCase , __UpperCAmelCase , kernel_size=1 ) , ResNetConvLayer(__UpperCAmelCase , __UpperCAmelCase , stride=__UpperCAmelCase ) , ResNetConvLayer(__UpperCAmelCase , __UpperCAmelCase , kernel_size=1 , activation=__UpperCAmelCase ) , ) UpperCAmelCase__ = ACTaFN[activation] def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[int] ) -> str: """simple docstring""" UpperCAmelCase__ = hidden_state UpperCAmelCase__ = self.layer(__UpperCAmelCase ) UpperCAmelCase__ = self.shortcut(__UpperCAmelCase ) hidden_state += residual UpperCAmelCase__ = self.activation(__UpperCAmelCase ) return hidden_state class A ( nn.Module ): def __init__(self : Tuple , __UpperCAmelCase : ResNetConfig , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int = 2 , __UpperCAmelCase : int = 2 , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = ResNetBottleNeckLayer if config.layer_type == "bottleneck" else ResNetBasicLayer UpperCAmelCase__ = nn.Sequential( # downsampling is done in the first layer with stride of 2 layer(__UpperCAmelCase , __UpperCAmelCase , stride=__UpperCAmelCase , activation=config.hidden_act ) , *[layer(__UpperCAmelCase , __UpperCAmelCase , activation=config.hidden_act ) for _ in range(depth - 1 )] , ) def lowercase_ (self : int , __UpperCAmelCase : Tensor ) -> Tensor: """simple docstring""" UpperCAmelCase__ = input for layer in self.layers: UpperCAmelCase__ = layer(__UpperCAmelCase ) return hidden_state class A ( nn.Module ): def __init__(self : Optional[int] , __UpperCAmelCase : ResNetConfig ) -> Union[str, Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = nn.ModuleList([] ) # based on `downsample_in_first_stage` the first layer of the first stage may or may not downsample the input self.stages.append( ResNetStage( __UpperCAmelCase , config.embedding_size , config.hidden_sizes[0] , stride=2 if config.downsample_in_first_stage else 1 , depth=config.depths[0] , ) ) UpperCAmelCase__ = zip(config.hidden_sizes , config.hidden_sizes[1:] ) for (in_channels, out_channels), depth in zip(__UpperCAmelCase , config.depths[1:] ): self.stages.append(ResNetStage(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , depth=__UpperCAmelCase ) ) def lowercase_ (self : Any , __UpperCAmelCase : Tensor , __UpperCAmelCase : bool = False , __UpperCAmelCase : bool = True ) -> BaseModelOutputWithNoAttention: """simple docstring""" UpperCAmelCase__ = () if output_hidden_states else None for stage_module in self.stages: if output_hidden_states: UpperCAmelCase__ = hidden_states + (hidden_state,) UpperCAmelCase__ = stage_module(__UpperCAmelCase ) if output_hidden_states: UpperCAmelCase__ = hidden_states + (hidden_state,) if not return_dict: return tuple(v for v in [hidden_state, hidden_states] if v is not None ) return BaseModelOutputWithNoAttention( last_hidden_state=__UpperCAmelCase , hidden_states=__UpperCAmelCase , ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ResNetConfig __UpperCAmelCase : List[Any] = 'resnet' __UpperCAmelCase : List[str] = 'pixel_values' __UpperCAmelCase : int = True def lowercase_ (self : Optional[int] , __UpperCAmelCase : int ) -> Any: """simple docstring""" if isinstance(__UpperCAmelCase , nn.Convad ): nn.init.kaiming_normal_(module.weight , mode="fan_out" , nonlinearity="relu" ) elif isinstance(__UpperCAmelCase , (nn.BatchNormad, nn.GroupNorm) ): nn.init.constant_(module.weight , 1 ) nn.init.constant_(module.bias , 0 ) def lowercase_ (self : Any , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Tuple=False ) -> Optional[int]: """simple docstring""" if isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = value UpperCamelCase__ = R'\n This model is a PyTorch [torch.nn.Module](https://pytorch.org/docs/stable/nn.html#torch.nn.Module) subclass. Use it\n as a regular PyTorch Module and refer to the PyTorch documentation for all matter related to general usage and\n behavior.\n\n Parameters:\n config ([`ResNetConfig`]): Model configuration class with all the parameters of the model.\n Initializing with a config file does not load the weights associated with the model, only the\n configuration. Check out the [`~PreTrainedModel.from_pretrained`] method to load the model weights.\n' UpperCamelCase__ = R'\n Args:\n pixel_values (`torch.FloatTensor` of shape `(batch_size, num_channels, height, width)`):\n Pixel values. Pixel values can be obtained using [`AutoImageProcessor`]. See\n [`ConvNextImageProcessor.__call__`] for details.\n\n output_hidden_states (`bool`, *optional*):\n Whether or not to return the hidden states of all layers. See `hidden_states` under returned tensors for\n more detail.\n return_dict (`bool`, *optional*):\n Whether or not to return a [`~utils.ModelOutput`] instead of a plain tuple.\n' @add_start_docstrings( 'The bare ResNet model outputting raw features without any specific head on top.' , UpperCAmelCase_ , ) class A ( UpperCAmelCase_ ): def __init__(self : str , __UpperCAmelCase : Optional[Any] ) -> List[Any]: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = config UpperCAmelCase__ = ResNetEmbeddings(__UpperCAmelCase ) UpperCAmelCase__ = ResNetEncoder(__UpperCAmelCase ) UpperCAmelCase__ = nn.AdaptiveAvgPoolad((1, 1) ) # Initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_CHECKPOINT_FOR_DOC , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , modality="vision" , expected_output=_EXPECTED_OUTPUT_SHAPE , ) def lowercase_ (self : Any , __UpperCAmelCase : Tensor , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None ) -> BaseModelOutputWithPoolingAndNoAttention: """simple docstring""" UpperCAmelCase__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.embedder(__UpperCAmelCase ) UpperCAmelCase__ = self.encoder( __UpperCAmelCase , output_hidden_states=__UpperCAmelCase , return_dict=__UpperCAmelCase ) UpperCAmelCase__ = encoder_outputs[0] UpperCAmelCase__ = self.pooler(__UpperCAmelCase ) if not return_dict: return (last_hidden_state, pooled_output) + encoder_outputs[1:] return BaseModelOutputWithPoolingAndNoAttention( last_hidden_state=__UpperCAmelCase , pooler_output=__UpperCAmelCase , hidden_states=encoder_outputs.hidden_states , ) @add_start_docstrings( '\n ResNet Model with an image classification head on top (a linear layer on top of the pooled features), e.g. for\n ImageNet.\n ' , UpperCAmelCase_ , ) class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = config.num_labels UpperCAmelCase__ = ResNetModel(__UpperCAmelCase ) # classification head UpperCAmelCase__ = nn.Sequential( nn.Flatten() , nn.Linear(config.hidden_sizes[-1] , config.num_labels ) if config.num_labels > 0 else nn.Identity() , ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__UpperCAmelCase ) @add_code_sample_docstrings( checkpoint=_IMAGE_CLASS_CHECKPOINT , output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC , expected_output=_IMAGE_CLASS_EXPECTED_OUTPUT , ) def lowercase_ (self : List[str] , __UpperCAmelCase : Optional[torch.FloatTensor] = None , __UpperCAmelCase : Optional[torch.LongTensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> ImageClassifierOutputWithNoAttention: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.resnet(__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , return_dict=__UpperCAmelCase ) UpperCAmelCase__ = outputs.pooler_output if return_dict else outputs[1] UpperCAmelCase__ = self.classifier(__UpperCAmelCase ) UpperCAmelCase__ = None if labels is not None: if self.config.problem_type is None: if self.num_labels == 1: UpperCAmelCase__ = "regression" elif self.num_labels > 1 and (labels.dtype == torch.long or labels.dtype == torch.int): UpperCAmelCase__ = "single_label_classification" else: UpperCAmelCase__ = "multi_label_classification" if self.config.problem_type == "regression": UpperCAmelCase__ = MSELoss() if self.num_labels == 1: UpperCAmelCase__ = loss_fct(logits.squeeze() , labels.squeeze() ) else: UpperCAmelCase__ = loss_fct(__UpperCAmelCase , __UpperCAmelCase ) elif self.config.problem_type == "single_label_classification": UpperCAmelCase__ = CrossEntropyLoss() UpperCAmelCase__ = loss_fct(logits.view(-1 , self.num_labels ) , labels.view(-1 ) ) elif self.config.problem_type == "multi_label_classification": UpperCAmelCase__ = BCEWithLogitsLoss() UpperCAmelCase__ = loss_fct(__UpperCAmelCase , __UpperCAmelCase ) if not return_dict: UpperCAmelCase__ = (logits,) + outputs[2:] return (loss,) + output if loss is not None else output return ImageClassifierOutputWithNoAttention(loss=__UpperCAmelCase , logits=__UpperCAmelCase , hidden_states=outputs.hidden_states ) @add_start_docstrings( '\n ResNet backbone, to be used with frameworks like DETR and MaskFormer.\n ' , UpperCAmelCase_ , ) class A ( UpperCAmelCase_ , UpperCAmelCase_ ): def __init__(self : Optional[Any] , __UpperCAmelCase : Any ) -> str: """simple docstring""" super().__init__(__UpperCAmelCase ) super()._init_backbone(__UpperCAmelCase ) UpperCAmelCase__ = [config.embedding_size] + config.hidden_sizes UpperCAmelCase__ = ResNetEmbeddings(__UpperCAmelCase ) UpperCAmelCase__ = ResNetEncoder(__UpperCAmelCase ) # initialize weights and apply final processing self.post_init() @add_start_docstrings_to_model_forward(__UpperCAmelCase ) @replace_return_docstrings(output_type=__UpperCAmelCase , config_class=_CONFIG_FOR_DOC ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Tensor , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None ) -> BackboneOutput: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = ( output_hidden_states if output_hidden_states is not None else self.config.output_hidden_states ) UpperCAmelCase__ = self.embedder(__UpperCAmelCase ) UpperCAmelCase__ = self.encoder(__UpperCAmelCase , output_hidden_states=__UpperCAmelCase , return_dict=__UpperCAmelCase ) UpperCAmelCase__ = outputs.hidden_states UpperCAmelCase__ = () for idx, stage in enumerate(self.stage_names ): if stage in self.out_features: feature_maps += (hidden_states[idx],) if not return_dict: UpperCAmelCase__ = (feature_maps,) if output_hidden_states: output += (outputs.hidden_states,) return output return BackboneOutput( feature_maps=__UpperCAmelCase , hidden_states=outputs.hidden_states if output_hidden_states else None , attentions=__UpperCAmelCase , )
65
def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' def get_matched_characters(__A, __A ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase__ = int(max(0, i - limit ) ) UpperCAmelCase__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(__A ) UpperCAmelCase__ = f"""{_stra[0:_stra.index(__A )]} {_stra[_stra.index(__A ) + 1:]}""" return "".join(__A ) # matching characters UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = len(__A ) # transposition UpperCAmelCase__ = ( len([(ca, ca) for ca, ca in zip(__A, __A ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase__ = 0.0 else: UpperCAmelCase__ = ( 1 / 3 * ( match_count / len(__A ) + match_count / len(__A ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
65
1
import itertools from dataclasses import dataclass from typing import List, Optional import pyarrow as pa import pyarrow.parquet as pq import datasets from datasets.table import table_cast UpperCamelCase__ = datasets.utils.logging.get_logger(__name__) @dataclass class A ( datasets.BuilderConfig ): __UpperCAmelCase : int = 1_00_00 __UpperCAmelCase : Optional[List[str]] = None __UpperCAmelCase : Optional[datasets.Features] = None class A ( datasets.ArrowBasedBuilder ): __UpperCAmelCase : List[Any] = ParquetConfig def lowercase_ (self : Optional[Any] ) -> Optional[Any]: """simple docstring""" return datasets.DatasetInfo(features=self.config.features ) def lowercase_ (self : Dict , __UpperCAmelCase : int ) -> Union[str, Any]: """simple docstring""" if not self.config.data_files: raise ValueError(f"""At least one data file must be specified, but got data_files={self.config.data_files}""" ) UpperCAmelCase__ = dl_manager.download_and_extract(self.config.data_files ) if isinstance(__UpperCAmelCase , (str, list, tuple) ): UpperCAmelCase__ = data_files if isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase__ = [dl_manager.iter_files(__UpperCAmelCase ) for file in files] return [datasets.SplitGenerator(name=datasets.Split.TRAIN , gen_kwargs={"files": files} )] UpperCAmelCase__ = [] for split_name, files in data_files.items(): if isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = [files] # Use `dl_manager.iter_files` to skip hidden files in an extracted archive UpperCAmelCase__ = [dl_manager.iter_files(__UpperCAmelCase ) for file in files] # Infer features is they are stoed in the arrow schema if self.info.features is None: for file in itertools.chain.from_iterable(__UpperCAmelCase ): with open(__UpperCAmelCase , "rb" ) as f: UpperCAmelCase__ = datasets.Features.from_arrow_schema(pq.read_schema(__UpperCAmelCase ) ) break splits.append(datasets.SplitGenerator(name=__UpperCAmelCase , gen_kwargs={"files": files} ) ) return splits def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : pa.Table ) -> pa.Table: """simple docstring""" if self.info.features is not None: # more expensive cast to support nested features with keys in a different order # allows str <-> int/float or str to Audio for example UpperCAmelCase__ = table_cast(__UpperCAmelCase , self.info.features.arrow_schema ) return pa_table def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.info.features.arrow_schema if self.info.features is not None else None if self.info.features is not None and self.config.columns is not None: if sorted(field.name for field in schema ) != sorted(self.config.columns ): raise ValueError( f"""Tried to load parquet data with columns '{self.config.columns}' with mismatching features '{self.info.features}'""" ) for file_idx, file in enumerate(itertools.chain.from_iterable(__UpperCAmelCase ) ): with open(__UpperCAmelCase , "rb" ) as f: UpperCAmelCase__ = pq.ParquetFile(__UpperCAmelCase ) try: for batch_idx, record_batch in enumerate( parquet_file.iter_batches(batch_size=self.config.batch_size , columns=self.config.columns ) ): UpperCAmelCase__ = pa.Table.from_batches([record_batch] ) # Uncomment for debugging (will print the Arrow table size and elements) # logger.warning(f"pa_table: {pa_table} num rows: {pa_table.num_rows}") # logger.warning('\n'.join(str(pa_table.slice(i, 1).to_pydict()) for i in range(pa_table.num_rows))) yield f"""{file_idx}_{batch_idx}""", self._cast_table(__UpperCAmelCase ) except ValueError as e: logger.error(f"""Failed to read file '{file}' with error {type(__UpperCAmelCase )}: {e}""" ) raise
65
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
1
import inspect import unittest from transformers import RegNetConfig from transformers.file_utils import cached_property, is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision, slow, torch_device from ...test_configuration_common import ConfigTester from ...test_modeling_common import ModelTesterMixin, floats_tensor, ids_tensor from ...test_pipeline_mixin import PipelineTesterMixin if is_torch_available(): import torch from torch import nn from transformers import RegNetForImageClassification, RegNetModel from transformers.models.regnet.modeling_regnet import REGNET_PRETRAINED_MODEL_ARCHIVE_LIST if is_vision_available(): from PIL import Image from transformers import AutoImageProcessor class A : def __init__(self : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : int=3_2 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : str=1_0 , __UpperCAmelCase : Dict=[1_0, 2_0, 3_0, 4_0] , __UpperCAmelCase : List[Any]=[1, 1, 2, 1] , __UpperCAmelCase : int=True , __UpperCAmelCase : List[str]=True , __UpperCAmelCase : Optional[int]="relu" , __UpperCAmelCase : Tuple=3 , __UpperCAmelCase : Any=None , ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = embeddings_size UpperCAmelCase__ = hidden_sizes UpperCAmelCase__ = depths UpperCAmelCase__ = is_training UpperCAmelCase__ = use_labels UpperCAmelCase__ = hidden_act UpperCAmelCase__ = num_labels UpperCAmelCase__ = scope UpperCAmelCase__ = len(__UpperCAmelCase ) def lowercase_ (self : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.num_labels ) UpperCAmelCase__ = self.get_config() return config, pixel_values, labels def lowercase_ (self : List[str] ) -> Optional[Any]: """simple docstring""" return RegNetConfig( num_channels=self.num_channels , embeddings_size=self.embeddings_size , hidden_sizes=self.hidden_sizes , depths=self.depths , hidden_act=self.hidden_act , num_labels=self.num_labels , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = RegNetModel(config=__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() UpperCAmelCase__ = model(__UpperCAmelCase ) # expected last hidden states: B, C, H // 32, W // 32 self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.hidden_sizes[-1], self.image_size // 3_2, self.image_size // 3_2) , ) def lowercase_ (self : Tuple , __UpperCAmelCase : Any , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = RegNetForImageClassification(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() UpperCAmelCase__ = model(__UpperCAmelCase , labels=__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ (self : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {"pixel_values": pixel_values} return config, inputs_dict @require_torch class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : str = (RegNetModel, RegNetForImageClassification) if is_torch_available() else () __UpperCAmelCase : Dict = ( {'feature-extraction': RegNetModel, 'image-classification': RegNetForImageClassification} if is_torch_available() else {} ) __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Dict = False __UpperCAmelCase : str = False __UpperCAmelCase : Any = False def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" UpperCAmelCase__ = RegNetModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__UpperCAmelCase , has_text_modality=__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> Dict: """simple docstring""" self.create_and_test_config_common_properties() self.config_tester.create_and_test_config_to_json_string() self.config_tester.create_and_test_config_to_json_file() self.config_tester.create_and_test_config_from_and_save_pretrained() self.config_tester.create_and_test_config_with_num_labels() self.config_tester.check_config_can_be_init_without_params() self.config_tester.check_config_arguments_init() def lowercase_ (self : List[Any] ) -> Dict: """simple docstring""" return @unittest.skip(reason="RegNet does not use inputs_embeds" ) def lowercase_ (self : Tuple ) -> Any: """simple docstring""" pass @unittest.skip(reason="RegNet does not support input and output embeddings" ) def lowercase_ (self : List[Any] ) -> int: """simple docstring""" pass def lowercase_ (self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = inspect.signature(model.forward ) # signature.parameters is an OrderedDict => so arg_names order is deterministic UpperCAmelCase__ = [*signature.parameters.keys()] UpperCAmelCase__ = ["pixel_values"] self.assertListEqual(arg_names[:1] , __UpperCAmelCase ) def lowercase_ (self : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowercase_ (self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(config=__UpperCAmelCase ) for name, module in model.named_modules(): if isinstance(__UpperCAmelCase , (nn.BatchNormad, nn.GroupNorm) ): self.assertTrue( torch.all(module.weight == 1 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) self.assertTrue( torch.all(module.bias == 0 ) , msg=f"""Parameter {name} of model {model_class} seems not properly initialized""" , ) def lowercase_ (self : Dict ) -> Optional[Any]: """simple docstring""" def check_hidden_states_output(__UpperCAmelCase : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] ): UpperCAmelCase__ = model_class(__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.eval() with torch.no_grad(): UpperCAmelCase__ = model(**self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) UpperCAmelCase__ = outputs.encoder_hidden_states if config.is_encoder_decoder else outputs.hidden_states UpperCAmelCase__ = self.model_tester.num_stages self.assertEqual(len(__UpperCAmelCase ) , expected_num_stages + 1 ) # RegNet's feature maps are of shape (batch_size, num_channels, height, width) self.assertListEqual( list(hidden_states[0].shape[-2:] ) , [self.model_tester.image_size // 2, self.model_tester.image_size // 2] , ) UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = ["basic", "bottleneck"] for model_class in self.all_model_classes: for layer_type in layers_type: UpperCAmelCase__ = layer_type UpperCAmelCase__ = True check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) # check that output_hidden_states also work using config del inputs_dict["output_hidden_states"] UpperCAmelCase__ = True check_hidden_states_output(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Dict ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_image_classification(*__UpperCAmelCase ) @slow def lowercase_ (self : int ) -> Dict: """simple docstring""" for model_name in REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = RegNetModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_torch @require_vision class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Any ) -> Union[str, Any]: """simple docstring""" return ( AutoImageProcessor.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ) if is_vision_available() else None ) @slow def lowercase_ (self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase__ = RegNetForImageClassification.from_pretrained(REGNET_PRETRAINED_MODEL_ARCHIVE_LIST[0] ).to(__UpperCAmelCase ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=__UpperCAmelCase , return_tensors="pt" ).to(__UpperCAmelCase ) # forward pass with torch.no_grad(): UpperCAmelCase__ = model(**__UpperCAmelCase ) # verify the logits UpperCAmelCase__ = torch.Size((1, 1_0_0_0) ) self.assertEqual(outputs.logits.shape , __UpperCAmelCase ) UpperCAmelCase__ = torch.tensor([-0.4180, -1.5051, -3.4836] ).to(__UpperCAmelCase ) self.assertTrue(torch.allclose(outputs.logits[0, :3] , __UpperCAmelCase , atol=1E-4 ) )
65
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCamelCase__ = 'base_with_context' def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["self_attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["MultiHeadDotProductAttention_0"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array, __A ) UpperCAmelCase__ = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] UpperCAmelCase__ = os.path.join(args.checkpoint_path, "..", "config.gin" ) UpperCAmelCase__ = inference.parse_training_gin_file(__A, __A ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path, __A ) UpperCAmelCase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2", variance_type="fixed_large" ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"], vocab_size=synth_model.model.module.config.vocab_size, d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims, targets_context_length=synth_model.sequence_length["targets_context"], d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims, targets_length=synth_model.sequence_length["targets_context"], max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time, d_model=synth_model.model.module.config.emb_dim, num_layers=synth_model.model.module.config.num_decoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, dropout_rate=synth_model.model.module.config.dropout_rate, ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint["target"]["token_encoder"], __A ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"], __A ) UpperCAmelCase__ = load_decoder(ta_checkpoint["target"]["decoder"], __A ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=__A, continuous_encoder=__A, decoder=__A, scheduler=__A, melgan=__A, ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help='Path to the original jax model checkpoint.', ) UpperCamelCase__ = parser.parse_args() main(args)
65
1
def lowerCAmelCase_ ( __A, __A, __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = len(__A ), len(grid[0] ) if ( min(__A, __A ) < 0 or row == row_length or col == col_length or (row, col) in visit or grid[row][col] == 1 ): return 0 if row == row_length - 1 and col == col_length - 1: return 1 visit.add((row, col) ) UpperCAmelCase__ = 0 count += depth_first_search(__A, row + 1, __A, __A ) count += depth_first_search(__A, row - 1, __A, __A ) count += depth_first_search(__A, __A, col + 1, __A ) count += depth_first_search(__A, __A, col - 1, __A ) visit.remove((row, col) ) return count if __name__ == "__main__": import doctest doctest.testmod()
65
import math def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' return math.sqrt(__A ) * math.sqrt(__A ) == num def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' UpperCAmelCase__ = 0 UpperCAmelCase__ = n while left <= right: UpperCAmelCase__ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase__ = mid - 1 else: UpperCAmelCase__ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
65
1
import argparse from pathlib import Path from transformers import AutoConfig, AutoTokenizer, RagConfig, RagSequenceForGeneration, RagTokenForGeneration def lowerCAmelCase_ ( __A, __A, __A, __A, __A = None, __A = None, __A = None, ) -> str: '''simple docstring''' if config_name_or_path is None: UpperCAmelCase__ = "facebook/rag-token-base" if model_type == "rag_token" else "facebook/rag-sequence-base" if generator_tokenizer_name_or_path is None: UpperCAmelCase__ = generator_name_or_path if question_encoder_tokenizer_name_or_path is None: UpperCAmelCase__ = question_encoder_name_or_path UpperCAmelCase__ = RagTokenForGeneration if model_type == "rag_token" else RagSequenceForGeneration # Save model. UpperCAmelCase__ = RagConfig.from_pretrained(__A ) UpperCAmelCase__ = AutoConfig.from_pretrained(__A ) UpperCAmelCase__ = AutoConfig.from_pretrained(__A ) UpperCAmelCase__ = gen_config UpperCAmelCase__ = question_encoder_config UpperCAmelCase__ = model_class.from_pretrained_question_encoder_generator( __A, __A, config=__A ) rag_model.save_pretrained(__A ) # Sanity check. model_class.from_pretrained(__A ) # Save tokenizers. UpperCAmelCase__ = AutoTokenizer.from_pretrained(__A ) gen_tokenizer.save_pretrained(dest_dir / "generator_tokenizer/" ) UpperCAmelCase__ = AutoTokenizer.from_pretrained(__A ) question_encoder_tokenizer.save_pretrained(dest_dir / "question_encoder_tokenizer/" ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument( '--model_type', choices=['rag_sequence', 'rag_token'], required=True, type=str, help='RAG model type: rag_sequence, rag_token', ) parser.add_argument('--dest', type=str, required=True, help='Path to the output checkpoint directory.') parser.add_argument('--generator_name_or_path', type=str, required=True, help='Generator model identifier') parser.add_argument( '--question_encoder_name_or_path', type=str, required=True, help='Question encoder model identifier' ) parser.add_argument( '--generator_tokenizer_name_or_path', type=str, help='Generator tokenizer identifier, if not specified, resolves to ``generator_name_or_path``', ) parser.add_argument( '--question_encoder_tokenizer_name_or_path', type=str, help='Question encoder tokenizer identifier, if not specified, resolves to ``question_encoder_name_or_path``', ) parser.add_argument( '--config_name_or_path', type=str, help=( 'Identifier of the model config to use, if not provided, resolves to a base config for a given' ' ``model_type``' ), ) UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = Path(args.dest) dest_dir.mkdir(exist_ok=True) consolidate( args.model_type, args.generator_name_or_path, args.question_encoder_name_or_path, dest_dir, args.config_name_or_path, args.generator_tokenizer_name_or_path, args.question_encoder_tokenizer_name_or_path, )
65
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor __UpperCAmelCase : Optional[torch.FloatTensor] = None def lowerCAmelCase_ ( __A, __A=0.999, __A="cosine", ) -> Tuple: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__A ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase__ = [] for i in range(__A ): UpperCAmelCase__ = i / num_diffusion_timesteps UpperCAmelCase__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__A ) / alpha_bar_fn(__A ), __A ) ) return torch.tensor(__A, dtype=torch.floataa ) class A ( UpperCAmelCase_ , UpperCAmelCase_ ): @register_to_config def __init__(self : List[str] , __UpperCAmelCase : int = 1_0_0_0 , __UpperCAmelCase : str = "fixed_small_log" , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[float] = 1.0 , __UpperCAmelCase : str = "epsilon" , __UpperCAmelCase : str = "squaredcos_cap_v2" , ) -> Optional[int]: """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) UpperCAmelCase__ = betas_for_alpha_bar(__UpperCAmelCase ) UpperCAmelCase__ = 1.0 - self.betas UpperCAmelCase__ = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase__ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase__ = 1.0 # setable values UpperCAmelCase__ = None UpperCAmelCase__ = torch.from_numpy(np.arange(0 , __UpperCAmelCase )[::-1].copy() ) UpperCAmelCase__ = variance_type def lowercase_ (self : List[str] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None ) -> torch.FloatTensor: """simple docstring""" return sample def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, torch.device] = None ) -> Any: """simple docstring""" UpperCAmelCase__ = num_inference_steps UpperCAmelCase__ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase__ = (np.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ).to(__UpperCAmelCase ) def lowercase_ (self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase__ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase__ = torch.log(torch.clamp(__UpperCAmelCase , min=1E-20 ) ) UpperCAmelCase__ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase__ = variance.log() UpperCAmelCase__ = beta.log() UpperCAmelCase__ = (predicted_variance + 1) / 2 UpperCAmelCase__ = frac * max_log + (1 - frac) * min_log return variance def lowercase_ (self : Optional[int] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : int , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : bool = True , ) -> Union[UnCLIPSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase__ , UpperCAmelCase__ = torch.split(__UpperCAmelCase , sample.shape[1] , dim=1 ) else: UpperCAmelCase__ = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] UpperCAmelCase__ = self.alphas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase__ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase__ = model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase__ = torch.clamp( __UpperCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase__ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase__ = 0 if t > 0: UpperCAmelCase__ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__UpperCAmelCase , device=model_output.device ) UpperCAmelCase__ = self._get_variance( __UpperCAmelCase , predicted_variance=__UpperCAmelCase , prev_timestep=__UpperCAmelCase , ) if self.variance_type == "fixed_small_log": UpperCAmelCase__ = variance elif self.variance_type == "learned_range": UpperCAmelCase__ = (0.5 * variance).exp() else: raise ValueError( f"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" " for the UnCLIPScheduler." ) UpperCAmelCase__ = variance * variance_noise UpperCAmelCase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__UpperCAmelCase , pred_original_sample=__UpperCAmelCase ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.IntTensor , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase__ = timesteps.to(original_samples.device ) UpperCAmelCase__ = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase__ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase__ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
65
1
import functools from typing import Any def lowerCAmelCase_ ( __A, __A ) -> bool: '''simple docstring''' if not isinstance(__A, __A ) or len(__A ) == 0: raise ValueError("the string should be not empty string" ) if not isinstance(__A, __A ) or not all( isinstance(__A, __A ) and len(__A ) > 0 for item in words ): raise ValueError("the words should be a list of non-empty strings" ) # Build trie UpperCAmelCase__ = {} UpperCAmelCase__ = "WORD_KEEPER" for word in words: UpperCAmelCase__ = trie for c in word: if c not in trie_node: UpperCAmelCase__ = {} UpperCAmelCase__ = trie_node[c] UpperCAmelCase__ = True UpperCAmelCase__ = len(__A ) # Dynamic programming method @functools.cache def is_breakable(__A ) -> bool: if index == len_string: return True UpperCAmelCase__ = trie for i in range(__A, __A ): UpperCAmelCase__ = trie_node.get(string[i], __A ) if trie_node is None: return False if trie_node.get(__A, __A ) and is_breakable(i + 1 ): return True return False return is_breakable(0 ) if __name__ == "__main__": import doctest doctest.testmod()
65
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class A ( unittest.TestCase ): def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) UpperCAmelCase__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def lowercase_ (self : List[str] ) -> Any: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : str ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.operation_file_path] print(f"""Command: {cmd}""" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Tuple ) -> int: """simple docstring""" UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Dict ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices, using 2 devices only""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase__ = Accelerator() UpperCamelCase__ = (accelerator.state.process_index + 2, 1_0) UpperCamelCase__ = torch.randint(0, 1_0, shape).to(accelerator.device) UpperCamelCase__ = '' UpperCamelCase__ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." UpperCamelCase__ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." UpperCamelCase__ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
65
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_xlnet import XLNetTokenizer else: UpperCamelCase__ = None UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'spiece.model', 'tokenizer_file': 'tokenizer.json'} UpperCamelCase__ = { 'vocab_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/spiece.model', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/spiece.model', }, 'tokenizer_file': { 'xlnet-base-cased': 'https://huggingface.co/xlnet-base-cased/resolve/main/tokenizer.json', 'xlnet-large-cased': 'https://huggingface.co/xlnet-large-cased/resolve/main/tokenizer.json', }, } UpperCamelCase__ = { 'xlnet-base-cased': None, 'xlnet-large-cased': None, } UpperCamelCase__ = '▁' # Segments (not really needed) UpperCamelCase__ = 0 UpperCamelCase__ = 1 UpperCamelCase__ = 2 UpperCamelCase__ = 3 UpperCamelCase__ = 4 class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = VOCAB_FILES_NAMES __UpperCAmelCase : List[str] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : int = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : str = 'left' __UpperCAmelCase : List[str] = XLNetTokenizer def __init__(self : Union[str, Any] , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : List[Any]=False , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : Union[str, Any]=False , __UpperCAmelCase : str="<s>" , __UpperCAmelCase : Tuple="</s>" , __UpperCAmelCase : List[Any]="<unk>" , __UpperCAmelCase : int="<sep>" , __UpperCAmelCase : Optional[Any]="<pad>" , __UpperCAmelCase : str="<cls>" , __UpperCAmelCase : Union[str, Any]="<mask>" , __UpperCAmelCase : Optional[int]=["<eop>", "<eod>"] , **__UpperCAmelCase : Any , ) -> Any: """simple docstring""" UpperCAmelCase__ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token super().__init__( vocab_file=__UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = 3 UpperCAmelCase__ = do_lower_case UpperCAmelCase__ = remove_space UpperCAmelCase__ = keep_accents UpperCAmelCase__ = vocab_file UpperCAmelCase__ = False if not self.vocab_file else True def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowercase_ (self : str , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowercase_ (self : int , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ): copyfile(self.vocab_file , __UpperCAmelCase ) return (out_vocab_file,)
65
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(__A, __A ) def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = emb.weight.shape UpperCAmelCase__ = nn.Linear(__A, __A, bias=__A ) UpperCAmelCase__ = emb.weight.data return lin_layer def lowerCAmelCase_ ( __A, __A="facebook/mbart-large-en-ro", __A=False, __A=False ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = torch.load(__A, map_location="cpu" )["model"] remove_ignore_keys_(__A ) UpperCAmelCase__ = state_dict["encoder.embed_tokens.weight"].shape[0] UpperCAmelCase__ = MBartConfig.from_pretrained(__A, vocab_size=__A ) if mbart_aa and finetuned: UpperCAmelCase__ = "relu" UpperCAmelCase__ = state_dict["decoder.embed_tokens.weight"] UpperCAmelCase__ = MBartForConditionalGeneration(__A ) model.model.load_state_dict(__A ) if finetuned: UpperCAmelCase__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
65
1
def lowerCAmelCase_ ( __A ) -> list: '''simple docstring''' if any(not isinstance(__A, __A ) or x < 0 for x in sequence ): raise TypeError("Sequence must be list of non-negative integers" ) for _ in range(len(__A ) ): for i, (rod_upper, rod_lower) in enumerate(zip(__A, sequence[1:] ) ): if rod_upper > rod_lower: sequence[i] -= rod_upper - rod_lower sequence[i + 1] += rod_upper - rod_lower return sequence if __name__ == "__main__": assert bead_sort([5, 4, 3, 2, 1]) == [1, 2, 3, 4, 5] assert bead_sort([7, 9, 4, 3, 5]) == [3, 4, 5, 7, 9]
65
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase__ = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def lowerCAmelCase_ ( __A, __A=None ) -> Dict: '''simple docstring''' require_version(deps[pkg], __A )
65
1
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path", type=__A, default="data/dump.txt", help="The path to the data." ) parser.add_argument("--tokenizer_type", type=__A, default="bert", choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name", type=__A, default="bert-base-uncased", help="The tokenizer to use." ) parser.add_argument("--dump_file", type=__A, default="data/dump", help="The dump file prefix." ) UpperCAmelCase__ = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": UpperCAmelCase__ = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `[CLS]` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCAmelCase__ = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `<s>` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": UpperCAmelCase__ = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` UpperCAmelCase__ = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path, "r", encoding="utf8" ) as fp: UpperCAmelCase__ = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__A )} examples to process.""" ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 10_000 UpperCAmelCase__ = time.time() for text in data: UpperCAmelCase__ = f"""{bos} {text.strip()} {sep}""" UpperCAmelCase__ = tokenizer.encode(__A, add_special_tokens=__A ) rslt.append(__A ) iter += 1 if iter % interval == 0: UpperCAmelCase__ = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) UpperCAmelCase__ = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__A )} examples processed.""" ) UpperCAmelCase__ = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" UpperCAmelCase__ = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCAmelCase__ = [np.uintaa(__A ) for d in rslt] else: UpperCAmelCase__ = [np.intaa(__A ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__A, "wb" ) as handle: pickle.dump(rslt_, __A, protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
65
1
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
from manim import * class A ( UpperCAmelCase_ ): def lowercase_ (self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase__ = Rectangle(height=0.25 , width=0.25 ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("CPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(4 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("GPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) gpu.move_to([-1, -1, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Model" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) model.move_to([3, -1.0, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = fill.copy().set_fill(__UpperCAmelCase , opacity=0.8 ) target.move_to(__UpperCAmelCase ) model_arr.append(__UpperCAmelCase ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(__UpperCAmelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__UpperCAmelCase ) self.add(*__UpperCAmelCase , *__UpperCAmelCase ) UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Disk" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) disk.move_to([-4, -1.25, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(__UpperCAmelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase ) ) UpperCAmelCase__ = Square(0.3 ) input.set_fill(__UpperCAmelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , __UpperCAmelCase , buff=0.5 ) self.play(Write(__UpperCAmelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=__UpperCAmelCase , buff=0.02 ) self.play(MoveToTarget(__UpperCAmelCase ) ) self.play(FadeOut(__UpperCAmelCase ) ) UpperCAmelCase__ = Arrow(start=__UpperCAmelCase , end=__UpperCAmelCase , color=__UpperCAmelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , __UpperCAmelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) UpperCAmelCase__ = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) ) UpperCAmelCase__ = {"run_time": 1, "fade_in": True, "fade_out": True, "buff": 0.02} self.play( Write(__UpperCAmelCase ) , Circumscribe(model_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_cpu_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) UpperCAmelCase__ = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , __UpperCAmelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) UpperCAmelCase__ = AnimationGroup( FadeOut(__UpperCAmelCase , run_time=0.5 ) , MoveToTarget(__UpperCAmelCase , run_time=0.5 ) , FadeIn(__UpperCAmelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(__UpperCAmelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: UpperCAmelCase__ = 0.7 self.play( Circumscribe(model_arr[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_arr[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) UpperCAmelCase__ = a_c UpperCAmelCase__ = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(__UpperCAmelCase ) , FadeOut(__UpperCAmelCase , run_time=0.5 ) , ) UpperCAmelCase__ = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) , MoveToTarget(__UpperCAmelCase ) ) self.wait()
65
1
import os # All paths are set with the intent you should run this script from the root of the repo with the command # python utils/check_doctest_list.py UpperCamelCase__ = '.' if __name__ == "__main__": UpperCamelCase__ = os.path.join(REPO_PATH, 'utils/documentation_tests.txt') UpperCamelCase__ = [] UpperCamelCase__ = [] with open(doctest_file_path) as fp: for line in fp: UpperCamelCase__ = line.strip() UpperCamelCase__ = os.path.join(REPO_PATH, line) if not (os.path.isfile(path) or os.path.isdir(path)): non_existent_paths.append(line) all_paths.append(path) if len(non_existent_paths) > 0: UpperCamelCase__ = '\n'.join(non_existent_paths) raise ValueError(f'''`utils/documentation_tests.txt` contains non-existent paths:\n{non_existent_paths}''') if all_paths != sorted(all_paths): raise ValueError('Files in `utils/documentation_tests.txt` are not in alphabetical order.')
65
from __future__ import annotations from scipy.special import comb # type: ignore class A : def __init__(self : List[Any] , __UpperCAmelCase : list[tuple[float, float]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. UpperCAmelCase__ = len(__UpperCAmelCase ) - 1 def lowercase_ (self : int , __UpperCAmelCase : float ) -> list[float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __UpperCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__UpperCAmelCase ) , 5 ) == 1 return output_values def lowercase_ (self : Dict , __UpperCAmelCase : float ) -> tuple[float, float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = self.basis_function(__UpperCAmelCase ) UpperCAmelCase__ = 0.0 UpperCAmelCase__ = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def lowercase_ (self : Optional[int] , __UpperCAmelCase : float = 0.01 ) -> Optional[int]: """simple docstring""" from matplotlib import pyplot as plt # type: ignore UpperCAmelCase__ = [] # x coordinates of points to plot UpperCAmelCase__ = [] # y coordinates of points to plot UpperCAmelCase__ = 0.0 while t <= 1: UpperCAmelCase__ = self.bezier_curve_function(__UpperCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size UpperCAmelCase__ = [i[0] for i in self.list_of_points] UpperCAmelCase__ = [i[1] for i in self.list_of_points] plt.plot( __UpperCAmelCase , __UpperCAmelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__UpperCAmelCase , __UpperCAmelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
65
1
from collections import OrderedDict from typing import Any, Mapping, Optional from ... import PreTrainedTokenizer, TensorType, is_torch_available from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfigWithPast from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'EleutherAI/gpt-neo-1.3B': 'https://huggingface.co/EleutherAI/gpt-neo-1.3B/resolve/main/config.json', # See all GPTNeo models at https://huggingface.co/models?filter=gpt_neo } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = 'gpt_neo' __UpperCAmelCase : Any = ['past_key_values'] __UpperCAmelCase : List[Any] = {'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers'} def __init__(self : Optional[Any] , __UpperCAmelCase : Any=5_0_2_5_7 , __UpperCAmelCase : List[Any]=2_0_4_8 , __UpperCAmelCase : int=2_0_4_8 , __UpperCAmelCase : Tuple=2_4 , __UpperCAmelCase : List[str]=[[["global", "local"], 1_2]] , __UpperCAmelCase : Union[str, Any]=1_6 , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : List[Any]=2_5_6 , __UpperCAmelCase : List[Any]="gelu_new" , __UpperCAmelCase : Optional[Any]=0.0 , __UpperCAmelCase : Optional[int]=0.0 , __UpperCAmelCase : Any=0.0 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : str=1E-5 , __UpperCAmelCase : List[str]=0.02 , __UpperCAmelCase : Optional[Any]=True , __UpperCAmelCase : Optional[int]=5_0_2_5_6 , __UpperCAmelCase : int=5_0_2_5_6 , **__UpperCAmelCase : str , ) -> int: """simple docstring""" UpperCAmelCase__ = vocab_size UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_layers UpperCAmelCase__ = num_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = window_size UpperCAmelCase__ = activation_function UpperCAmelCase__ = resid_dropout UpperCAmelCase__ = embed_dropout UpperCAmelCase__ = attention_dropout UpperCAmelCase__ = classifier_dropout UpperCAmelCase__ = layer_norm_epsilon UpperCAmelCase__ = initializer_range UpperCAmelCase__ = use_cache UpperCAmelCase__ = bos_token_id UpperCAmelCase__ = eos_token_id UpperCAmelCase__ = attention_types UpperCAmelCase__ = self.expand_attention_types_params(__UpperCAmelCase ) if len(self.attention_layers ) != self.num_layers: raise ValueError( "Configuration for convolutional module is incorrect. " "It is required that `len(config.attention_layers)` == `config.num_layers` " f"""but is `len(config.attention_layers) = {len(self.attention_layers )}`, """ f"""`config.num_layers = {self.num_layers}`. """ "`config.attention_layers` is prepared using `config.attention_types`. " "Please verify the value of `config.attention_types` argument." ) super().__init__(bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) @staticmethod def lowercase_ (__UpperCAmelCase : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] for item in attention_types: for _ in range(item[1] ): attentions.extend(item[0] ) return attentions def lowerCAmelCase_ ( __A, __A, __A, __A ) -> int: '''simple docstring''' import torch UpperCAmelCase__ = input.size() UpperCAmelCase__ = len(__A ) UpperCAmelCase__ = shape[dimension] UpperCAmelCase__ = torch.arange(0, __A, __A ) UpperCAmelCase__ = torch.div(sizedim - size, __A, rounding_mode="floor" ) + 1 UpperCAmelCase__ = torch.arange(__A ) + low_indices[:min_length][:, None] UpperCAmelCase__ = [slice(__A )] * rank UpperCAmelCase__ = indices UpperCAmelCase__ = input[s] UpperCAmelCase__ = list(range(0, rank + 1 ) ) perm.append(perm.pop(dimension + 1 ) ) return sliced.permute(__A ) def lowerCAmelCase_ ( __A, __A ) -> Any: '''simple docstring''' import torch UpperCAmelCase__ = torch.arange(1, __A ) UpperCAmelCase__ = torch.remainder(__A, __A ) UpperCAmelCase__ = remainders == 0 UpperCAmelCase__ = candidates[divisor_indices] UpperCAmelCase__ = torch.max(__A ) return largest_divisor, torch.div(__A, __A, rounding_mode="floor" ) class A ( UpperCAmelCase_ ): @property def lowercase_ (self : Any ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" UpperCAmelCase__ = OrderedDict({"input_ids": {0: "batch", 1: "sequence"}} ) if self.use_past: self.fill_with_past_key_values_(__UpperCAmelCase , direction="inputs" ) UpperCAmelCase__ = {0: "batch", 1: "past_sequence + sequence"} else: UpperCAmelCase__ = {0: "batch", 1: "sequence"} return common_inputs @property def lowercase_ (self : List[Any] ) -> int: """simple docstring""" return self._config.num_heads def lowercase_ (self : str , __UpperCAmelCase : PreTrainedTokenizer , __UpperCAmelCase : int = -1 , __UpperCAmelCase : int = -1 , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[TensorType] = None , ) -> Mapping[str, Any]: """simple docstring""" UpperCAmelCase__ = super(__UpperCAmelCase , self ).generate_dummy_inputs( __UpperCAmelCase , batch_size=__UpperCAmelCase , seq_length=__UpperCAmelCase , is_pair=__UpperCAmelCase , framework=__UpperCAmelCase ) # We need to order the input in the way they appears in the forward() UpperCAmelCase__ = OrderedDict({"input_ids": common_inputs["input_ids"]} ) # Need to add the past_keys if self.use_past: if not is_torch_available(): raise ValueError("Cannot generate dummy past_keys inputs without PyTorch installed." ) else: import torch UpperCAmelCase__ , UpperCAmelCase__ = common_inputs["input_ids"].shape # Not using the same length for past_key_values UpperCAmelCase__ = seqlen + 2 UpperCAmelCase__ = ( batch, self.num_attention_heads, past_key_values_length, self._config.hidden_size // self.num_attention_heads, ) UpperCAmelCase__ = [ (torch.zeros(__UpperCAmelCase ), torch.zeros(__UpperCAmelCase )) for _ in range(self.num_layers ) ] UpperCAmelCase__ = common_inputs["attention_mask"] if self.use_past: UpperCAmelCase__ = ordered_inputs["attention_mask"].dtype UpperCAmelCase__ = torch.cat( [ordered_inputs["attention_mask"], torch.ones(__UpperCAmelCase , __UpperCAmelCase , dtype=__UpperCAmelCase )] , dim=1 ) return ordered_inputs @property def lowercase_ (self : Optional[Any] ) -> int: """simple docstring""" return 1_3
65
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
1
from math import factorial def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' if n < k or k < 0: raise ValueError("Please enter positive integers for n and k where n >= k" ) return factorial(__A ) // (factorial(__A ) * factorial(n - k )) if __name__ == "__main__": print( 'The number of five-card hands possible from a standard', f'''fifty-two card deck is: {combinations(5_2, 5)}\n''', ) print( 'If a class of 40 students must be arranged into groups of', f'''4 for group projects, there are {combinations(4_0, 4)} ways''', 'to arrange them.\n', ) print( 'If 10 teams are competing in a Formula One race, there', f'''are {combinations(1_0, 3)} ways that first, second and''', 'third place can be awarded.', )
65
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
1
import warnings from ...utils import logging from .image_processing_glpn import GLPNImageProcessor UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): def __init__(self : Any , *__UpperCAmelCase : Any , **__UpperCAmelCase : Tuple ) -> None: """simple docstring""" warnings.warn( "The class GLPNFeatureExtractor is deprecated and will be removed in version 5 of Transformers. Please" " use GLPNImageProcessor instead." , __UpperCAmelCase , ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase )
65
from __future__ import annotations from collections import deque class A : def __init__(self : Dict , __UpperCAmelCase : list[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__UpperCAmelCase ) self.set_fail_transitions() def lowercase_ (self : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : str ) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = 0 for character in keyword: UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , __UpperCAmelCase ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase__ = len(self.adlist ) - 1 else: UpperCAmelCase__ = next_state self.adlist[current_state]["output"].append(__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = deque() for node in self.adlist[0]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = 0 while q: UpperCAmelCase__ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = self.adlist[r]["fail_state"] while ( self.find_next_state(__UpperCAmelCase , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase__ = self.adlist[state]["fail_state"] UpperCAmelCase__ = self.find_next_state( __UpperCAmelCase , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase__ = 0 UpperCAmelCase__ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> dict[str, list[int]]: """simple docstring""" UpperCAmelCase__ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase__ = 0 for i in range(len(__UpperCAmelCase ) ): while ( self.find_next_state(__UpperCAmelCase , string[i] ) is None and current_state != 0 ): UpperCAmelCase__ = self.adlist[current_state]["fail_state"] UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , string[i] ) if next_state is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase__ = [] result[key].append(i - len(__UpperCAmelCase ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
65
1
from copy import deepcopy from typing import Optional, Union import numpy as np from ...processing_utils import ProcessorMixin from ...tokenization_utils_base import BatchEncoding from ...utils import TensorType, is_tf_available, is_torch_available if is_torch_available(): import torch if is_tf_available(): import tensorflow as tf class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[int] = ['image_processor'] __UpperCAmelCase : Any = 'SamImageProcessor' def __init__(self : Optional[Any] , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = self.image_processor UpperCAmelCase__ = -1_0 UpperCAmelCase__ = self.image_processor.size["longest_edge"] def __call__(self : Optional[Any] , __UpperCAmelCase : Any=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : List[Any] , ) -> BatchEncoding: """simple docstring""" UpperCAmelCase__ = self.image_processor( __UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase , ) # pop arguments that are not used in the foward but used nevertheless UpperCAmelCase__ = encoding_image_processor["original_sizes"] if hasattr(__UpperCAmelCase , "numpy" ): # Checks if Torch or TF tensor UpperCAmelCase__ = original_sizes.numpy() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = self._check_and_preprocess_points( input_points=__UpperCAmelCase , input_labels=__UpperCAmelCase , input_boxes=__UpperCAmelCase , ) UpperCAmelCase__ = self._normalize_and_convert( __UpperCAmelCase , __UpperCAmelCase , input_points=__UpperCAmelCase , input_labels=__UpperCAmelCase , input_boxes=__UpperCAmelCase , return_tensors=__UpperCAmelCase , ) return encoding_image_processor def lowercase_ (self : Dict , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Any=None , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : Optional[Any]="pt" , ) -> int: """simple docstring""" if input_points is not None: if len(__UpperCAmelCase ) != len(__UpperCAmelCase ): UpperCAmelCase__ = [ self._normalize_coordinates(self.target_size , __UpperCAmelCase , original_sizes[0] ) for point in input_points ] else: UpperCAmelCase__ = [ self._normalize_coordinates(self.target_size , __UpperCAmelCase , __UpperCAmelCase ) for point, original_size in zip(__UpperCAmelCase , __UpperCAmelCase ) ] # check that all arrays have the same shape if not all(point.shape == input_points[0].shape for point in input_points ): if input_labels is not None: UpperCAmelCase__ , UpperCAmelCase__ = self._pad_points_and_labels(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = np.array(__UpperCAmelCase ) if input_labels is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase ) if input_boxes is not None: if len(__UpperCAmelCase ) != len(__UpperCAmelCase ): UpperCAmelCase__ = [ self._normalize_coordinates(self.target_size , __UpperCAmelCase , original_sizes[0] , is_bounding_box=__UpperCAmelCase ) for box in input_boxes ] else: UpperCAmelCase__ = [ self._normalize_coordinates(self.target_size , __UpperCAmelCase , __UpperCAmelCase , is_bounding_box=__UpperCAmelCase ) for box, original_size in zip(__UpperCAmelCase , __UpperCAmelCase ) ] UpperCAmelCase__ = np.array(__UpperCAmelCase ) if input_boxes is not None: if return_tensors == "pt": UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ) # boxes batch size of 1 by default UpperCAmelCase__ = input_boxes.unsqueeze(1 ) if len(input_boxes.shape ) != 3 else input_boxes elif return_tensors == "tf": UpperCAmelCase__ = tf.convert_to_tensor(__UpperCAmelCase ) # boxes batch size of 1 by default UpperCAmelCase__ = tf.expand_dims(__UpperCAmelCase , 1 ) if len(input_boxes.shape ) != 3 else input_boxes encoding_image_processor.update({"input_boxes": input_boxes} ) if input_points is not None: if return_tensors == "pt": UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ) # point batch size of 1 by default UpperCAmelCase__ = input_points.unsqueeze(1 ) if len(input_points.shape ) != 4 else input_points elif return_tensors == "tf": UpperCAmelCase__ = tf.convert_to_tensor(__UpperCAmelCase ) # point batch size of 1 by default UpperCAmelCase__ = tf.expand_dims(__UpperCAmelCase , 1 ) if len(input_points.shape ) != 4 else input_points encoding_image_processor.update({"input_points": input_points} ) if input_labels is not None: if return_tensors == "pt": UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ) # point batch size of 1 by default UpperCAmelCase__ = input_labels.unsqueeze(1 ) if len(input_labels.shape ) != 3 else input_labels elif return_tensors == "tf": UpperCAmelCase__ = tf.convert_to_tensor(__UpperCAmelCase ) # point batch size of 1 by default UpperCAmelCase__ = tf.expand_dims(__UpperCAmelCase , 1 ) if len(input_labels.shape ) != 3 else input_labels encoding_image_processor.update({"input_labels": input_labels} ) return encoding_image_processor def lowercase_ (self : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Tuple ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = max([point.shape[0] for point in input_points] ) UpperCAmelCase__ = [] for i, point in enumerate(__UpperCAmelCase ): if point.shape[0] != expected_nb_points: UpperCAmelCase__ = np.concatenate( [point, np.zeros((expected_nb_points - point.shape[0], 2) ) + self.point_pad_value] , axis=0 ) UpperCAmelCase__ = np.append(input_labels[i] , [self.point_pad_value] ) processed_input_points.append(__UpperCAmelCase ) UpperCAmelCase__ = processed_input_points return input_points, input_labels def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : int , __UpperCAmelCase : np.ndarray , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any]=False ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = original_size UpperCAmelCase__ , UpperCAmelCase__ = self.image_processor._get_preprocess_shape(__UpperCAmelCase , longest_edge=__UpperCAmelCase ) UpperCAmelCase__ = deepcopy(__UpperCAmelCase ).astype(__UpperCAmelCase ) if is_bounding_box: UpperCAmelCase__ = coords.reshape(-1 , 2 , 2 ) UpperCAmelCase__ = coords[..., 0] * (new_w / old_w) UpperCAmelCase__ = coords[..., 1] * (new_h / old_h) if is_bounding_box: UpperCAmelCase__ = coords.reshape(-1 , 4 ) return coords def lowercase_ (self : Any , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : List[Any]=None , ) -> List[str]: """simple docstring""" if input_points is not None: if hasattr(__UpperCAmelCase , "numpy" ): # Checks for TF or Torch tensor UpperCAmelCase__ = input_points.numpy().tolist() if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not isinstance(input_points[0] , __UpperCAmelCase ): raise ValueError("Input points must be a list of list of floating points." ) UpperCAmelCase__ = [np.array(__UpperCAmelCase ) for input_point in input_points] else: UpperCAmelCase__ = None if input_labels is not None: if hasattr(__UpperCAmelCase , "numpy" ): UpperCAmelCase__ = input_labels.numpy().tolist() if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not isinstance(input_labels[0] , __UpperCAmelCase ): raise ValueError("Input labels must be a list of list integers." ) UpperCAmelCase__ = [np.array(__UpperCAmelCase ) for label in input_labels] else: UpperCAmelCase__ = None if input_boxes is not None: if hasattr(__UpperCAmelCase , "numpy" ): UpperCAmelCase__ = input_boxes.numpy().tolist() if ( not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not isinstance(input_boxes[0] , __UpperCAmelCase ) or not isinstance(input_boxes[0][0] , __UpperCAmelCase ) ): raise ValueError("Input boxes must be a list of list of list of floating points." ) UpperCAmelCase__ = [np.array(__UpperCAmelCase ).astype(np.floataa ) for box in input_boxes] else: UpperCAmelCase__ = None return input_points, input_labels, input_boxes @property def lowercase_ (self : str ) -> Any: """simple docstring""" UpperCAmelCase__ = self.image_processor.model_input_names return list(dict.fromkeys(__UpperCAmelCase ) ) def lowercase_ (self : Optional[Any] , *__UpperCAmelCase : List[Any] , **__UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return self.image_processor.post_process_masks(*__UpperCAmelCase , **__UpperCAmelCase )
65
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ['input_values', 'attention_mask'] def __init__(self : Any , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 1_6_0_0_0 , __UpperCAmelCase : float = 0.0 , __UpperCAmelCase : bool = False , __UpperCAmelCase : int = 8_0 , __UpperCAmelCase : int = 1_6 , __UpperCAmelCase : int = 6_4 , __UpperCAmelCase : str = "hann_window" , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : float = 8_0 , __UpperCAmelCase : float = 7_6_0_0 , __UpperCAmelCase : float = 1E-10 , __UpperCAmelCase : int = 2 , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Any , ) -> str: """simple docstring""" super().__init__(feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = do_normalize UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = frame_signal_scale UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = reduction_factor UpperCAmelCase__ = win_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = hop_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = optimal_fft_length(self.sample_size ) UpperCAmelCase__ = (self.n_fft // 2) + 1 UpperCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCAmelCase ) UpperCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowercase_ (__UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase , np.intaa ) UpperCAmelCase__ = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): UpperCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: UpperCAmelCase__ = padding_value normed_input_values.append(__UpperCAmelCase ) else: UpperCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowercase_ (self : Optional[int] , __UpperCAmelCase : np.ndarray , ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ = spectrogram( __UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__(self : Any , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[int] = None , **__UpperCAmelCase : str , ) -> BatchFeature: """simple docstring""" if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) else: UpperCAmelCase__ = None if audio_target is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) if inputs is None: return inputs_target else: UpperCAmelCase__ = inputs_target["input_values"] UpperCAmelCase__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: UpperCAmelCase__ = decoder_attention_mask return inputs def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCAmelCase : bool = False , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : Any , ) -> BatchFeature: """simple docstring""" UpperCAmelCase__ = isinstance(__UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): UpperCAmelCase__ = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs UpperCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: UpperCAmelCase__ = [self._extract_mel_features(__UpperCAmelCase ) for waveform in speech] UpperCAmelCase__ = BatchFeature({"input_values": features} ) UpperCAmelCase__ = self.num_mel_bins else: UpperCAmelCase__ = BatchFeature({"input_values": speech} ) UpperCAmelCase__ = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = feature_size_hack # convert input values to correct format UpperCAmelCase__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(__UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): UpperCAmelCase__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(__UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = input_values.astype(np.floataa ) # convert attention_mask to correct format UpperCAmelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: UpperCAmelCase__ = ( attention_mask if self._get_padding_strategies(__UpperCAmelCase , max_length=__UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=__UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def lowercase_ (self : Tuple ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. UpperCAmelCase__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
65
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'spm_char.model'} UpperCamelCase__ = { 'vocab_file': { 'microsoft/speecht5_asr': 'https://huggingface.co/microsoft/speecht5_asr/resolve/main/spm_char.model', 'microsoft/speecht5_tts': 'https://huggingface.co/microsoft/speecht5_tts/resolve/main/spm_char.model', 'microsoft/speecht5_vc': 'https://huggingface.co/microsoft/speecht5_vc/resolve/main/spm_char.model', } } UpperCamelCase__ = { 'microsoft/speecht5_asr': 1_0_2_4, 'microsoft/speecht5_tts': 1_0_2_4, 'microsoft/speecht5_vc': 1_0_2_4, } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Any = VOCAB_FILES_NAMES __UpperCAmelCase : int = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Dict = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : str = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str]="<s>" , __UpperCAmelCase : Union[str, Any]="</s>" , __UpperCAmelCase : List[str]="<unk>" , __UpperCAmelCase : Any="<pad>" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Optional[Any] , ) -> None: """simple docstring""" UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) UpperCAmelCase__ = vocab_file UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) @property def lowercase_ (self : Optional[Any] ) -> Optional[int]: """simple docstring""" return self.sp_model.get_piece_size() def lowercase_ (self : List[Any] ) -> Any: """simple docstring""" UpperCAmelCase__ = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__(self : Tuple ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None return state def __setstate__(self : Union[str, Any] , __UpperCAmelCase : str ) -> str: """simple docstring""" UpperCAmelCase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" return self.sp_model.piece_to_id(__UpperCAmelCase ) def lowercase_ (self : str , __UpperCAmelCase : Dict ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.sp_model.IdToPiece(__UpperCAmelCase ) return token def lowercase_ (self : Any , __UpperCAmelCase : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = [] UpperCAmelCase__ = "" for token in tokens: # make sure that special tokens are not decoded using sentencepiece model if token in self.all_special_tokens: out_string += self.sp_model.decode(__UpperCAmelCase ) + token UpperCAmelCase__ = [] else: current_sub_tokens.append(__UpperCAmelCase ) out_string += self.sp_model.decode(__UpperCAmelCase ) return out_string.strip() def lowercase_ (self : Any , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any]=None ) -> List[int]: """simple docstring""" if token_ids_a is None: return token_ids_a + [self.eos_token_id] # We don't expect to process pairs, but leave the pair logic for API consistency return token_ids_a + token_ids_a + [self.eos_token_id] def lowercase_ (self : int , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) UpperCAmelCase__ = [1] if token_ids_a is None: return ([0] * len(__UpperCAmelCase )) + suffix_ones return ([0] * len(__UpperCAmelCase )) + ([0] * len(__UpperCAmelCase )) + suffix_ones def lowercase_ (self : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , "wb" ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,)
65
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[torch.FloatTensor] = None __UpperCAmelCase : torch.FloatTensor = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Union[str, Any]=5_1_2 , __UpperCAmelCase : List[str]="cls" , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=True , **__UpperCAmelCase : str , ) -> int: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = project_dim UpperCAmelCase__ = pooler_fn UpperCAmelCase__ = learn_encoder UpperCAmelCase__ = use_attention_mask class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = [r'pooler', r'logit_scale'] __UpperCAmelCase : int = [r'position_ids', r'predictions.decoder.bias'] __UpperCAmelCase : Any = 'roberta' __UpperCAmelCase : List[str] = RobertaSeriesConfig def __init__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = XLMRobertaModel(__UpperCAmelCase ) UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = getattr(__UpperCAmelCase , "has_pre_transformation" , __UpperCAmelCase ) if self.has_pre_transformation: UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.base_model( input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , position_ids=__UpperCAmelCase , head_mask=__UpperCAmelCase , inputs_embeds=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__UpperCAmelCase , ) if self.has_pre_transformation: UpperCAmelCase__ = outputs["hidden_states"][-2] UpperCAmelCase__ = self.pre_LN(__UpperCAmelCase ) UpperCAmelCase__ = self.transformation_pre(__UpperCAmelCase ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: UpperCAmelCase__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
65
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_torch_available UpperCamelCase__ = { 'configuration_jukebox': [ 'JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP', 'JukeboxConfig', 'JukeboxPriorConfig', 'JukeboxVQVAEConfig', ], 'tokenization_jukebox': ['JukeboxTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST', 'JukeboxModel', 'JukeboxPreTrainedModel', 'JukeboxVQVAE', 'JukeboxPrior', ] if TYPE_CHECKING: from .configuration_jukebox import ( JUKEBOX_PRETRAINED_CONFIG_ARCHIVE_MAP, JukeboxConfig, JukeboxPriorConfig, JukeboxVQVAEConfig, ) from .tokenization_jukebox import JukeboxTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_jukebox import ( JUKEBOX_PRETRAINED_MODEL_ARCHIVE_LIST, JukeboxModel, JukeboxPreTrainedModel, JukeboxPrior, JukeboxVQVAE, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
1
# DISCLAIMER: This file is strongly influenced by https://github.com/ermongroup/ddim from dataclasses import dataclass from typing import Optional, Tuple, Union import flax import jax import jax.numpy as jnp from ..configuration_utils import ConfigMixin, register_to_config from .scheduling_utils_flax import ( CommonSchedulerState, FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, add_noise_common, get_velocity_common, ) @flax.struct.dataclass class A : __UpperCAmelCase : CommonSchedulerState # setable values __UpperCAmelCase : jnp.ndarray __UpperCAmelCase : jnp.ndarray __UpperCAmelCase : Optional[int] = None @classmethod def lowercase_ (cls : Any , __UpperCAmelCase : CommonSchedulerState , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray ) -> List[str]: """simple docstring""" return cls(common=__UpperCAmelCase , init_noise_sigma=__UpperCAmelCase , timesteps=__UpperCAmelCase ) @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : DDPMSchedulerState class A ( UpperCAmelCase_ , UpperCAmelCase_ ): __UpperCAmelCase : Any = [e.name for e in FlaxKarrasDiffusionSchedulers] __UpperCAmelCase : jnp.dtype @property def lowercase_ (self : Any ) -> Union[str, Any]: """simple docstring""" return True @register_to_config def __init__(self : Optional[Any] , __UpperCAmelCase : int = 1_0_0_0 , __UpperCAmelCase : float = 0.0001 , __UpperCAmelCase : float = 0.02 , __UpperCAmelCase : str = "linear" , __UpperCAmelCase : Optional[jnp.ndarray] = None , __UpperCAmelCase : str = "fixed_small" , __UpperCAmelCase : bool = True , __UpperCAmelCase : str = "epsilon" , __UpperCAmelCase : jnp.dtype = jnp.floataa , ) -> List[str]: """simple docstring""" UpperCAmelCase__ = dtype def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[CommonSchedulerState] = None ) -> DDPMSchedulerState: """simple docstring""" if common is None: UpperCAmelCase__ = CommonSchedulerState.create(self ) # standard deviation of the initial noise distribution UpperCAmelCase__ = jnp.array(1.0 , dtype=self.dtype ) UpperCAmelCase__ = jnp.arange(0 , self.config.num_train_timesteps ).round()[::-1] return DDPMSchedulerState.create( common=__UpperCAmelCase , init_noise_sigma=__UpperCAmelCase , timesteps=__UpperCAmelCase , ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : Optional[int] = None ) -> jnp.ndarray: """simple docstring""" return sample def lowercase_ (self : Any , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : int , __UpperCAmelCase : Tuple = () ) -> DDPMSchedulerState: """simple docstring""" UpperCAmelCase__ = self.config.num_train_timesteps // num_inference_steps # creates integer timesteps by multiplying by ratio # rounding to avoid issues when num_inference_step is power of 3 UpperCAmelCase__ = (jnp.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1] return state.replace( num_inference_steps=__UpperCAmelCase , timesteps=__UpperCAmelCase , ) def lowercase_ (self : str , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : Optional[int]=None ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = state.common.alphas_cumprod[t] UpperCAmelCase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase__ = (1 - alpha_prod_t_prev) / (1 - alpha_prod_t) * state.common.betas[t] if variance_type is None: UpperCAmelCase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small": UpperCAmelCase__ = jnp.clip(__UpperCAmelCase , a_min=1E-20 ) # for rl-diffuser https://arxiv.org/abs/2205.09991 elif variance_type == "fixed_small_log": UpperCAmelCase__ = jnp.log(jnp.clip(__UpperCAmelCase , a_min=1E-20 ) ) elif variance_type == "fixed_large": UpperCAmelCase__ = state.common.betas[t] elif variance_type == "fixed_large_log": # Glide max_log UpperCAmelCase__ = jnp.log(state.common.betas[t] ) elif variance_type == "learned": return predicted_variance elif variance_type == "learned_range": UpperCAmelCase__ = variance UpperCAmelCase__ = state.common.betas[t] UpperCAmelCase__ = (predicted_variance + 1) / 2 UpperCAmelCase__ = frac * max_log + (1 - frac) * min_log return variance def lowercase_ (self : int , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : Optional[jax.random.KeyArray] = None , __UpperCAmelCase : bool = True , ) -> Union[FlaxDDPMSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = timestep if key is None: UpperCAmelCase__ = jax.random.PRNGKey(0 ) if model_output.shape[1] == sample.shape[1] * 2 and self.config.variance_type in ["learned", "learned_range"]: UpperCAmelCase__ , UpperCAmelCase__ = jnp.split(__UpperCAmelCase , sample.shape[1] , axis=1 ) else: UpperCAmelCase__ = None # 1. compute alphas, betas UpperCAmelCase__ = state.common.alphas_cumprod[t] UpperCAmelCase__ = jnp.where(t > 0 , state.common.alphas_cumprod[t - 1] , jnp.array(1.0 , dtype=self.dtype ) ) UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase__ = model_output elif self.config.prediction_type == "v_prediction": UpperCAmelCase__ = (alpha_prod_t**0.5) * sample - (beta_prod_t**0.5) * model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon`, `sample` """ " for the FlaxDDPMScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase__ = jnp.clip(__UpperCAmelCase , -1 , 1 ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = (alpha_prod_t_prev ** 0.5 * state.common.betas[t]) / beta_prod_t UpperCAmelCase__ = state.common.alphas[t] ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise def random_variance(): UpperCAmelCase__ = jax.random.split(__UpperCAmelCase , num=1 ) UpperCAmelCase__ = jax.random.normal(__UpperCAmelCase , shape=model_output.shape , dtype=self.dtype ) return (self._get_variance(__UpperCAmelCase , __UpperCAmelCase , predicted_variance=__UpperCAmelCase ) ** 0.5) * noise UpperCAmelCase__ = jnp.where(t > 0 , random_variance() , jnp.zeros(model_output.shape , dtype=self.dtype ) ) UpperCAmelCase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample, state) return FlaxDDPMSchedulerOutput(prev_sample=__UpperCAmelCase , state=__UpperCAmelCase ) def lowercase_ (self : int , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , ) -> jnp.ndarray: """simple docstring""" return add_noise_common(state.common , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : DDPMSchedulerState , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , ) -> jnp.ndarray: """simple docstring""" return get_velocity_common(state.common , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def __len__(self : int ) -> Union[str, Any]: """simple docstring""" return self.config.num_train_timesteps
65
import math import random def lowerCAmelCase_ ( __A, __A = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase__ = 0.0_2 def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' UpperCAmelCase__ = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__A ): # Forward propagation UpperCAmelCase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase__ = (expected / 100) - layer_a # Error delta UpperCAmelCase__ = layer_1_error * sigmoid_function(__A, __A ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input('Expected value: ')) UpperCamelCase__ = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
65
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_sentencepiece_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { 'configuration_rembert': ['REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'RemBertConfig', 'RemBertOnnxConfig'] } try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['RemBertTokenizer'] try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['RemBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'RemBertForCausalLM', 'RemBertForMaskedLM', 'RemBertForMultipleChoice', 'RemBertForQuestionAnswering', 'RemBertForSequenceClassification', 'RemBertForTokenClassification', 'RemBertLayer', 'RemBertModel', 'RemBertPreTrainedModel', 'load_tf_weights_in_rembert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFRemBertForCausalLM', 'TFRemBertForMaskedLM', 'TFRemBertForMultipleChoice', 'TFRemBertForQuestionAnswering', 'TFRemBertForSequenceClassification', 'TFRemBertForTokenClassification', 'TFRemBertLayer', 'TFRemBertModel', 'TFRemBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_rembert import REMBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, RemBertConfig, RemBertOnnxConfig try: if not is_sentencepiece_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert import RemBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_rembert_fast import RemBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_rembert import ( REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, RemBertForCausalLM, RemBertForMaskedLM, RemBertForMultipleChoice, RemBertForQuestionAnswering, RemBertForSequenceClassification, RemBertForTokenClassification, RemBertLayer, RemBertModel, RemBertPreTrainedModel, load_tf_weights_in_rembert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_rembert import ( TF_REMBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFRemBertForCausalLM, TFRemBertForMaskedLM, TFRemBertForMultipleChoice, TFRemBertForQuestionAnswering, TFRemBertForSequenceClassification, TFRemBertForTokenClassification, TFRemBertLayer, TFRemBertModel, TFRemBertPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
from __future__ import annotations class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : list[list[int]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed from a list of zero or more lists containing at " "least one and the same number of values, each of which must be of type " "int or float." ) if len(__UpperCAmelCase ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__UpperCAmelCase ) != cols: raise error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def lowercase_ (self : Any ) -> list[list[int]]: """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def lowercase_ (self : Any ) -> int: """simple docstring""" return len(self.rows ) @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" return len(self.rows[0] ) @property def lowercase_ (self : List[Any] ) -> tuple[int, int]: """simple docstring""" return (self.num_rows, self.num_columns) @property def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return self.order[0] == self.order[1] def lowercase_ (self : Any ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : int ) -> int: """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return bool(self.determinant() ) def lowercase_ (self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__UpperCAmelCase ).determinant() def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) return -1 * self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Matrix: """simple docstring""" return Matrix( [ [self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def lowercase_ (self : List[str] ) -> Matrix: """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def lowercase_ (self : Optional[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError("Only matrices with a non-zero determinant have an inverse" ) return self.adjugate() * (1 / determinant) def __repr__(self : Dict ) -> str: """simple docstring""" return str(self.rows ) def __str__(self : Optional[Any] ) -> str: """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ "[" + ". ".join([str(__UpperCAmelCase ) for value in row] ) + ".]" for row in self.rows ] ) + "]" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError("Row must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_columns: raise ValueError( "Row must be equal in length to the other rows in the matrix" ) if position is None: self.rows.append(__UpperCAmelCase ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError( "Column must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in column: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_rows: raise ValueError( "Column must be equal in length to the other columns in the matrix" ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__(self : Any , __UpperCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return NotImplemented return self.rows == other.rows def __ne__(self : int , __UpperCAmelCase : object ) -> bool: """simple docstring""" return not self == other def __neg__(self : Dict ) -> Matrix: """simple docstring""" return self * -1 def __add__(self : Dict , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Addition requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__(self : Optional[Any] , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Subtraction requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__(self : Tuple , __UpperCAmelCase : Matrix | int | float ) -> Matrix: """simple docstring""" if isinstance(__UpperCAmelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): if self.num_columns != other.num_rows: raise ValueError( "The number of columns in the first matrix must " "be equal to the number of rows in the second" ) return Matrix( [ [Matrix.dot_product(__UpperCAmelCase , __UpperCAmelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( "A Matrix can only be multiplied by an int, float, or another matrix" ) def __pow__(self : List[Any] , __UpperCAmelCase : int ) -> Matrix: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("A Matrix can only be raised to the power of an int" ) if not self.is_square: raise ValueError("Only square matrices can be raised to a power" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( "Only invertable matrices can be raised to a negative power" ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : list[int] , __UpperCAmelCase : list[int] ) -> int: """simple docstring""" return sum(row[i] * column[i] for i in range(len(__UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
1
import io import os import unicodedata from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = '▁' UpperCamelCase__ = {'vocab_file': 'vocab.txt', 'sentencepiece_model_ckpt': 'sentencepiece.bpe.model'} UpperCamelCase__ = { 'sentencepiece_model_file': 'sentencepiece.bpe.model', 'vocab_file': 'vocab.txt', } UpperCamelCase__ = { 'vocab_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/vocab.txt', }, 'sentencepiece_model_file': { 'ernie-m-base': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', 'ernie-m-large': 'https://huggingface.co/susnato/ernie-m-base_pytorch/blob/main/sentencepiece.bpe.model', }, } UpperCamelCase__ = { 'ernie-m-base': 5_1_4, 'ernie-m-large': 5_1_4, } UpperCamelCase__ = { 'ernie-m-base': {'do_lower_case': False}, 'ernie-m-large': {'do_lower_case': False}, } class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[str] = ["input_ids"] __UpperCAmelCase : int = VOCAB_FILES_NAMES __UpperCAmelCase : Tuple = PRETRAINED_INIT_CONFIGURATION __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Optional[Any] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : str = RESOURCE_FILES_NAMES def __init__(self : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : Optional[Any]="utf8" , __UpperCAmelCase : Union[str, Any]="[UNK]" , __UpperCAmelCase : Any="[SEP]" , __UpperCAmelCase : Union[str, Any]="[PAD]" , __UpperCAmelCase : Optional[Any]="[CLS]" , __UpperCAmelCase : Tuple="[MASK]" , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Optional[Any] , ) -> None: """simple docstring""" UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , vocab_file=__UpperCAmelCase , encoding=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case UpperCAmelCase__ = sentencepiece_model_ckpt UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) # to mimic paddlenlp.transformers.ernie_m.tokenizer.ErnieMTokenizer functioning if vocab_file is not None: UpperCAmelCase__ = self.load_vocab(filepath=__UpperCAmelCase ) else: UpperCAmelCase__ = {self.sp_model.id_to_piece(__UpperCAmelCase ): id for id in range(self.sp_model.get_piece_size() )} UpperCAmelCase__ = {v: k for k, v in self.vocab.items()} def lowercase_ (self : Optional[Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" if text is None: return None UpperCAmelCase__ = self.tokenize(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = "", [] for i, ch in enumerate(__UpperCAmelCase ): if ch in self.SP_CHAR_MAPPING: UpperCAmelCase__ = self.SP_CHAR_MAPPING.get(__UpperCAmelCase ) else: UpperCAmelCase__ = unicodedata.normalize("NFKC" , __UpperCAmelCase ) if self.is_whitespace(__UpperCAmelCase ): continue normalized_text += ch char_mapping.extend([i] * len(__UpperCAmelCase ) ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = normalized_text, [], 0 if self.do_lower_case: UpperCAmelCase__ = text.lower() for token in split_tokens: if token[:1] == "▁": UpperCAmelCase__ = token[1:] UpperCAmelCase__ = text[offset:].index(__UpperCAmelCase ) + offset UpperCAmelCase__ = start + len(__UpperCAmelCase ) token_mapping.append((char_mapping[start], char_mapping[end - 1] + 1) ) UpperCAmelCase__ = end return token_mapping @property def lowercase_ (self : str ) -> List[str]: """simple docstring""" return len(self.vocab ) def lowercase_ (self : Dict ) -> Tuple: """simple docstring""" return dict(self.vocab , **self.added_tokens_encoder ) def __getstate__(self : List[str] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None return state def __setstate__(self : int , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.sentencepiece_model_ckpt ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> Any: """simple docstring""" return "".join((self.SP_CHAR_MAPPING.get(__UpperCAmelCase , __UpperCAmelCase ) for c in text) ) def lowercase_ (self : str , __UpperCAmelCase : Any , __UpperCAmelCase : Dict=False , __UpperCAmelCase : str=6_4 , __UpperCAmelCase : str=0.1 ) -> str: """simple docstring""" if self.sp_model_kwargs.get("enable_sampling" ) is True: UpperCAmelCase__ = True if self.sp_model_kwargs.get("alpha" ) is not None: UpperCAmelCase__ = self.sp_model_kwargs.get("alpha" ) if self.sp_model_kwargs.get("nbest_size" ) is not None: UpperCAmelCase__ = self.sp_model_kwargs.get("nbest_size" ) if not enable_sampling: UpperCAmelCase__ = self.sp_model.EncodeAsPieces(__UpperCAmelCase ) else: UpperCAmelCase__ = self.sp_model.SampleEncodeAsPieces(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = [] for pi, piece in enumerate(__UpperCAmelCase ): if piece == SPIECE_UNDERLINE: if not pieces[pi + 1].startswith(__UpperCAmelCase ) and pi != 0: new_pieces.append(__UpperCAmelCase ) continue else: continue UpperCAmelCase__ = 0 for i, chunk in enumerate(__UpperCAmelCase ): if chunk == SPIECE_UNDERLINE: continue if self.is_ch_char(__UpperCAmelCase ) or self.is_punct(__UpperCAmelCase ): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) new_pieces.append(__UpperCAmelCase ) UpperCAmelCase__ = i + 1 elif chunk.isdigit() and i > 0 and not piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) UpperCAmelCase__ = i elif not chunk.isdigit() and i > 0 and piece[i - 1].isdigit(): if i > lst_i and piece[lst_i:i] != SPIECE_UNDERLINE: new_pieces.append(piece[lst_i:i] ) UpperCAmelCase__ = i if len(__UpperCAmelCase ) > lst_i: new_pieces.append(piece[lst_i:] ) return new_pieces def lowercase_ (self : Tuple , __UpperCAmelCase : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = "".join(__UpperCAmelCase ).replace(__UpperCAmelCase , " " ).strip() return out_string def lowercase_ (self : Any , __UpperCAmelCase : Any ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.convert_ids_to_tokens(__UpperCAmelCase ) UpperCAmelCase__ = "".join(__UpperCAmelCase ).replace(__UpperCAmelCase , " " ).strip() return out_string def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> Tuple: """simple docstring""" return self.vocab.get(__UpperCAmelCase , self.vocab.get(self.unk_token ) ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[int] ) -> Optional[Any]: """simple docstring""" return self.reverse_vocab.get(__UpperCAmelCase , self.unk_token ) def lowercase_ (self : int , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any]=None ) -> Dict: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] UpperCAmelCase__ = [self.sep_token_id] return _cls + token_ids_a + _sep + _sep + token_ids_a + _sep def lowercase_ (self : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : str=None ) -> Dict: """simple docstring""" if offset_mapping_a is None: return [(0, 0)] + offset_mapping_a + [(0, 0)] return [(0, 0)] + offset_mapping_a + [(0, 0), (0, 0)] + offset_mapping_a + [(0, 0)] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Optional[int]=False ) -> Tuple: """simple docstring""" if already_has_special_tokens: if token_ids_a is not None: raise ValueError( "You should not supply a second sequence if the provided sequence of " "ids is already formatted with special tokens for the model." ) return [1 if x in [self.sep_token_id, self.cls_token_id] else 0 for x in token_ids_a] if token_ids_a is not None: return [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] + ([0] * len(__UpperCAmelCase )) + [1] return [1] + ([0] * len(__UpperCAmelCase )) + [1] def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: # [CLS] X [SEP] return (len(__UpperCAmelCase ) + 2) * [0] # [CLS] A [SEP] [SEP] B [SEP] return [0] * (len(__UpperCAmelCase ) + 1) + [1] * (len(__UpperCAmelCase ) + 3) def lowercase_ (self : Any , __UpperCAmelCase : Dict ) -> Optional[Any]: """simple docstring""" if "\u4e00" <= char <= "\u9fff": return True return False def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> Dict: """simple docstring""" if ("a" <= char <= "z") or ("A" <= char <= "Z"): return True return False def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> Optional[int]: """simple docstring""" if char in ",;:.?!~,;:。?!《》【】": return True return False def lowercase_ (self : str , __UpperCAmelCase : Union[str, Any] ) -> Any: """simple docstring""" if char == " " or char == "\t" or char == "\n" or char == "\r": return True if len(__UpperCAmelCase ) == 1: UpperCAmelCase__ = unicodedata.category(__UpperCAmelCase ) if cat == "Zs": return True return False def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = {} with io.open(__UpperCAmelCase , "r" , encoding="utf-8" ) as f: for index, line in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = line.rstrip("\n" ) UpperCAmelCase__ = int(__UpperCAmelCase ) return token_to_idx def lowercase_ (self : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase__ = 0 if os.path.isdir(__UpperCAmelCase ): UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) else: UpperCAmelCase__ = (filename_prefix + "-" if filename_prefix else "") + save_directory with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for token, token_index in sorted(self.vocab.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {vocab_file}: vocabulary indices are not consecutive.""" " Please check that the vocabulary is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(token + "\n" ) index += 1 UpperCAmelCase__ = os.path.join(__UpperCAmelCase , "sentencepiece.bpe.model" ) with open(__UpperCAmelCase , "wb" ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (vocab_file,)
65
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
1
import argparse import os import torch from transformers import FlavaImageCodebook, FlavaImageCodebookConfig def lowerCAmelCase_ ( __A, __A, __A, __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = s.rsplit(__A, __A ) return new.join(__A ) def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' return sum(param.float().sum() if "encoder.embeddings" not in key else 0 for key, param in state_dict.items() ) def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = {} UpperCAmelCase__ = ["group_1", "group_2", "group_3", "group_4"] for key, value in state_dict.items(): for group_key in group_keys: if group_key in key: UpperCAmelCase__ = key.replace(f"""{group_key}.""", f"""{group_key}.group.""" ) if "res_path" in key: UpperCAmelCase__ = key.replace("res_path.", "res_path.path." ) if key.endswith(".w" ): UpperCAmelCase__ = rreplace(__A, ".w", ".weight", 1 ) if key.endswith(".b" ): UpperCAmelCase__ = rreplace(__A, ".b", ".bias", 1 ) UpperCAmelCase__ = value.float() return upgrade @torch.no_grad() def lowerCAmelCase_ ( __A, __A, __A=None, __A=True ) -> List[str]: '''simple docstring''' from dall_e import Encoder UpperCAmelCase__ = Encoder() if os.path.exists(__A ): UpperCAmelCase__ = torch.load(__A ) else: UpperCAmelCase__ = torch.hub.load_state_dict_from_url(__A ) if isinstance(__A, __A ): UpperCAmelCase__ = ckpt.state_dict() encoder.load_state_dict(__A ) if config_path is not None: UpperCAmelCase__ = FlavaImageCodebookConfig.from_pretrained(__A ) else: UpperCAmelCase__ = FlavaImageCodebookConfig() UpperCAmelCase__ = FlavaImageCodebook(__A ).eval() UpperCAmelCase__ = encoder.state_dict() UpperCAmelCase__ = upgrade_state_dict(__A ) hf_model.load_state_dict(__A ) UpperCAmelCase__ = hf_model.state_dict() UpperCAmelCase__ = count_parameters(__A ) UpperCAmelCase__ = count_parameters(__A ) assert torch.allclose(__A, __A, atol=1e-3 ) if save_checkpoint: hf_model.save_pretrained(__A ) else: return hf_state_dict if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument('--checkpoint_path', default=None, type=str, help='Path to flava checkpoint') parser.add_argument('--config_path', default=None, type=str, help='Path to hf config.json of model to convert') UpperCamelCase__ = parser.parse_args() convert_dalle_checkpoint(args.checkpoint_path, args.pytorch_dump_folder_path, args.config_path)
65
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor class A ( nn.Module ): def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , __UpperCAmelCase : int=(6_4,) , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Any=3_2 , __UpperCAmelCase : str="silu" , __UpperCAmelCase : Any=True , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = torch.nn.Convad( __UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) # down UpperCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_down_block( __UpperCAmelCase , num_layers=self.layers_per_block , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) self.down_blocks.append(__UpperCAmelCase ) # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # out UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = 2 * out_channels if double_z else out_channels UpperCAmelCase__ = nn.Convad(block_out_channels[-1] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : List[Any] , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = x UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : int ): def custom_forward(*__UpperCAmelCase : Optional[Any] ): return module(*__UpperCAmelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __UpperCAmelCase ) else: # down for down_block in self.down_blocks: UpperCAmelCase__ = down_block(__UpperCAmelCase ) # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase ) # post-process UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : List[Any] , __UpperCAmelCase : str=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=("UpDecoderBlock2D",) , __UpperCAmelCase : str=(6_4,) , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : Any="silu" , __UpperCAmelCase : Any="group" , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) UpperCAmelCase__ = in_channels if norm_type == "spatial" else None # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # up UpperCAmelCase__ = list(reversed(__UpperCAmelCase ) ) UpperCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = reversed_block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_up_block( __UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , prev_output_channel=__UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , resnet_time_scale_shift=__UpperCAmelCase , ) self.up_blocks.append(__UpperCAmelCase ) UpperCAmelCase__ = output_channel # out if norm_type == "spatial": UpperCAmelCase__ = SpatialNorm(block_out_channels[0] , __UpperCAmelCase ) else: UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = nn.Convad(block_out_channels[0] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = z UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) UpperCAmelCase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : str ): def custom_forward(*__UpperCAmelCase : List[str] ): return module(*__UpperCAmelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase ) else: # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = up_block(__UpperCAmelCase , __UpperCAmelCase ) # post-process if latent_embeds is None: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) else: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Union[str, Any]="random" , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=True ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = n_e UpperCAmelCase__ = vq_embed_dim UpperCAmelCase__ = beta UpperCAmelCase__ = legacy UpperCAmelCase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) UpperCAmelCase__ = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) UpperCAmelCase__ = self.used.shape[0] UpperCAmelCase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": UpperCAmelCase__ = self.re_embed UpperCAmelCase__ = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: UpperCAmelCase__ = n_e UpperCAmelCase__ = sane_index_shape def lowercase_ (self : str , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) UpperCAmelCase__ = (inds[:, :, None] == used[None, None, ...]).long() UpperCAmelCase__ = match.argmax(-1 ) UpperCAmelCase__ = match.sum(2 ) < 1 if self.unknown_index == "random": UpperCAmelCase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: UpperCAmelCase__ = self.unknown_index return new.reshape(__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token UpperCAmelCase__ = 0 # simply set to zero UpperCAmelCase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __UpperCAmelCase ) return back.reshape(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() UpperCAmelCase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z UpperCAmelCase__ = torch.argmin(torch.cdist(__UpperCAmelCase , self.embedding.weight ) , dim=1 ) UpperCAmelCase__ = self.embedding(__UpperCAmelCase ).view(z.shape ) UpperCAmelCase__ = None UpperCAmelCase__ = None # compute loss for embedding if not self.legacy: UpperCAmelCase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: UpperCAmelCase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients UpperCAmelCase__ = z + (z_q - z).detach() # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: UpperCAmelCase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.remap_to_used(__UpperCAmelCase ) UpperCAmelCase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: UpperCAmelCase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Any: """simple docstring""" if self.remap is not None: UpperCAmelCase__ = indices.reshape(shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.unmap_to_all(__UpperCAmelCase ) UpperCAmelCase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors UpperCAmelCase__ = self.embedding(__UpperCAmelCase ) if shape is not None: UpperCAmelCase__ = z_q.view(__UpperCAmelCase ) # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parameters UpperCAmelCase__ , UpperCAmelCase__ = torch.chunk(__UpperCAmelCase , 2 , dim=1 ) UpperCAmelCase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) UpperCAmelCase__ = deterministic UpperCAmelCase__ = torch.exp(0.5 * self.logvar ) UpperCAmelCase__ = torch.exp(self.logvar ) if self.deterministic: UpperCAmelCase__ = UpperCAmelCase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = randn_tensor( self.mean.shape , generator=__UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) UpperCAmelCase__ = self.mean + self.std * sample return x def lowercase_ (self : str , __UpperCAmelCase : int=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase_ (self : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=[1, 2, 3] ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) UpperCAmelCase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" return self.mean
65
1
import importlib.util import os import platform from argparse import ArgumentParser import huggingface_hub from .. import __version__ as version from ..utils import ( is_accelerate_available, is_flax_available, is_safetensors_available, is_tf_available, is_torch_available, ) from . import BaseTransformersCLICommand def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' return EnvironmentCommand() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' return EnvironmentCommand(args.accelerate_config_file ) class A ( UpperCAmelCase_ ): @staticmethod def lowercase_ (__UpperCAmelCase : ArgumentParser ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = parser.add_parser("env" ) download_parser.set_defaults(func=__UpperCAmelCase ) download_parser.add_argument( "--accelerate-config_file" , default=__UpperCAmelCase , help="The accelerate config file to use for the default values in the launching script." , ) download_parser.set_defaults(func=__UpperCAmelCase ) def __init__(self : Optional[int] , __UpperCAmelCase : str , *__UpperCAmelCase : Tuple ) -> None: """simple docstring""" UpperCAmelCase__ = accelerate_config_file def lowercase_ (self : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = "not installed" if is_safetensors_available(): import safetensors UpperCAmelCase__ = safetensors.__version__ elif importlib.util.find_spec("safetensors" ) is not None: import safetensors UpperCAmelCase__ = f"""{safetensors.__version__} but is ignored because of PyTorch version too old.""" UpperCAmelCase__ = "not installed" UpperCAmelCase__ = UpperCAmelCase__ = "not found" if is_accelerate_available(): import accelerate from accelerate.commands.config import default_config_file, load_config_from_file UpperCAmelCase__ = accelerate.__version__ # Get the default from the config file. if self._accelerate_config_file is not None or os.path.isfile(__UpperCAmelCase ): UpperCAmelCase__ = load_config_from_file(self._accelerate_config_file ).to_dict() UpperCAmelCase__ = ( "\n".join([f"""\t- {prop}: {val}""" for prop, val in accelerate_config.items()] ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else f"""\t{accelerate_config}""" ) UpperCAmelCase__ = "not installed" UpperCAmelCase__ = "NA" if is_torch_available(): import torch UpperCAmelCase__ = torch.__version__ UpperCAmelCase__ = torch.cuda.is_available() UpperCAmelCase__ = "not installed" UpperCAmelCase__ = "NA" if is_tf_available(): import tensorflow as tf UpperCAmelCase__ = tf.__version__ try: # deprecated in v2.1 UpperCAmelCase__ = tf.test.is_gpu_available() except AttributeError: # returns list of devices, convert to bool UpperCAmelCase__ = bool(tf.config.list_physical_devices("GPU" ) ) UpperCAmelCase__ = "not installed" UpperCAmelCase__ = "not installed" UpperCAmelCase__ = "not installed" UpperCAmelCase__ = "NA" if is_flax_available(): import flax import jax import jaxlib UpperCAmelCase__ = flax.__version__ UpperCAmelCase__ = jax.__version__ UpperCAmelCase__ = jaxlib.__version__ UpperCAmelCase__ = jax.lib.xla_bridge.get_backend().platform UpperCAmelCase__ = { "`transformers` version": version, "Platform": platform.platform(), "Python version": platform.python_version(), "Huggingface_hub version": huggingface_hub.__version__, "Safetensors version": f"""{safetensors_version}""", "Accelerate version": f"""{accelerate_version}""", "Accelerate config": f"""{accelerate_config_str}""", "PyTorch version (GPU?)": f"""{pt_version} ({pt_cuda_available})""", "Tensorflow version (GPU?)": f"""{tf_version} ({tf_cuda_available})""", "Flax version (CPU?/GPU?/TPU?)": f"""{flax_version} ({jax_backend})""", "Jax version": f"""{jax_version}""", "JaxLib version": f"""{jaxlib_version}""", "Using GPU in script?": "<fill in>", "Using distributed or parallel set-up in script?": "<fill in>", } print("\nCopy-and-paste the text below in your GitHub issue and FILL OUT the two last points.\n" ) print(self.format_dict(__UpperCAmelCase ) ) return info @staticmethod def lowercase_ (__UpperCAmelCase : Dict ) -> Optional[int]: """simple docstring""" return "\n".join([f"""- {prop}: {val}""" for prop, val in d.items()] ) + "\n"
65
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
1
import random import timeit from functools import wraps from typing import Callable, Optional from ..configuration_utils import PretrainedConfig from ..models.auto.modeling_tf_auto import TF_MODEL_MAPPING, TF_MODEL_WITH_LM_HEAD_MAPPING from ..utils import is_pyanvml_available, is_tf_available, logging from .benchmark_utils import ( Benchmark, Memory, MemorySummary, measure_peak_memory_cpu, start_memory_tracing, stop_memory_tracing, ) if is_tf_available(): import tensorflow as tf from tensorflow.python.framework.errors_impl import ResourceExhaustedError from .benchmark_args_tf import TensorFlowBenchmarkArguments if is_pyanvml_available(): import pyanvml.pyanvml as nvml UpperCamelCase__ = logging.get_logger(__name__) def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' def run_func(__A ): @wraps(__A ) def run_in_eager_mode(*__A, **__A ): return func(*__A, **__A ) @wraps(__A ) @tf.function(experimental_compile=__A ) def run_in_graph_mode(*__A, **__A ): return func(*__A, **__A ) if do_eager_mode is True: if use_xla is not False: raise ValueError( "Cannot run model in XLA, if `args.eager_mode` is set to `True`. Please set `args.eager_mode=False`." ) return run_in_eager_mode else: return run_in_graph_mode return run_func def lowerCAmelCase_ ( __A, __A, __A ) -> ["tf.Tensor"]: '''simple docstring''' UpperCAmelCase__ = random.Random() UpperCAmelCase__ = [rng.randint(0, vocab_size - 1 ) for i in range(batch_size * sequence_length )] return tf.constant(__A, shape=(batch_size, sequence_length), dtype=tf.intaa ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : TensorFlowBenchmarkArguments __UpperCAmelCase : PretrainedConfig __UpperCAmelCase : str = "TensorFlow" @property def lowercase_ (self : Any ) -> str: """simple docstring""" return tf.__version__ def lowercase_ (self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> float: """simple docstring""" UpperCAmelCase__ = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase__ = self._prepare_inference_func(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return self._measure_speed(_inference ) def lowercase_ (self : str , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> float: """simple docstring""" UpperCAmelCase__ = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase__ = self._prepare_train_func(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return self._measure_speed(_train ) def lowercase_ (self : Tuple , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> [Memory, Optional[MemorySummary]]: """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , __UpperCAmelCase ) UpperCAmelCase__ = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase__ = self._prepare_inference_func(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return self._measure_memory(_inference ) def lowercase_ (self : Dict , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> [Memory, Optional[MemorySummary]]: """simple docstring""" if self.args.is_gpu: tf.config.experimental.set_memory_growth(self.args.gpu_list[self.args.device_idx] , __UpperCAmelCase ) UpperCAmelCase__ = self.args.strategy if strategy is None: raise ValueError("A device strategy has to be initialized before using TensorFlow." ) UpperCAmelCase__ = self._prepare_train_func(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return self._measure_memory(_train ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> Callable[[], None]: """simple docstring""" UpperCAmelCase__ = self.config_dict[model_name] if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) UpperCAmelCase__ = ( hasattr(__UpperCAmelCase , "architectures" ) and isinstance(config.architectures , __UpperCAmelCase ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: UpperCAmelCase__ = "TF" + config.architectures[0] # prepend 'TF' for tensorflow model UpperCAmelCase__ = __import__("transformers" , fromlist=[model_class] ) UpperCAmelCase__ = getattr(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = model_cls(__UpperCAmelCase ) except ImportError: raise ImportError( f"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: UpperCAmelCase__ = TF_MODEL_MAPPING[config.__class__](__UpperCAmelCase ) # encoder-decoder has vocab size saved differently UpperCAmelCase__ = config.vocab_size if hasattr(__UpperCAmelCase , "vocab_size" ) else config.encoder.vocab_size UpperCAmelCase__ = random_input_ids(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_forward(): return model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase , training=__UpperCAmelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_forward(): return model(__UpperCAmelCase , training=__UpperCAmelCase ) UpperCAmelCase__ = encoder_decoder_forward if config.is_encoder_decoder else encoder_forward return _inference def lowercase_ (self : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> Callable[[], None]: """simple docstring""" UpperCAmelCase__ = self.config_dict[model_name] if self.args.eager_mode is not False: raise ValueError("Training cannot be done in eager mode. Please make sure that `args.eager_mode = False`." ) if self.args.fpaa: raise NotImplementedError("Mixed precision is currently not supported." ) UpperCAmelCase__ = ( hasattr(__UpperCAmelCase , "architectures" ) and isinstance(config.architectures , __UpperCAmelCase ) and len(config.architectures ) > 0 ) if not self.args.only_pretrain_model and has_model_class_in_config: try: UpperCAmelCase__ = "TF" + config.architectures[0] # prepend 'TF' for tensorflow model UpperCAmelCase__ = __import__("transformers" , fromlist=[model_class] ) UpperCAmelCase__ = getattr(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = model_cls(__UpperCAmelCase ) except ImportError: raise ImportError( f"""{model_class} does not exist. If you just want to test the pretrained model, you might want to""" " set `--only_pretrain_model` or `args.only_pretrain_model=True`." ) else: UpperCAmelCase__ = TF_MODEL_WITH_LM_HEAD_MAPPING[config.__class__](__UpperCAmelCase ) # encoder-decoder has vocab size saved differently UpperCAmelCase__ = config.vocab_size if hasattr(__UpperCAmelCase , "vocab_size" ) else config.encoder.vocab_size UpperCAmelCase__ = random_input_ids(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_decoder_train(): UpperCAmelCase__ = model(__UpperCAmelCase , decoder_input_ids=__UpperCAmelCase , labels=__UpperCAmelCase , training=__UpperCAmelCase )[0] UpperCAmelCase__ = tf.gradients(__UpperCAmelCase , model.trainable_variables ) return gradients @run_with_tf_optimizations(self.args.eager_mode , self.args.use_xla ) def encoder_train(): UpperCAmelCase__ = model(__UpperCAmelCase , labels=__UpperCAmelCase , training=__UpperCAmelCase )[0] UpperCAmelCase__ = tf.gradients(__UpperCAmelCase , model.trainable_variables ) return gradients UpperCAmelCase__ = encoder_decoder_train if config.is_encoder_decoder else encoder_train return _train def lowercase_ (self : Tuple , __UpperCAmelCase : Dict ) -> float: """simple docstring""" with self.args.strategy.scope(): try: if self.args.is_tpu or self.args.use_xla: # run additional 10 times to stabilize compilation for tpu logger.info("Do inference on TPU. Running model 5 times to stabilize compilation" ) timeit.repeat(__UpperCAmelCase , repeat=1 , number=5 ) # as written in https://docs.python.org/2/library/timeit.html#timeit.Timer.repeat, min should be taken rather than the average UpperCAmelCase__ = timeit.repeat( __UpperCAmelCase , repeat=self.args.repeat , number=1_0 , ) return min(__UpperCAmelCase ) / 10.0 except ResourceExhaustedError as e: self.print_fn(f"""Doesn't fit on GPU. {e}""" ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Callable[[], None] ) -> [Memory, MemorySummary]: """simple docstring""" logger.info( "Note that TensorFlow allocates more memory than " "it might need to speed up computation. " "The memory reported here corresponds to the memory " "reported by `nvidia-smi`, which can vary depending " "on total available memory on the GPU that is used." ) with self.args.strategy.scope(): try: if self.args.trace_memory_line_by_line: if not self.args.eager_mode: raise ValueError( "`args.eager_mode` is set to `False`. Make sure to run model in eager mode to measure memory" " consumption line by line." ) UpperCAmelCase__ = start_memory_tracing("transformers" ) if self.args.is_tpu: # tpu raise NotImplementedError( "Memory Benchmarking is currently not implemented for TPU. Please disable memory benchmarking" " with `args.memory=False`" ) elif self.args.is_gpu: # gpu if not is_pyanvml_available(): logger.warning( "py3nvml not installed, we won't log GPU memory usage. " "Install py3nvml (pip install py3nvml) to log information about GPU." ) UpperCAmelCase__ = "N/A" else: logger.info( "Measuring total GPU usage on GPU device. Make sure to not have additional processes" " running on the same GPU." ) # init nvml nvml.nvmlInit() func() UpperCAmelCase__ = nvml.nvmlDeviceGetHandleByIndex(self.args.device_idx ) UpperCAmelCase__ = nvml.nvmlDeviceGetMemoryInfo(__UpperCAmelCase ) UpperCAmelCase__ = meminfo.used UpperCAmelCase__ = Memory(__UpperCAmelCase ) # shutdown nvml nvml.nvmlShutdown() else: # cpu if self.args.trace_memory_line_by_line: logger.info( "When enabling line by line tracing, the max peak memory for CPU is inaccurate in" " TensorFlow." ) UpperCAmelCase__ = None else: UpperCAmelCase__ = measure_peak_memory_cpu(__UpperCAmelCase ) UpperCAmelCase__ = Memory(__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else memory_bytes if self.args.trace_memory_line_by_line: UpperCAmelCase__ = stop_memory_tracing(__UpperCAmelCase ) if memory is None: UpperCAmelCase__ = summary.total else: UpperCAmelCase__ = None return memory, summary except ResourceExhaustedError as e: self.print_fn(f"""Doesn't fit on GPU. {e}""" ) return "N/A", None
65
def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' def get_matched_characters(__A, __A ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase__ = int(max(0, i - limit ) ) UpperCAmelCase__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(__A ) UpperCAmelCase__ = f"""{_stra[0:_stra.index(__A )]} {_stra[_stra.index(__A ) + 1:]}""" return "".join(__A ) # matching characters UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = len(__A ) # transposition UpperCAmelCase__ = ( len([(ca, ca) for ca, ca in zip(__A, __A ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase__ = 0.0 else: UpperCAmelCase__ = ( 1 / 3 * ( match_count / len(__A ) + match_count / len(__A ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
65
1
import torch from torch import nn class A ( nn.Module ): def __init__(self : Any , __UpperCAmelCase : Any , __UpperCAmelCase : str , __UpperCAmelCase : str , __UpperCAmelCase : str , __UpperCAmelCase : Dict=1 , __UpperCAmelCase : int=False ) -> List[Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = n_token UpperCAmelCase__ = d_embed UpperCAmelCase__ = d_proj UpperCAmelCase__ = cutoffs + [n_token] UpperCAmelCase__ = [0] + self.cutoffs UpperCAmelCase__ = div_val UpperCAmelCase__ = self.cutoffs[0] UpperCAmelCase__ = len(self.cutoffs ) - 1 UpperCAmelCase__ = self.shortlist_size + self.n_clusters if self.n_clusters > 0: UpperCAmelCase__ = nn.Parameter(torch.zeros(self.n_clusters , self.d_embed ) ) UpperCAmelCase__ = nn.Parameter(torch.zeros(self.n_clusters ) ) UpperCAmelCase__ = nn.ModuleList() UpperCAmelCase__ = nn.ParameterList() if div_val == 1: for i in range(len(self.cutoffs ) ): if d_proj != d_embed: self.out_projs.append(nn.Parameter(torch.FloatTensor(__UpperCAmelCase , __UpperCAmelCase ) ) ) else: self.out_projs.append(__UpperCAmelCase ) self.out_layers.append(nn.Linear(__UpperCAmelCase , __UpperCAmelCase ) ) else: for i in range(len(self.cutoffs ) ): UpperCAmelCase__ , UpperCAmelCase__ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ = d_embed // (div_val**i) self.out_projs.append(nn.Parameter(torch.FloatTensor(__UpperCAmelCase , __UpperCAmelCase ) ) ) self.out_layers.append(nn.Linear(__UpperCAmelCase , r_idx - l_idx ) ) UpperCAmelCase__ = keep_order def lowercase_ (self : int , __UpperCAmelCase : Any , __UpperCAmelCase : Any , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Any ) -> int: """simple docstring""" if proj is None: UpperCAmelCase__ = nn.functional.linear(__UpperCAmelCase , __UpperCAmelCase , bias=__UpperCAmelCase ) else: # if CUDA_MAJOR <= 9 and CUDA_MINOR <= 1: UpperCAmelCase__ = nn.functional.linear(__UpperCAmelCase , proj.t().contiguous() ) UpperCAmelCase__ = nn.functional.linear(__UpperCAmelCase , __UpperCAmelCase , bias=__UpperCAmelCase ) # else: # logit = torch.einsum('bd,de,ev->bv', (hidden, proj, weight.t())) # if bias is not None: # logit = logit + bias return logit def lowercase_ (self : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" if labels is not None: # Shift so that tokens < n predict n UpperCAmelCase__ = hidden[..., :-1, :].contiguous() UpperCAmelCase__ = labels[..., 1:].contiguous() UpperCAmelCase__ = hidden.view(-1 , hidden.size(-1 ) ) UpperCAmelCase__ = labels.view(-1 ) if hidden.size(0 ) != labels.size(0 ): raise RuntimeError("Input and labels should have the same size in the batch dimension." ) else: UpperCAmelCase__ = hidden.view(-1 , hidden.size(-1 ) ) if self.n_clusters == 0: UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) if labels is not None: UpperCAmelCase__ = labels != -1_0_0 UpperCAmelCase__ = torch.zeros_like(__UpperCAmelCase , dtype=hidden.dtype , device=hidden.device ) UpperCAmelCase__ = ( -nn.functional.log_softmax(__UpperCAmelCase , dim=-1 )[mask].gather(1 , labels[mask].unsqueeze(1 ) ).squeeze(1 ) ) else: UpperCAmelCase__ = nn.functional.log_softmax(__UpperCAmelCase , dim=-1 ) else: # construct weights and biases UpperCAmelCase__ , UpperCAmelCase__ = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: UpperCAmelCase__ , UpperCAmelCase__ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase__ = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase__ = self.out_layers[i].weight UpperCAmelCase__ = self.out_layers[i].bias if i == 0: UpperCAmelCase__ = torch.cat([weight_i, self.cluster_weight] , dim=0 ) UpperCAmelCase__ = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(__UpperCAmelCase ) biases.append(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = weights[0], biases[0], self.out_projs[0] UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = nn.functional.log_softmax(__UpperCAmelCase , dim=1 ) if labels is None: UpperCAmelCase__ = hidden.new_empty((head_logit.size(0 ), self.n_token) ) else: UpperCAmelCase__ = torch.zeros_like(__UpperCAmelCase , dtype=hidden.dtype , device=hidden.device ) UpperCAmelCase__ = 0 UpperCAmelCase__ = [0] + self.cutoffs for i in range(len(__UpperCAmelCase ) - 1 ): UpperCAmelCase__ , UpperCAmelCase__ = cutoff_values[i], cutoff_values[i + 1] if labels is not None: UpperCAmelCase__ = (labels >= l_idx) & (labels < r_idx) UpperCAmelCase__ = mask_i.nonzero().squeeze() if indices_i.numel() == 0: continue UpperCAmelCase__ = labels.index_select(0 , __UpperCAmelCase ) - l_idx UpperCAmelCase__ = head_logprob.index_select(0 , __UpperCAmelCase ) UpperCAmelCase__ = hidden.index_select(0 , __UpperCAmelCase ) else: UpperCAmelCase__ = hidden if i == 0: if labels is not None: UpperCAmelCase__ = head_logprob_i.gather(1 , target_i[:, None] ).squeeze(1 ) else: UpperCAmelCase__ = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = weights[i], biases[i], self.out_projs[i] UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = nn.functional.log_softmax(__UpperCAmelCase , dim=1 ) UpperCAmelCase__ = self.cutoffs[0] + i - 1 # No probability for the head cluster if labels is not None: UpperCAmelCase__ = head_logprob_i[:, cluster_prob_idx] + tail_logprob_i.gather( 1 , target_i[:, None] ).squeeze(1 ) else: UpperCAmelCase__ = head_logprob[:, cluster_prob_idx, None] + tail_logprob_i UpperCAmelCase__ = logprob_i if labels is not None: if (hasattr(self , "keep_order" ) and self.keep_order) or keep_order: out.index_copy_(0 , __UpperCAmelCase , -logprob_i ) else: out[offset : offset + logprob_i.size(0 )].copy_(-logprob_i ) offset += logprob_i.size(0 ) return out def lowercase_ (self : Optional[int] , __UpperCAmelCase : Any ) -> str: """simple docstring""" if self.n_clusters == 0: UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , self.out_layers[0].weight , self.out_layers[0].bias , self.out_projs[0] ) return nn.functional.log_softmax(__UpperCAmelCase , dim=-1 ) else: # construct weights and biases UpperCAmelCase__ , UpperCAmelCase__ = [], [] for i in range(len(self.cutoffs ) ): if self.div_val == 1: UpperCAmelCase__ , UpperCAmelCase__ = self.cutoff_ends[i], self.cutoff_ends[i + 1] UpperCAmelCase__ = self.out_layers[0].weight[l_idx:r_idx] UpperCAmelCase__ = self.out_layers[0].bias[l_idx:r_idx] else: UpperCAmelCase__ = self.out_layers[i].weight UpperCAmelCase__ = self.out_layers[i].bias if i == 0: UpperCAmelCase__ = torch.cat([weight_i, self.cluster_weight] , dim=0 ) UpperCAmelCase__ = torch.cat([bias_i, self.cluster_bias] , dim=0 ) weights.append(__UpperCAmelCase ) biases.append(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = weights[0], biases[0], self.out_projs[0] UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = hidden.new_empty((head_logit.size(0 ), self.n_token) ) UpperCAmelCase__ = nn.functional.log_softmax(__UpperCAmelCase , dim=1 ) UpperCAmelCase__ = [0] + self.cutoffs for i in range(len(__UpperCAmelCase ) - 1 ): UpperCAmelCase__ , UpperCAmelCase__ = cutoff_values[i], cutoff_values[i + 1] if i == 0: UpperCAmelCase__ = head_logprob[:, : self.cutoffs[0]] else: UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = weights[i], biases[i], self.out_projs[i] UpperCAmelCase__ = self._compute_logit(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = nn.functional.log_softmax(__UpperCAmelCase , dim=1 ) UpperCAmelCase__ = head_logprob[:, -i] + tail_logprob_i UpperCAmelCase__ = logprob_i return out
65
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
1
from __future__ import annotations def lowerCAmelCase_ ( __A, __A, __A, ) -> tuple: '''simple docstring''' if (electron_conc, hole_conc, intrinsic_conc).count(0 ) != 1: raise ValueError("You cannot supply more or less than 2 values" ) elif electron_conc < 0: raise ValueError("Electron concentration cannot be negative in a semiconductor" ) elif hole_conc < 0: raise ValueError("Hole concentration cannot be negative in a semiconductor" ) elif intrinsic_conc < 0: raise ValueError( "Intrinsic concentration cannot be negative in a semiconductor" ) elif electron_conc == 0: return ( "electron_conc", intrinsic_conc**2 / hole_conc, ) elif hole_conc == 0: return ( "hole_conc", intrinsic_conc**2 / electron_conc, ) elif intrinsic_conc == 0: return ( "intrinsic_conc", (electron_conc * hole_conc) ** 0.5, ) else: return (-1, -1) if __name__ == "__main__": import doctest doctest.testmod()
65
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCamelCase__ = 'base_with_context' def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["self_attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["MultiHeadDotProductAttention_0"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array, __A ) UpperCAmelCase__ = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] UpperCAmelCase__ = os.path.join(args.checkpoint_path, "..", "config.gin" ) UpperCAmelCase__ = inference.parse_training_gin_file(__A, __A ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path, __A ) UpperCAmelCase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2", variance_type="fixed_large" ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"], vocab_size=synth_model.model.module.config.vocab_size, d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims, targets_context_length=synth_model.sequence_length["targets_context"], d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims, targets_length=synth_model.sequence_length["targets_context"], max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time, d_model=synth_model.model.module.config.emb_dim, num_layers=synth_model.model.module.config.num_decoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, dropout_rate=synth_model.model.module.config.dropout_rate, ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint["target"]["token_encoder"], __A ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"], __A ) UpperCAmelCase__ = load_decoder(ta_checkpoint["target"]["decoder"], __A ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=__A, continuous_encoder=__A, decoder=__A, scheduler=__A, melgan=__A, ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help='Path to the original jax model checkpoint.', ) UpperCamelCase__ = parser.parse_args() main(args)
65
1
from typing import List, Optional, Tuple, Union import torch from ...models import UNetaDModel from ...schedulers import KarrasVeScheduler from ...utils import randn_tensor from ..pipeline_utils import DiffusionPipeline, ImagePipelineOutput class A ( UpperCAmelCase_ ): __UpperCAmelCase : UNetaDModel __UpperCAmelCase : KarrasVeScheduler def __init__(self : int , __UpperCAmelCase : UNetaDModel , __UpperCAmelCase : KarrasVeScheduler ) -> List[Any]: """simple docstring""" super().__init__() self.register_modules(unet=__UpperCAmelCase , scheduler=__UpperCAmelCase ) @torch.no_grad() def __call__(self : Optional[int] , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 5_0 , __UpperCAmelCase : Optional[Union[torch.Generator, List[torch.Generator]]] = None , __UpperCAmelCase : Optional[str] = "pil" , __UpperCAmelCase : bool = True , **__UpperCAmelCase : List[Any] , ) -> Union[Tuple, ImagePipelineOutput]: """simple docstring""" UpperCAmelCase__ = self.unet.config.sample_size UpperCAmelCase__ = (batch_size, 3, img_size, img_size) UpperCAmelCase__ = self.unet # sample x_0 ~ N(0, sigma_0^2 * I) UpperCAmelCase__ = randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase , device=self.device ) * self.scheduler.init_noise_sigma self.scheduler.set_timesteps(__UpperCAmelCase ) for t in self.progress_bar(self.scheduler.timesteps ): # here sigma_t == t_i from the paper UpperCAmelCase__ = self.scheduler.schedule[t] UpperCAmelCase__ = self.scheduler.schedule[t - 1] if t > 0 else 0 # 1. Select temporarily increased noise level sigma_hat # 2. Add new noise to move from sample_i to sample_hat UpperCAmelCase__ , UpperCAmelCase__ = self.scheduler.add_noise_to_input(__UpperCAmelCase , __UpperCAmelCase , generator=__UpperCAmelCase ) # 3. Predict the noise residual given the noise magnitude `sigma_hat` # The model inputs and output are adjusted by following eq. (213) in [1]. UpperCAmelCase__ = (sigma_hat / 2) * model((sample_hat + 1) / 2 , sigma_hat / 2 ).sample # 4. Evaluate dx/dt at sigma_hat # 5. Take Euler step from sigma to sigma_prev UpperCAmelCase__ = self.scheduler.step(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) if sigma_prev != 0: # 6. Apply 2nd order correction # The model inputs and output are adjusted by following eq. (213) in [1]. UpperCAmelCase__ = (sigma_prev / 2) * model((step_output.prev_sample + 1) / 2 , sigma_prev / 2 ).sample UpperCAmelCase__ = self.scheduler.step_correct( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , step_output.prev_sample , step_output["derivative"] , ) UpperCAmelCase__ = step_output.prev_sample UpperCAmelCase__ = (sample / 2 + 0.5).clamp(0 , 1 ) UpperCAmelCase__ = sample.cpu().permute(0 , 2 , 3 , 1 ).numpy() if output_type == "pil": UpperCAmelCase__ = self.numpy_to_pil(__UpperCAmelCase ) if not return_dict: return (image,) return ImagePipelineOutput(images=__UpperCAmelCase )
65
import math def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' return math.sqrt(__A ) * math.sqrt(__A ) == num def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' UpperCAmelCase__ = 0 UpperCAmelCase__ = n while left <= right: UpperCAmelCase__ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase__ = mid - 1 else: UpperCAmelCase__ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
65
1
import mpmath # for roots of unity import numpy as np class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : Any=None , __UpperCAmelCase : Optional[int]=None ) -> Dict: """simple docstring""" UpperCAmelCase__ = list(poly_a or [0] )[:] UpperCAmelCase__ = list(poly_b or [0] )[:] # Remove leading zero coefficients while self.polyA[-1] == 0: self.polyA.pop() UpperCAmelCase__ = len(self.polyA ) while self.polyB[-1] == 0: self.polyB.pop() UpperCAmelCase__ = len(self.polyB ) # Add 0 to make lengths equal a power of 2 UpperCAmelCase__ = int( 2 ** np.ceil(np.loga(len(self.polyA ) + len(self.polyB ) - 1 ) ) ) while len(self.polyA ) < self.c_max_length: self.polyA.append(0 ) while len(self.polyB ) < self.c_max_length: self.polyB.append(0 ) # A complex root used for the fourier transform UpperCAmelCase__ = complex(mpmath.root(x=1 , n=self.c_max_length , k=1 ) ) # The product UpperCAmelCase__ = self.__multiply() def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = [[x] for x in self.polyA] if which == "A" else [[x] for x in self.polyB] # Corner case if len(__UpperCAmelCase ) <= 1: return dft[0] # UpperCAmelCase__ = self.c_max_length // 2 while next_ncol > 0: UpperCAmelCase__ = [[] for i in range(__UpperCAmelCase )] UpperCAmelCase__ = self.root**next_ncol # First half of next step UpperCAmelCase__ = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__UpperCAmelCase ): new_dft[i].append(dft[i][j] + current_root * dft[i + next_ncol][j] ) current_root *= root # Second half of next step UpperCAmelCase__ = 1 for j in range(self.c_max_length // (next_ncol * 2) ): for i in range(__UpperCAmelCase ): new_dft[i].append(dft[i][j] - current_root * dft[i + next_ncol][j] ) current_root *= root # Update UpperCAmelCase__ = new_dft UpperCAmelCase__ = next_ncol // 2 return dft[0] def lowercase_ (self : str ) -> int: """simple docstring""" UpperCAmelCase__ = self.__dft("A" ) UpperCAmelCase__ = self.__dft("B" ) UpperCAmelCase__ = [[dft_a[i] * dft_b[i] for i in range(self.c_max_length )]] del dft_a del dft_b # Corner Case if len(inverce_c[0] ) <= 1: return inverce_c[0] # Inverse DFT UpperCAmelCase__ = 2 while next_ncol <= self.c_max_length: UpperCAmelCase__ = [[] for i in range(__UpperCAmelCase )] UpperCAmelCase__ = self.root ** (next_ncol // 2) UpperCAmelCase__ = 1 # First half of next step for j in range(self.c_max_length // next_ncol ): for i in range(next_ncol // 2 ): # Even positions new_inverse_c[i].append( ( inverce_c[i][j] + inverce_c[i][j + self.c_max_length // next_ncol] ) / 2 ) # Odd positions new_inverse_c[i + next_ncol // 2].append( ( inverce_c[i][j] - inverce_c[i][j + self.c_max_length // next_ncol] ) / (2 * current_root) ) current_root *= root # Update UpperCAmelCase__ = new_inverse_c next_ncol *= 2 # Unpack UpperCAmelCase__ = [round(x[0].real , 8 ) + round(x[0].imag , 8 ) * 1j for x in inverce_c] # Remove leading 0's while inverce_c[-1] == 0: inverce_c.pop() return inverce_c def __str__(self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase__ = "A = " + " + ".join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.polyA[: self.len_A] ) ) UpperCAmelCase__ = "B = " + " + ".join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.polyB[: self.len_B] ) ) UpperCAmelCase__ = "A*B = " + " + ".join( f"""{coef}*x^{i}""" for coef, i in enumerate(self.product ) ) return f"""{a}\n{b}\n{c}""" # Unit tests if __name__ == "__main__": import doctest doctest.testmod()
65
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor __UpperCAmelCase : Optional[torch.FloatTensor] = None def lowerCAmelCase_ ( __A, __A=0.999, __A="cosine", ) -> Tuple: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__A ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase__ = [] for i in range(__A ): UpperCAmelCase__ = i / num_diffusion_timesteps UpperCAmelCase__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__A ) / alpha_bar_fn(__A ), __A ) ) return torch.tensor(__A, dtype=torch.floataa ) class A ( UpperCAmelCase_ , UpperCAmelCase_ ): @register_to_config def __init__(self : List[str] , __UpperCAmelCase : int = 1_0_0_0 , __UpperCAmelCase : str = "fixed_small_log" , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[float] = 1.0 , __UpperCAmelCase : str = "epsilon" , __UpperCAmelCase : str = "squaredcos_cap_v2" , ) -> Optional[int]: """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) UpperCAmelCase__ = betas_for_alpha_bar(__UpperCAmelCase ) UpperCAmelCase__ = 1.0 - self.betas UpperCAmelCase__ = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase__ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase__ = 1.0 # setable values UpperCAmelCase__ = None UpperCAmelCase__ = torch.from_numpy(np.arange(0 , __UpperCAmelCase )[::-1].copy() ) UpperCAmelCase__ = variance_type def lowercase_ (self : List[str] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None ) -> torch.FloatTensor: """simple docstring""" return sample def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, torch.device] = None ) -> Any: """simple docstring""" UpperCAmelCase__ = num_inference_steps UpperCAmelCase__ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase__ = (np.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ).to(__UpperCAmelCase ) def lowercase_ (self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase__ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase__ = torch.log(torch.clamp(__UpperCAmelCase , min=1E-20 ) ) UpperCAmelCase__ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase__ = variance.log() UpperCAmelCase__ = beta.log() UpperCAmelCase__ = (predicted_variance + 1) / 2 UpperCAmelCase__ = frac * max_log + (1 - frac) * min_log return variance def lowercase_ (self : Optional[int] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : int , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : bool = True , ) -> Union[UnCLIPSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase__ , UpperCAmelCase__ = torch.split(__UpperCAmelCase , sample.shape[1] , dim=1 ) else: UpperCAmelCase__ = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] UpperCAmelCase__ = self.alphas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase__ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase__ = model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase__ = torch.clamp( __UpperCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase__ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase__ = 0 if t > 0: UpperCAmelCase__ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__UpperCAmelCase , device=model_output.device ) UpperCAmelCase__ = self._get_variance( __UpperCAmelCase , predicted_variance=__UpperCAmelCase , prev_timestep=__UpperCAmelCase , ) if self.variance_type == "fixed_small_log": UpperCAmelCase__ = variance elif self.variance_type == "learned_range": UpperCAmelCase__ = (0.5 * variance).exp() else: raise ValueError( f"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" " for the UnCLIPScheduler." ) UpperCAmelCase__ = variance * variance_noise UpperCAmelCase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__UpperCAmelCase , pred_original_sample=__UpperCAmelCase ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.IntTensor , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase__ = timesteps.to(original_samples.device ) UpperCAmelCase__ = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase__ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase__ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
65
1
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class A ( unittest.TestCase ): def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) UpperCAmelCase__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def lowercase_ (self : List[str] ) -> Any: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : str ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.operation_file_path] print(f"""Command: {cmd}""" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Tuple ) -> int: """simple docstring""" UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Dict ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices, using 2 devices only""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase__ = Accelerator() UpperCamelCase__ = (accelerator.state.process_index + 2, 1_0) UpperCamelCase__ = torch.randint(0, 1_0, shape).to(accelerator.device) UpperCamelCase__ = '' UpperCamelCase__ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." UpperCamelCase__ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." UpperCamelCase__ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
65
1
from __future__ import annotations def lowerCAmelCase_ ( __A ) -> float: '''simple docstring''' UpperCAmelCase__ = 0.00 UpperCAmelCase__ = 0 for resistor in resistors: if resistor <= 0: UpperCAmelCase__ = f"""Resistor at index {index} has a negative or zero value!""" raise ValueError(__A ) first_sum += 1 / float(__A ) index += 1 return 1 / first_sum def lowerCAmelCase_ ( __A ) -> float: '''simple docstring''' UpperCAmelCase__ = 0.00 UpperCAmelCase__ = 0 for resistor in resistors: sum_r += resistor if resistor < 0: UpperCAmelCase__ = f"""Resistor at index {index} has a negative value!""" raise ValueError(__A ) index += 1 return sum_r if __name__ == "__main__": import doctest doctest.testmod()
65
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(__A, __A ) def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = emb.weight.shape UpperCAmelCase__ = nn.Linear(__A, __A, bias=__A ) UpperCAmelCase__ = emb.weight.data return lin_layer def lowerCAmelCase_ ( __A, __A="facebook/mbart-large-en-ro", __A=False, __A=False ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = torch.load(__A, map_location="cpu" )["model"] remove_ignore_keys_(__A ) UpperCAmelCase__ = state_dict["encoder.embed_tokens.weight"].shape[0] UpperCAmelCase__ = MBartConfig.from_pretrained(__A, vocab_size=__A ) if mbart_aa and finetuned: UpperCAmelCase__ = "relu" UpperCAmelCase__ = state_dict["decoder.embed_tokens.weight"] UpperCAmelCase__ = MBartForConditionalGeneration(__A ) model.model.load_state_dict(__A ) if finetuned: UpperCAmelCase__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
65
1
import inspect import jax import jax.lax as lax import jax.numpy as jnp from ..utils import add_start_docstrings from ..utils.logging import get_logger UpperCamelCase__ = get_logger(__name__) UpperCamelCase__ = R'\n Args:\n input_ids (`jnp.ndarray` of shape `(batch_size, sequence_length)`):\n Indices of input sequence tokens in the vocabulary.\n\n Indices can be obtained using [`PreTrainedTokenizer`]. See [`PreTrainedTokenizer.encode`] and\n [`PreTrainedTokenizer.__call__`] for details.\n\n [What are input IDs?](../glossary#input-ids)\n scores (`jnp.ndarray` of shape `(batch_size, config.vocab_size)`):\n Prediction scores of a language modeling head. These can be logits for each vocabulary when not using beam\n search or log softmax for each vocabulary token when using beam search\n kwargs (`Dict[str, Any]`, *optional*):\n Additional logits processor specific kwargs.\n\n Return:\n `jnp.ndarray` of shape `(batch_size, config.vocab_size)`: The processed prediction scores.\n\n' class A : @add_start_docstrings(__UpperCAmelCase ) def __call__(self : Optional[Any] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray ) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class A : @add_start_docstrings(__UpperCAmelCase ) def __call__(self : Tuple , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray ) -> jnp.ndarray: """simple docstring""" raise NotImplementedError( f"""{self.__class__} is an abstract class. Only classes inheriting this class can be called.""" ) class A ( UpperCAmelCase_ ): @add_start_docstrings(__UpperCAmelCase ) def __call__(self : Tuple , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int , **__UpperCAmelCase : str ) -> jnp.ndarray: """simple docstring""" for processor in self: UpperCAmelCase__ = inspect.signature(processor.__call__ ).parameters if len(__UpperCAmelCase ) > 3: if not all(arg in kwargs for arg in list(function_args.keys() )[2:] ): raise ValueError( f"""Make sure that all the required parameters: {list(function_args.keys() )} for """ f"""{processor.__class__} are passed to the logits processor.""" ) UpperCAmelCase__ = processor(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) else: UpperCAmelCase__ = processor(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) return scores class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : float ) -> Union[str, Any]: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or not (temperature > 0): raise ValueError(f"""`temperature` has to be a strictly positive float, but is {temperature}""" ) UpperCAmelCase__ = temperature def __call__(self : Union[str, Any] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ = scores / self.temperature return scores class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : float , __UpperCAmelCase : float = -float("Inf" ) , __UpperCAmelCase : int = 1 ) -> str: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or (top_p < 0 or top_p > 1.0): raise ValueError(f"""`top_p` has to be a float > 0 and < 1, but is {top_p}""" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or (min_tokens_to_keep < 1): raise ValueError(f"""`min_tokens_to_keep` has to be a positive integer, but is {min_tokens_to_keep}""" ) UpperCAmelCase__ = top_p UpperCAmelCase__ = filter_value UpperCAmelCase__ = min_tokens_to_keep def __call__(self : Any , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = lax.top_k(__UpperCAmelCase , scores.shape[-1] ) UpperCAmelCase__ = jnp.full_like(__UpperCAmelCase , self.filter_value ) UpperCAmelCase__ = jax.nn.softmax(__UpperCAmelCase , axis=-1 ).cumsum(axis=-1 ) UpperCAmelCase__ = cumulative_probs < self.top_p # include the token that is higher than top_p as well UpperCAmelCase__ = jnp.roll(__UpperCAmelCase , 1 ) score_mask |= score_mask.at[:, 0].set(__UpperCAmelCase ) # min tokens to keep UpperCAmelCase__ = score_mask.at[:, : self.min_tokens_to_keep].set(__UpperCAmelCase ) UpperCAmelCase__ = jnp.where(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jax.lax.sort_key_val(__UpperCAmelCase , __UpperCAmelCase )[-1] return next_scores class A ( UpperCAmelCase_ ): def __init__(self : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : float = -float("Inf" ) , __UpperCAmelCase : int = 1 ) -> str: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or top_k <= 0: raise ValueError(f"""`top_k` has to be a strictly positive integer, but is {top_k}""" ) UpperCAmelCase__ = max(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = filter_value def __call__(self : Any , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = scores.shape UpperCAmelCase__ = jnp.full(batch_size * vocab_size , self.filter_value ) UpperCAmelCase__ = min(self.top_k , scores.shape[-1] ) # Safety check UpperCAmelCase__ , UpperCAmelCase__ = lax.top_k(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jnp.broadcast_to((jnp.arange(__UpperCAmelCase ) * vocab_size)[:, None] , (batch_size, topk) ).flatten() UpperCAmelCase__ = topk_scores.flatten() UpperCAmelCase__ = topk_indices.flatten() + shift UpperCAmelCase__ = next_scores_flat.at[topk_indices_flat].set(__UpperCAmelCase ) UpperCAmelCase__ = next_scores_flat.reshape(__UpperCAmelCase , __UpperCAmelCase ) return next_scores class A ( UpperCAmelCase_ ): def __init__(self : Optional[Any] , __UpperCAmelCase : int ) -> Tuple: """simple docstring""" UpperCAmelCase__ = bos_token_id def __call__(self : Optional[Any] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ = jnp.full(scores.shape , -float("inf" ) ) UpperCAmelCase__ = 1 - jnp.bool_(cur_len - 1 ) UpperCAmelCase__ = jnp.where(__UpperCAmelCase , new_scores.at[:, self.bos_token_id].set(0 ) , __UpperCAmelCase ) return scores class A ( UpperCAmelCase_ ): def __init__(self : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = max_length UpperCAmelCase__ = eos_token_id def __call__(self : Optional[int] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ = jnp.full(scores.shape , -float("inf" ) ) UpperCAmelCase__ = 1 - jnp.bool_(cur_len - self.max_length + 1 ) UpperCAmelCase__ = jnp.where(__UpperCAmelCase , new_scores.at[:, self.eos_token_id].set(0 ) , __UpperCAmelCase ) return scores class A ( UpperCAmelCase_ ): def __init__(self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> Tuple: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or min_length < 0: raise ValueError(f"""`min_length` has to be a positive integer, but is {min_length}""" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ) or eos_token_id < 0: raise ValueError(f"""`eos_token_id` has to be a positive integer, but is {eos_token_id}""" ) UpperCAmelCase__ = min_length UpperCAmelCase__ = eos_token_id def __call__(self : Optional[int] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ = 1 - jnp.clip(cur_len - self.min_length , 0 , 1 ) UpperCAmelCase__ = jnp.where(__UpperCAmelCase , scores.at[:, self.eos_token_id].set(-float("inf" ) ) , __UpperCAmelCase ) return scores class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : Tuple ) -> List[str]: """simple docstring""" UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = begin_index def __call__(self : Any , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = 1 - jnp.bool_(cur_len - self.begin_index ) UpperCAmelCase__ = jnp.where(__UpperCAmelCase , scores.at[:, self.begin_suppress_tokens].set(-float("inf" ) ) , __UpperCAmelCase ) return scores class A ( UpperCAmelCase_ ): def __init__(self : List[Any] , __UpperCAmelCase : list ) -> str: """simple docstring""" UpperCAmelCase__ = list(__UpperCAmelCase ) def __call__(self : List[str] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" UpperCAmelCase__ = scores.at[..., self.suppress_tokens].set(-float("inf" ) ) return scores class A ( UpperCAmelCase_ ): def __init__(self : List[Any] , __UpperCAmelCase : List[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = dict(__UpperCAmelCase ) # Converts the dictionary of format {index: token} containing the tokens to be forced to an array, where the # index of the array corresponds to the index of the token to be forced, for XLA compatibility. # Indexes without forced tokens will have a negative value. UpperCAmelCase__ = jnp.ones((max(force_token_map.keys() ) + 1) , dtype=jnp.intaa ) * -1 for index, token in force_token_map.items(): if token is not None: UpperCAmelCase__ = force_token_array.at[index].set(__UpperCAmelCase ) UpperCAmelCase__ = jnp.intaa(__UpperCAmelCase ) def __call__(self : List[str] , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : jnp.ndarray , __UpperCAmelCase : int ) -> jnp.ndarray: """simple docstring""" def _force_token(__UpperCAmelCase : List[Any] ): UpperCAmelCase__ = scores.shape[0] UpperCAmelCase__ = self.force_token_array[generation_idx] UpperCAmelCase__ = jnp.ones_like(__UpperCAmelCase , dtype=scores.dtype ) * -float("inf" ) UpperCAmelCase__ = jnp.zeros((batch_size, 1) , dtype=scores.dtype ) UpperCAmelCase__ = lax.dynamic_update_slice(__UpperCAmelCase , __UpperCAmelCase , (0, current_token) ) return new_scores UpperCAmelCase__ = lax.cond( cur_len >= self.force_token_array.shape[0] , lambda: scores , lambda: lax.cond( self.force_token_array[cur_len] >= 0 , lambda: _force_token(__UpperCAmelCase ) , lambda: scores , ) , ) return scores class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = generate_config.eos_token_id UpperCAmelCase__ = generate_config.no_timestamps_token_id UpperCAmelCase__ = generate_config.no_timestamps_token_id + 1 UpperCAmelCase__ = decoder_input_length + 1 if generate_config.is_multilingual: # room for language token and task token self.begin_index += 2 if hasattr(__UpperCAmelCase , "max_initial_timestamp_index" ): UpperCAmelCase__ = generate_config.max_initial_timestamp_index else: UpperCAmelCase__ = model_config.vocab_size if self.max_initial_timestamp_index is None: UpperCAmelCase__ = model_config.vocab_size def __call__(self : int , __UpperCAmelCase : Any , __UpperCAmelCase : Any , __UpperCAmelCase : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = scores.at[:, self.no_timestamps_token_id].set(-float("inf" ) ) def handle_pairs(__UpperCAmelCase : Any , __UpperCAmelCase : Optional[int] ): UpperCAmelCase__ = jnp.where((cur_len - self.begin_index) >= 1 , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jnp.where( input_ids_k[cur_len - 1] >= self.timestamp_begin , True and last_was_timestamp , __UpperCAmelCase , ) UpperCAmelCase__ = jnp.where((cur_len - self.begin_index) < 2 , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jnp.where( input_ids_k[cur_len - 2] >= self.timestamp_begin , __UpperCAmelCase , __UpperCAmelCase , ) return jnp.where( __UpperCAmelCase , jnp.where( penultimate_was_timestamp > 0 , scores_k.at[self.timestamp_begin :].set(-float("inf" ) ) , scores_k.at[: self.eos_token_id].set(-float("inf" ) ) , ) , __UpperCAmelCase , ) UpperCAmelCase__ = jax.vmap(__UpperCAmelCase )(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jnp.where(cur_len == self.begin_index , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = jnp.where( self.max_initial_timestamp_index is not None , True and apply_max_initial_timestamp , __UpperCAmelCase , ) UpperCAmelCase__ = self.timestamp_begin + self.max_initial_timestamp_index UpperCAmelCase__ = jnp.where( __UpperCAmelCase , scores.at[:, last_allowed + 1 :].set(-float("inf" ) ) , __UpperCAmelCase , ) # if sum of probability over timestamps is above any other token, sample timestamp UpperCAmelCase__ = jax.nn.log_softmax(__UpperCAmelCase , axis=-1 ) def handle_cumulative_probs(__UpperCAmelCase : int , __UpperCAmelCase : Any ): UpperCAmelCase__ = jax.nn.logsumexp(logprobs_k[self.timestamp_begin :] , axis=-1 ) UpperCAmelCase__ = jnp.max(logprobs_k[: self.timestamp_begin] ) return jnp.where( timestamp_logprob > max_text_token_logprob , scores_k.at[: self.timestamp_begin].set(-float("inf" ) ) , __UpperCAmelCase , ) UpperCAmelCase__ = jax.vmap(__UpperCAmelCase )(__UpperCAmelCase , __UpperCAmelCase ) return scores
65
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase__ = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def lowerCAmelCase_ ( __A, __A=None ) -> Dict: '''simple docstring''' require_version(deps[pkg], __A )
65
1
from __future__ import annotations import os import tempfile import unittest from transformers import ConvBertConfig, is_tf_available from transformers.testing_utils import require_tf, slow from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertModel, ) class A : def __init__(self : Any , __UpperCAmelCase : Any , __UpperCAmelCase : Tuple=1_3 , __UpperCAmelCase : int=7 , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Any=True , __UpperCAmelCase : Dict=True , __UpperCAmelCase : str=True , __UpperCAmelCase : Union[str, Any]=9_9 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : List[Any]=2 , __UpperCAmelCase : List[Any]=4 , __UpperCAmelCase : str=3_7 , __UpperCAmelCase : Union[str, Any]="gelu" , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : List[Any]=0.1 , __UpperCAmelCase : Optional[Any]=5_1_2 , __UpperCAmelCase : Tuple=1_6 , __UpperCAmelCase : Any=2 , __UpperCAmelCase : Union[str, Any]=0.02 , __UpperCAmelCase : Optional[int]=3 , __UpperCAmelCase : str=4 , __UpperCAmelCase : str=None , ) -> Any: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = 1_3 UpperCAmelCase__ = 7 UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = 9_9 UpperCAmelCase__ = 3_8_4 UpperCAmelCase__ = 2 UpperCAmelCase__ = 4 UpperCAmelCase__ = 3_7 UpperCAmelCase__ = "gelu" UpperCAmelCase__ = 0.1 UpperCAmelCase__ = 0.1 UpperCAmelCase__ = 5_1_2 UpperCAmelCase__ = 1_6 UpperCAmelCase__ = 2 UpperCAmelCase__ = 0.02 UpperCAmelCase__ = 3 UpperCAmelCase__ = 4 UpperCAmelCase__ = 1_2_8 UpperCAmelCase__ = 2 UpperCAmelCase__ = 9 UpperCAmelCase__ = 1 UpperCAmelCase__ = None def lowercase_ (self : Any ) -> str: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_input_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.num_labels ) UpperCAmelCase__ = ids_tensor([self.batch_size] , self.num_choices ) UpperCAmelCase__ = ConvBertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , return_dict=__UpperCAmelCase , ) return config, input_ids, token_type_ids, input_mask, sequence_labels, token_labels, choice_labels def lowercase_ (self : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = TFConvBertModel(config=__UpperCAmelCase ) UpperCAmelCase__ = {"input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids} UpperCAmelCase__ = [input_ids, input_mask] UpperCAmelCase__ = model(__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict , __UpperCAmelCase : Dict , __UpperCAmelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = TFConvBertForMaskedLM(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.vocab_size) ) def lowercase_ (self : List[Any] , __UpperCAmelCase : int , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[Any] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFConvBertForSequenceClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = self.num_choices UpperCAmelCase__ = TFConvBertForMultipleChoice(config=__UpperCAmelCase ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.num_choices, 1) ) UpperCAmelCase__ = { "input_ids": multiple_choice_inputs_ids, "attention_mask": multiple_choice_input_mask, "token_type_ids": multiple_choice_token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_choices) ) def lowercase_ (self : Any , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : str , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFConvBertForTokenClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.seq_length, self.num_labels) ) def lowercase_ (self : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Any , __UpperCAmelCase : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = TFConvBertForQuestionAnswering(config=__UpperCAmelCase ) UpperCAmelCase__ = { "input_ids": input_ids, "attention_mask": input_mask, "token_type_ids": token_type_ids, } UpperCAmelCase__ = model(__UpperCAmelCase ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ (self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = config_and_inputs UpperCAmelCase__ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": input_mask} return config, inputs_dict @require_tf class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Tuple = ( ( TFConvBertModel, TFConvBertForMaskedLM, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertForMultipleChoice, ) if is_tf_available() else () ) __UpperCAmelCase : Optional[Any] = ( { 'feature-extraction': TFConvBertModel, 'fill-mask': TFConvBertForMaskedLM, 'question-answering': TFConvBertForQuestionAnswering, 'text-classification': TFConvBertForSequenceClassification, 'token-classification': TFConvBertForTokenClassification, 'zero-shot': TFConvBertForSequenceClassification, } if is_tf_available() else {} ) __UpperCAmelCase : List[Any] = False __UpperCAmelCase : Optional[Any] = False __UpperCAmelCase : str = False def lowercase_ (self : Tuple ) -> str: """simple docstring""" UpperCAmelCase__ = TFConvBertModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=3_7 ) def lowercase_ (self : str ) -> Dict: """simple docstring""" self.config_tester.run_common_tests() def lowercase_ (self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(*__UpperCAmelCase ) def lowercase_ (self : str ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_masked_lm(*__UpperCAmelCase ) def lowercase_ (self : int ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_multiple_choice(*__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> str: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering(*__UpperCAmelCase ) def lowercase_ (self : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification(*__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> str: """simple docstring""" UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification(*__UpperCAmelCase ) @slow def lowercase_ (self : Union[str, Any] ) -> Any: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = True UpperCAmelCase__ = True if hasattr(__UpperCAmelCase , "use_cache" ): UpperCAmelCase__ = True UpperCAmelCase__ = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) for model_class in self.all_model_classes: UpperCAmelCase__ = self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = len(model(__UpperCAmelCase ) ) with tempfile.TemporaryDirectory() as tmpdirname: model.save_pretrained(__UpperCAmelCase , saved_model=__UpperCAmelCase ) UpperCAmelCase__ = os.path.join(__UpperCAmelCase , "saved_model" , "1" ) UpperCAmelCase__ = tf.keras.models.load_model(__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase ) if self.is_encoder_decoder: UpperCAmelCase__ = outputs["encoder_hidden_states"] UpperCAmelCase__ = outputs["encoder_attentions"] else: UpperCAmelCase__ = outputs["hidden_states"] UpperCAmelCase__ = outputs["attentions"] self.assertEqual(len(__UpperCAmelCase ) , __UpperCAmelCase ) UpperCAmelCase__ = getattr( self.model_tester , "expected_num_hidden_layers" , self.model_tester.num_hidden_layers + 1 ) self.assertEqual(len(__UpperCAmelCase ) , __UpperCAmelCase ) self.assertListEqual( list(output_hidden_states[0].shape[-2:] ) , [self.model_tester.seq_length, self.model_tester.hidden_size] , ) self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(output_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) @slow def lowercase_ (self : Any ) -> Any: """simple docstring""" UpperCAmelCase__ = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) self.assertIsNotNone(__UpperCAmelCase ) def lowercase_ (self : int ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() UpperCAmelCase__ = True UpperCAmelCase__ = getattr(self.model_tester , "decoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "encoder_seq_length" , self.model_tester.seq_length ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) UpperCAmelCase__ = getattr(self.model_tester , "key_length" , __UpperCAmelCase ) def check_decoder_attentions_output(__UpperCAmelCase : List[str] ): UpperCAmelCase__ = len(__UpperCAmelCase ) self.assertEqual(out_len % 2 , 0 ) UpperCAmelCase__ = outputs.decoder_attentions self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(decoder_attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, decoder_seq_length, decoder_key_length] , ) def check_encoder_attentions_output(__UpperCAmelCase : List[str] ): UpperCAmelCase__ = [ t.numpy() for t in (outputs.encoder_attentions if config.is_encoder_decoder else outputs.attentions) ] self.assertEqual(len(__UpperCAmelCase ) , self.model_tester.num_hidden_layers ) self.assertListEqual( list(attentions[0].shape[-3:] ) , [self.model_tester.num_attention_heads / 2, encoder_seq_length, encoder_key_length] , ) for model_class in self.all_model_classes: UpperCAmelCase__ = True UpperCAmelCase__ = False UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) UpperCAmelCase__ = len(__UpperCAmelCase ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) if self.is_encoder_decoder: UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_decoder_attentions_output(__UpperCAmelCase ) # Check that output attentions can also be changed via the config del inputs_dict["output_attentions"] UpperCAmelCase__ = True UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) # Check attention is always last and order is fine UpperCAmelCase__ = True UpperCAmelCase__ = True UpperCAmelCase__ = model_class(__UpperCAmelCase ) UpperCAmelCase__ = model(self._prepare_for_class(__UpperCAmelCase , __UpperCAmelCase ) ) self.assertEqual(out_len + (2 if self.is_encoder_decoder else 1) , len(__UpperCAmelCase ) ) self.assertEqual(model.config.output_hidden_states , __UpperCAmelCase ) check_encoder_attentions_output(__UpperCAmelCase ) @require_tf class A ( unittest.TestCase ): @slow def lowercase_ (self : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = TFConvBertModel.from_pretrained("YituTech/conv-bert-base" ) UpperCAmelCase__ = tf.constant([[0, 1, 2, 3, 4, 5]] ) UpperCAmelCase__ = model(__UpperCAmelCase )[0] UpperCAmelCase__ = [1, 6, 7_6_8] self.assertEqual(output.shape , __UpperCAmelCase ) UpperCAmelCase__ = tf.constant( [ [ [-0.03475493, -0.4686034, -0.30638832], [0.22637248, -0.26988646, -0.7423424], [0.10324868, -0.45013508, -0.58280784], ] ] ) tf.debugging.assert_near(output[:, :3, :3] , __UpperCAmelCase , atol=1E-4 )
65
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path", type=__A, default="data/dump.txt", help="The path to the data." ) parser.add_argument("--tokenizer_type", type=__A, default="bert", choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name", type=__A, default="bert-base-uncased", help="The tokenizer to use." ) parser.add_argument("--dump_file", type=__A, default="data/dump", help="The dump file prefix." ) UpperCAmelCase__ = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": UpperCAmelCase__ = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `[CLS]` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCAmelCase__ = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `<s>` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": UpperCAmelCase__ = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` UpperCAmelCase__ = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path, "r", encoding="utf8" ) as fp: UpperCAmelCase__ = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__A )} examples to process.""" ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 10_000 UpperCAmelCase__ = time.time() for text in data: UpperCAmelCase__ = f"""{bos} {text.strip()} {sep}""" UpperCAmelCase__ = tokenizer.encode(__A, add_special_tokens=__A ) rslt.append(__A ) iter += 1 if iter % interval == 0: UpperCAmelCase__ = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) UpperCAmelCase__ = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__A )} examples processed.""" ) UpperCAmelCase__ = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" UpperCAmelCase__ = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCAmelCase__ = [np.uintaa(__A ) for d in rslt] else: UpperCAmelCase__ = [np.intaa(__A ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__A, "wb" ) as handle: pickle.dump(rslt_, __A, protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
65
1
# Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import numpy as np import torch from ..models.clipseg import CLIPSegForImageSegmentation from ..utils import is_vision_available, requires_backends from .base import PipelineTool if is_vision_available(): from PIL import Image class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[str] = ( 'This is a tool that creates a segmentation mask of an image according to a label. It cannot create an image.' 'It takes two arguments named `image` which should be the original image, and `label` which should be a text ' 'describing the elements what should be identified in the segmentation mask. The tool returns the mask.' ) __UpperCAmelCase : Tuple = 'CIDAS/clipseg-rd64-refined' __UpperCAmelCase : str = 'image_segmenter' __UpperCAmelCase : Any = CLIPSegForImageSegmentation __UpperCAmelCase : str = ['image', 'text'] __UpperCAmelCase : Tuple = ['image'] def __init__(self : Union[str, Any] , *__UpperCAmelCase : int , **__UpperCAmelCase : Dict ) -> Optional[int]: """simple docstring""" requires_backends(self , ["vision"] ) super().__init__(*__UpperCAmelCase , **__UpperCAmelCase ) def lowercase_ (self : List[str] , __UpperCAmelCase : "Image" , __UpperCAmelCase : str ) -> Tuple: """simple docstring""" return self.pre_processor(text=[label] , images=[image] , padding=__UpperCAmelCase , return_tensors="pt" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : str ) -> int: """simple docstring""" with torch.no_grad(): UpperCAmelCase__ = self.model(**__UpperCAmelCase ).logits return logits def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = outputs.cpu().detach().numpy() UpperCAmelCase__ = 0 UpperCAmelCase__ = 1 return Image.fromarray((array * 2_5_5).astype(np.uinta ) )
65
from manim import * class A ( UpperCAmelCase_ ): def lowercase_ (self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase__ = Rectangle(height=0.25 , width=0.25 ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("CPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(4 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("GPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) gpu.move_to([-1, -1, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Model" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) model.move_to([3, -1.0, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = fill.copy().set_fill(__UpperCAmelCase , opacity=0.8 ) target.move_to(__UpperCAmelCase ) model_arr.append(__UpperCAmelCase ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(__UpperCAmelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__UpperCAmelCase ) self.add(*__UpperCAmelCase , *__UpperCAmelCase ) UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Disk" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) disk.move_to([-4, -1.25, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(__UpperCAmelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase ) ) UpperCAmelCase__ = Square(0.3 ) input.set_fill(__UpperCAmelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , __UpperCAmelCase , buff=0.5 ) self.play(Write(__UpperCAmelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=__UpperCAmelCase , buff=0.02 ) self.play(MoveToTarget(__UpperCAmelCase ) ) self.play(FadeOut(__UpperCAmelCase ) ) UpperCAmelCase__ = Arrow(start=__UpperCAmelCase , end=__UpperCAmelCase , color=__UpperCAmelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , __UpperCAmelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) UpperCAmelCase__ = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) ) UpperCAmelCase__ = {"run_time": 1, "fade_in": True, "fade_out": True, "buff": 0.02} self.play( Write(__UpperCAmelCase ) , Circumscribe(model_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_cpu_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) UpperCAmelCase__ = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , __UpperCAmelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) UpperCAmelCase__ = AnimationGroup( FadeOut(__UpperCAmelCase , run_time=0.5 ) , MoveToTarget(__UpperCAmelCase , run_time=0.5 ) , FadeIn(__UpperCAmelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(__UpperCAmelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: UpperCAmelCase__ = 0.7 self.play( Circumscribe(model_arr[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_arr[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) UpperCAmelCase__ = a_c UpperCAmelCase__ = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(__UpperCAmelCase ) , FadeOut(__UpperCAmelCase , run_time=0.5 ) , ) UpperCAmelCase__ = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) , MoveToTarget(__UpperCAmelCase ) ) self.wait()
65
1
from collections import OrderedDict from typing import Mapping from ...configuration_utils import PretrainedConfig from ...onnx import OnnxConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'bert-base-uncased': 'https://huggingface.co/bert-base-uncased/resolve/main/config.json', 'bert-large-uncased': 'https://huggingface.co/bert-large-uncased/resolve/main/config.json', 'bert-base-cased': 'https://huggingface.co/bert-base-cased/resolve/main/config.json', 'bert-large-cased': 'https://huggingface.co/bert-large-cased/resolve/main/config.json', 'bert-base-multilingual-uncased': 'https://huggingface.co/bert-base-multilingual-uncased/resolve/main/config.json', 'bert-base-multilingual-cased': 'https://huggingface.co/bert-base-multilingual-cased/resolve/main/config.json', 'bert-base-chinese': 'https://huggingface.co/bert-base-chinese/resolve/main/config.json', 'bert-base-german-cased': 'https://huggingface.co/bert-base-german-cased/resolve/main/config.json', 'bert-large-uncased-whole-word-masking': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking': ( 'https://huggingface.co/bert-large-cased-whole-word-masking/resolve/main/config.json' ), 'bert-large-uncased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-uncased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-large-cased-whole-word-masking-finetuned-squad': ( 'https://huggingface.co/bert-large-cased-whole-word-masking-finetuned-squad/resolve/main/config.json' ), 'bert-base-cased-finetuned-mrpc': 'https://huggingface.co/bert-base-cased-finetuned-mrpc/resolve/main/config.json', 'bert-base-german-dbmdz-cased': 'https://huggingface.co/bert-base-german-dbmdz-cased/resolve/main/config.json', 'bert-base-german-dbmdz-uncased': 'https://huggingface.co/bert-base-german-dbmdz-uncased/resolve/main/config.json', 'cl-tohoku/bert-base-japanese': 'https://huggingface.co/cl-tohoku/bert-base-japanese/resolve/main/config.json', 'cl-tohoku/bert-base-japanese-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-whole-word-masking/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char/resolve/main/config.json' ), 'cl-tohoku/bert-base-japanese-char-whole-word-masking': ( 'https://huggingface.co/cl-tohoku/bert-base-japanese-char-whole-word-masking/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-cased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-cased-v1/resolve/main/config.json' ), 'TurkuNLP/bert-base-finnish-uncased-v1': ( 'https://huggingface.co/TurkuNLP/bert-base-finnish-uncased-v1/resolve/main/config.json' ), 'wietsedv/bert-base-dutch-cased': 'https://huggingface.co/wietsedv/bert-base-dutch-cased/resolve/main/config.json', # See all BERT models at https://huggingface.co/models?filter=bert } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = 'bert' def __init__(self : List[str] , __UpperCAmelCase : Tuple=3_0_5_2_2 , __UpperCAmelCase : List[Any]=7_6_8 , __UpperCAmelCase : List[Any]=1_2 , __UpperCAmelCase : Any=1_2 , __UpperCAmelCase : Any=3_0_7_2 , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : List[str]=0.1 , __UpperCAmelCase : List[str]=5_1_2 , __UpperCAmelCase : List[str]=2 , __UpperCAmelCase : List[str]=0.02 , __UpperCAmelCase : Union[str, Any]=1E-12 , __UpperCAmelCase : List[str]=0 , __UpperCAmelCase : List[Any]="absolute" , __UpperCAmelCase : int=True , __UpperCAmelCase : List[Any]=None , **__UpperCAmelCase : Tuple , ) -> Union[str, Any]: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_act UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = position_embedding_type UpperCAmelCase__ = use_cache UpperCAmelCase__ = classifier_dropout class A ( UpperCAmelCase_ ): @property def lowercase_ (self : List[Any] ) -> Mapping[str, Mapping[int, str]]: """simple docstring""" if self.task == "multiple-choice": UpperCAmelCase__ = {0: "batch", 1: "choice", 2: "sequence"} else: UpperCAmelCase__ = {0: "batch", 1: "sequence"} return OrderedDict( [ ("input_ids", dynamic_axis), ("attention_mask", dynamic_axis), ("token_type_ids", dynamic_axis), ] )
65
from __future__ import annotations from scipy.special import comb # type: ignore class A : def __init__(self : List[Any] , __UpperCAmelCase : list[tuple[float, float]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. UpperCAmelCase__ = len(__UpperCAmelCase ) - 1 def lowercase_ (self : int , __UpperCAmelCase : float ) -> list[float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __UpperCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__UpperCAmelCase ) , 5 ) == 1 return output_values def lowercase_ (self : Dict , __UpperCAmelCase : float ) -> tuple[float, float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = self.basis_function(__UpperCAmelCase ) UpperCAmelCase__ = 0.0 UpperCAmelCase__ = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def lowercase_ (self : Optional[int] , __UpperCAmelCase : float = 0.01 ) -> Optional[int]: """simple docstring""" from matplotlib import pyplot as plt # type: ignore UpperCAmelCase__ = [] # x coordinates of points to plot UpperCAmelCase__ = [] # y coordinates of points to plot UpperCAmelCase__ = 0.0 while t <= 1: UpperCAmelCase__ = self.bezier_curve_function(__UpperCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size UpperCAmelCase__ = [i[0] for i in self.list_of_points] UpperCAmelCase__ = [i[1] for i in self.list_of_points] plt.plot( __UpperCAmelCase , __UpperCAmelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__UpperCAmelCase , __UpperCAmelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
65
1
import pytest from datasets import inspect_metric, list_metrics, load_metric @pytest.fixture def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' monkeypatch.setattr("datasets.utils.deprecation_utils._emitted_deprecation_warnings", set() ) @pytest.fixture def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' class A : def __init__(self : Dict , __UpperCAmelCase : List[Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = metric_id class A : __UpperCAmelCase : Union[str, Any] = [MetricMock(UpperCAmelCase_ ) for metric_id in ['accuracy', 'mse', 'precision', 'codeparrot/apps_metric']] def lowercase_ (self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" return self._metrics monkeypatch.setattr("datasets.inspect.huggingface_hub", HfhMock() ) @pytest.mark.parametrize( "func, args", [(load_metric, ("metrics/mse",)), (list_metrics, ()), (inspect_metric, ("metrics/mse", "tmp_path"))] ) def lowerCAmelCase_ ( __A, __A, __A, __A, __A ) -> Optional[int]: '''simple docstring''' if "tmp_path" in args: UpperCAmelCase__ = tuple(arg if arg != "tmp_path" else tmp_path for arg in args ) with pytest.warns(__A, match="https://huggingface.co/docs/evaluate" ): func(*__A )
65
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
1
import functools import operator from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'microsoft/wavlm-base': 'https://huggingface.co/microsoft/wavlm-base/resolve/main/config.json', # See all WavLM models at https://huggingface.co/models?filter=wavlm } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = 'wavlm' def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3_2 , __UpperCAmelCase : int=7_6_8 , __UpperCAmelCase : Dict=1_2 , __UpperCAmelCase : List[Any]=1_2 , __UpperCAmelCase : int=3_0_7_2 , __UpperCAmelCase : List[str]="gelu" , __UpperCAmelCase : Optional[Any]=0.1 , __UpperCAmelCase : Dict=0.1 , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : Any=0.0 , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : Dict=0.02 , __UpperCAmelCase : Optional[Any]=1E-5 , __UpperCAmelCase : List[Any]="group" , __UpperCAmelCase : Dict="gelu" , __UpperCAmelCase : Tuple=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2, 5_1_2) , __UpperCAmelCase : Any=(5, 2, 2, 2, 2, 2, 2) , __UpperCAmelCase : int=(1_0, 3, 3, 3, 3, 2, 2) , __UpperCAmelCase : Any=False , __UpperCAmelCase : Optional[int]=1_2_8 , __UpperCAmelCase : Optional[int]=1_6 , __UpperCAmelCase : List[str]=3_2_0 , __UpperCAmelCase : List[str]=8_0_0 , __UpperCAmelCase : int=False , __UpperCAmelCase : Optional[int]=True , __UpperCAmelCase : int=0.05 , __UpperCAmelCase : List[Any]=1_0 , __UpperCAmelCase : Optional[int]=2 , __UpperCAmelCase : List[str]=0.0 , __UpperCAmelCase : Optional[Any]=1_0 , __UpperCAmelCase : List[str]=3_2_0 , __UpperCAmelCase : List[str]=2 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : Union[str, Any]=1_0_0 , __UpperCAmelCase : int=2_5_6 , __UpperCAmelCase : Union[str, Any]=2_5_6 , __UpperCAmelCase : List[str]=0.1 , __UpperCAmelCase : str="mean" , __UpperCAmelCase : Union[str, Any]=False , __UpperCAmelCase : str=False , __UpperCAmelCase : Optional[int]=2_5_6 , __UpperCAmelCase : Tuple=(5_1_2, 5_1_2, 5_1_2, 5_1_2, 1_5_0_0) , __UpperCAmelCase : List[Any]=(5, 3, 3, 1, 1) , __UpperCAmelCase : List[Any]=(1, 2, 3, 1, 1) , __UpperCAmelCase : Optional[Any]=5_1_2 , __UpperCAmelCase : Optional[int]=8_0 , __UpperCAmelCase : int=0 , __UpperCAmelCase : Optional[int]=1 , __UpperCAmelCase : Tuple=2 , __UpperCAmelCase : Tuple=False , __UpperCAmelCase : List[str]=3 , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Optional[Any]=3 , __UpperCAmelCase : Optional[int]=None , **__UpperCAmelCase : Optional[int] , ) -> List[Any]: """simple docstring""" super().__init__(**__UpperCAmelCase , pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase ) UpperCAmelCase__ = hidden_size UpperCAmelCase__ = feat_extract_norm UpperCAmelCase__ = feat_extract_activation UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = conv_bias UpperCAmelCase__ = num_buckets UpperCAmelCase__ = max_bucket_distance UpperCAmelCase__ = num_conv_pos_embeddings UpperCAmelCase__ = num_conv_pos_embedding_groups UpperCAmelCase__ = len(self.conv_dim ) UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_dropout UpperCAmelCase__ = attention_dropout UpperCAmelCase__ = activation_dropout UpperCAmelCase__ = feat_proj_dropout UpperCAmelCase__ = final_dropout UpperCAmelCase__ = layerdrop UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_ctc_classes UpperCAmelCase__ = vocab_size UpperCAmelCase__ = do_stable_layer_norm UpperCAmelCase__ = use_weighted_layer_sum UpperCAmelCase__ = classifier_proj_size if ( (len(self.conv_stride ) != self.num_feat_extract_layers) or (len(self.conv_kernel ) != self.num_feat_extract_layers) or (len(self.conv_dim ) != self.num_feat_extract_layers) ): raise ValueError( "Configuration for convolutional layers is incorrect. It is required that `len(config.conv_dim)` ==" " `len(config.conv_stride)` == `len(config.conv_kernel)`, but is `len(config.conv_dim) =" f""" {len(self.conv_dim )}`, `len(config.conv_stride) = {len(self.conv_stride )}`,""" f""" `len(config.conv_kernel) = {len(self.conv_kernel )}`.""" ) # fine-tuning config parameters for SpecAugment: https://arxiv.org/abs/1904.08779 UpperCAmelCase__ = apply_spec_augment UpperCAmelCase__ = mask_time_prob UpperCAmelCase__ = mask_time_length UpperCAmelCase__ = mask_time_min_masks UpperCAmelCase__ = mask_feature_prob UpperCAmelCase__ = mask_feature_length # parameters for pretraining with codevector quantized representations UpperCAmelCase__ = num_codevectors_per_group UpperCAmelCase__ = num_codevector_groups UpperCAmelCase__ = contrastive_logits_temperature UpperCAmelCase__ = num_negatives UpperCAmelCase__ = codevector_dim UpperCAmelCase__ = proj_codevector_dim UpperCAmelCase__ = diversity_loss_weight # ctc loss UpperCAmelCase__ = ctc_loss_reduction UpperCAmelCase__ = ctc_zero_infinity # adapter UpperCAmelCase__ = add_adapter UpperCAmelCase__ = adapter_kernel_size UpperCAmelCase__ = adapter_stride UpperCAmelCase__ = num_adapter_layers UpperCAmelCase__ = output_hidden_size or hidden_size # SequenceClassification-specific parameter. Feel free to ignore for other classes. UpperCAmelCase__ = classifier_proj_size # XVector-specific parameters. Feel free to ignore for other classes. UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = list(__UpperCAmelCase ) UpperCAmelCase__ = xvector_output_dim @property def lowercase_ (self : Tuple ) -> Tuple: """simple docstring""" return functools.reduce(operator.mul , self.conv_stride , 1 )
65
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
1
import unittest import numpy as np from transformers import AlbertConfig, is_flax_available from transformers.testing_utils import require_flax, slow from ...test_modeling_flax_common import FlaxModelTesterMixin, ids_tensor, random_attention_mask if is_flax_available(): import jax.numpy as jnp from transformers.models.albert.modeling_flax_albert import ( FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForPreTraining, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertModel, ) class A ( unittest.TestCase ): def __init__(self : int , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : str=1_3 , __UpperCAmelCase : Dict=7 , __UpperCAmelCase : str=True , __UpperCAmelCase : int=True , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : int=True , __UpperCAmelCase : Tuple=9_9 , __UpperCAmelCase : int=3_2 , __UpperCAmelCase : List[Any]=5 , __UpperCAmelCase : List[str]=4 , __UpperCAmelCase : Tuple=3_7 , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : Tuple=0.1 , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : List[str]=5_1_2 , __UpperCAmelCase : Optional[Any]=1_6 , __UpperCAmelCase : int=2 , __UpperCAmelCase : List[str]=0.02 , __UpperCAmelCase : int=4 , ) -> Any: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = seq_length UpperCAmelCase__ = is_training UpperCAmelCase__ = use_attention_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = num_choices def lowercase_ (self : str ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.vocab_size ) UpperCAmelCase__ = None if self.use_attention_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.seq_length] , self.type_vocab_size ) UpperCAmelCase__ = AlbertConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , is_decoder=__UpperCAmelCase , initializer_range=self.initializer_range , ) return config, input_ids, token_type_ids, attention_mask def lowercase_ (self : Tuple ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = config_and_inputs UpperCAmelCase__ = {"input_ids": input_ids, "token_type_ids": token_type_ids, "attention_mask": attention_mask} return config, inputs_dict @require_flax class A ( UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Tuple = ( ( FlaxAlbertModel, FlaxAlbertForPreTraining, FlaxAlbertForMaskedLM, FlaxAlbertForMultipleChoice, FlaxAlbertForQuestionAnswering, FlaxAlbertForSequenceClassification, FlaxAlbertForTokenClassification, FlaxAlbertForQuestionAnswering, ) if is_flax_available() else () ) def lowercase_ (self : Tuple ) -> Any: """simple docstring""" UpperCAmelCase__ = FlaxAlbertModelTester(self ) @slow def lowercase_ (self : Optional[Any] ) -> List[str]: """simple docstring""" for model_class_name in self.all_model_classes: UpperCAmelCase__ = model_class_name.from_pretrained("albert-base-v2" ) UpperCAmelCase__ = model(np.ones((1, 1) ) ) self.assertIsNotNone(__UpperCAmelCase ) @require_flax class A ( unittest.TestCase ): @slow def lowercase_ (self : int ) -> str: """simple docstring""" UpperCAmelCase__ = FlaxAlbertModel.from_pretrained("albert-base-v2" ) UpperCAmelCase__ = np.array([[0, 3_4_5, 2_3_2, 3_2_8, 7_4_0, 1_4_0, 1_6_9_5, 6_9, 6_0_7_8, 1_5_8_8, 2]] ) UpperCAmelCase__ = np.array([[0, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1]] ) UpperCAmelCase__ = model(__UpperCAmelCase , attention_mask=__UpperCAmelCase )[0] UpperCAmelCase__ = (1, 1_1, 7_6_8) self.assertEqual(output.shape , __UpperCAmelCase ) UpperCAmelCase__ = np.array( [[[-0.6513, 1.5035, -0.2766], [-0.6515, 1.5046, -0.2780], [-0.6512, 1.5049, -0.2784]]] ) self.assertTrue(jnp.allclose(output[:, 1:4, 1:4] , __UpperCAmelCase , atol=1E-4 ) )
65
from __future__ import annotations from collections import deque class A : def __init__(self : Dict , __UpperCAmelCase : list[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__UpperCAmelCase ) self.set_fail_transitions() def lowercase_ (self : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : str ) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = 0 for character in keyword: UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , __UpperCAmelCase ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase__ = len(self.adlist ) - 1 else: UpperCAmelCase__ = next_state self.adlist[current_state]["output"].append(__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = deque() for node in self.adlist[0]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = 0 while q: UpperCAmelCase__ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = self.adlist[r]["fail_state"] while ( self.find_next_state(__UpperCAmelCase , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase__ = self.adlist[state]["fail_state"] UpperCAmelCase__ = self.find_next_state( __UpperCAmelCase , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase__ = 0 UpperCAmelCase__ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> dict[str, list[int]]: """simple docstring""" UpperCAmelCase__ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase__ = 0 for i in range(len(__UpperCAmelCase ) ): while ( self.find_next_state(__UpperCAmelCase , string[i] ) is None and current_state != 0 ): UpperCAmelCase__ = self.adlist[current_state]["fail_state"] UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , string[i] ) if next_state is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase__ = [] result[key].append(i - len(__UpperCAmelCase ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
65
1
import fire from torch.utils.data import DataLoader from tqdm import tqdm from transformers import AutoTokenizer from utils import SeqaSeqDataset, pickle_save def lowerCAmelCase_ ( __A, __A, __A=1_024, __A=1_024, __A=False, **__A ) -> int: '''simple docstring''' UpperCAmelCase__ = AutoTokenizer.from_pretrained(__A ) UpperCAmelCase__ = SeqaSeqDataset(__A, __A, __A, __A, type_path="train", **__A ) UpperCAmelCase__ = tok.pad_token_id def get_lens(__A ): UpperCAmelCase__ = tqdm( DataLoader(__A, batch_size=512, num_workers=8, shuffle=__A, collate_fn=ds.collate_fn ), desc=str(ds.len_file ), ) UpperCAmelCase__ = [] for batch in dl: UpperCAmelCase__ = batch["input_ids"].ne(__A ).sum(1 ).tolist() UpperCAmelCase__ = batch["labels"].ne(__A ).sum(1 ).tolist() if consider_target: for src, tgt in zip(__A, __A ): max_lens.append(max(__A, __A ) ) else: max_lens.extend(__A ) return max_lens UpperCAmelCase__ = get_lens(__A ) UpperCAmelCase__ = SeqaSeqDataset(__A, __A, __A, __A, type_path="val", **__A ) UpperCAmelCase__ = get_lens(__A ) pickle_save(__A, train_ds.len_file ) pickle_save(__A, val_ds.len_file ) if __name__ == "__main__": fire.Fire(save_len_file)
65
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ['input_values', 'attention_mask'] def __init__(self : Any , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 1_6_0_0_0 , __UpperCAmelCase : float = 0.0 , __UpperCAmelCase : bool = False , __UpperCAmelCase : int = 8_0 , __UpperCAmelCase : int = 1_6 , __UpperCAmelCase : int = 6_4 , __UpperCAmelCase : str = "hann_window" , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : float = 8_0 , __UpperCAmelCase : float = 7_6_0_0 , __UpperCAmelCase : float = 1E-10 , __UpperCAmelCase : int = 2 , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Any , ) -> str: """simple docstring""" super().__init__(feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = do_normalize UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = frame_signal_scale UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = reduction_factor UpperCAmelCase__ = win_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = hop_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = optimal_fft_length(self.sample_size ) UpperCAmelCase__ = (self.n_fft // 2) + 1 UpperCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCAmelCase ) UpperCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowercase_ (__UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase , np.intaa ) UpperCAmelCase__ = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): UpperCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: UpperCAmelCase__ = padding_value normed_input_values.append(__UpperCAmelCase ) else: UpperCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowercase_ (self : Optional[int] , __UpperCAmelCase : np.ndarray , ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ = spectrogram( __UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__(self : Any , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[int] = None , **__UpperCAmelCase : str , ) -> BatchFeature: """simple docstring""" if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) else: UpperCAmelCase__ = None if audio_target is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) if inputs is None: return inputs_target else: UpperCAmelCase__ = inputs_target["input_values"] UpperCAmelCase__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: UpperCAmelCase__ = decoder_attention_mask return inputs def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCAmelCase : bool = False , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : Any , ) -> BatchFeature: """simple docstring""" UpperCAmelCase__ = isinstance(__UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): UpperCAmelCase__ = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs UpperCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: UpperCAmelCase__ = [self._extract_mel_features(__UpperCAmelCase ) for waveform in speech] UpperCAmelCase__ = BatchFeature({"input_values": features} ) UpperCAmelCase__ = self.num_mel_bins else: UpperCAmelCase__ = BatchFeature({"input_values": speech} ) UpperCAmelCase__ = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = feature_size_hack # convert input values to correct format UpperCAmelCase__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(__UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): UpperCAmelCase__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(__UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = input_values.astype(np.floataa ) # convert attention_mask to correct format UpperCAmelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: UpperCAmelCase__ = ( attention_mask if self._get_padding_strategies(__UpperCAmelCase , max_length=__UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=__UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def lowercase_ (self : Tuple ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. UpperCAmelCase__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
65
1
import json from typing import List, Optional, Tuple from tokenizers import normalizers from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import logging from .tokenization_mobilebert import MobileBertTokenizer UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'vocab.txt', 'tokenizer_file': 'tokenizer.json'} UpperCamelCase__ = { 'vocab_file': {'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/vocab.txt'}, 'tokenizer_file': { 'mobilebert-uncased': 'https://huggingface.co/google/mobilebert-uncased/resolve/main/tokenizer.json' }, } UpperCamelCase__ = {'mobilebert-uncased': 5_1_2} UpperCamelCase__ = {} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Any = PRETRAINED_INIT_CONFIGURATION __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = MobileBertTokenizer def __init__(self : List[str] , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : str=None , __UpperCAmelCase : str=True , __UpperCAmelCase : Optional[int]="[UNK]" , __UpperCAmelCase : List[str]="[SEP]" , __UpperCAmelCase : Tuple="[PAD]" , __UpperCAmelCase : int="[CLS]" , __UpperCAmelCase : Tuple="[MASK]" , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : str=None , **__UpperCAmelCase : Tuple , ) -> Tuple: """simple docstring""" super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , tokenize_chinese_chars=__UpperCAmelCase , strip_accents=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = json.loads(self.backend_tokenizer.normalizer.__getstate__() ) if ( normalizer_state.get("lowercase" , __UpperCAmelCase ) != do_lower_case or normalizer_state.get("strip_accents" , __UpperCAmelCase ) != strip_accents or normalizer_state.get("handle_chinese_chars" , __UpperCAmelCase ) != tokenize_chinese_chars ): UpperCAmelCase__ = getattr(__UpperCAmelCase , normalizer_state.pop("type" ) ) UpperCAmelCase__ = do_lower_case UpperCAmelCase__ = strip_accents UpperCAmelCase__ = tokenize_chinese_chars UpperCAmelCase__ = normalizer_class(**__UpperCAmelCase ) UpperCAmelCase__ = do_lower_case def lowercase_ (self : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any]=None ) -> List[str]: """simple docstring""" UpperCAmelCase__ = [self.cls_token_id] + token_ids_a + [self.sep_token_id] if token_ids_a: output += token_ids_a + [self.sep_token_id] return output def lowercase_ (self : List[str] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] def lowercase_ (self : str , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" UpperCAmelCase__ = self._tokenizer.model.save(__UpperCAmelCase , name=__UpperCAmelCase ) return tuple(__UpperCAmelCase )
65
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[torch.FloatTensor] = None __UpperCAmelCase : torch.FloatTensor = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Union[str, Any]=5_1_2 , __UpperCAmelCase : List[str]="cls" , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=True , **__UpperCAmelCase : str , ) -> int: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = project_dim UpperCAmelCase__ = pooler_fn UpperCAmelCase__ = learn_encoder UpperCAmelCase__ = use_attention_mask class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = [r'pooler', r'logit_scale'] __UpperCAmelCase : int = [r'position_ids', r'predictions.decoder.bias'] __UpperCAmelCase : Any = 'roberta' __UpperCAmelCase : List[str] = RobertaSeriesConfig def __init__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = XLMRobertaModel(__UpperCAmelCase ) UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = getattr(__UpperCAmelCase , "has_pre_transformation" , __UpperCAmelCase ) if self.has_pre_transformation: UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.base_model( input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , position_ids=__UpperCAmelCase , head_mask=__UpperCAmelCase , inputs_embeds=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__UpperCAmelCase , ) if self.has_pre_transformation: UpperCAmelCase__ = outputs["hidden_states"][-2] UpperCAmelCase__ = self.pre_LN(__UpperCAmelCase ) UpperCAmelCase__ = self.transformation_pre(__UpperCAmelCase ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: UpperCAmelCase__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
65
1
import unittest import numpy as np from transformers.file_utils import is_torch_available, is_vision_available from transformers.testing_utils import require_torch, require_vision from ...test_image_processing_common import ImageProcessingSavingTestMixin, prepare_image_inputs if is_torch_available(): import torch if is_vision_available(): from PIL import Image from transformers import DPTImageProcessor class A ( unittest.TestCase ): def __init__(self : int , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[Any]=7 , __UpperCAmelCase : str=3 , __UpperCAmelCase : Tuple=1_8 , __UpperCAmelCase : int=3_0 , __UpperCAmelCase : str=4_0_0 , __UpperCAmelCase : str=True , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Tuple=True , __UpperCAmelCase : Any=[0.5, 0.5, 0.5] , __UpperCAmelCase : str=[0.5, 0.5, 0.5] , ) -> Tuple: """simple docstring""" UpperCAmelCase__ = size if size is not None else {"height": 1_8, "width": 1_8} UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = min_resolution UpperCAmelCase__ = max_resolution UpperCAmelCase__ = do_resize UpperCAmelCase__ = size UpperCAmelCase__ = do_normalize UpperCAmelCase__ = image_mean UpperCAmelCase__ = image_std def lowercase_ (self : Optional[Any] ) -> int: """simple docstring""" return { "image_mean": self.image_mean, "image_std": self.image_std, "do_normalize": self.do_normalize, "do_resize": self.do_resize, "size": self.size, } @require_torch @require_vision class A ( UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Dict = DPTImageProcessor if is_vision_available() else None def lowercase_ (self : List[Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = DPTImageProcessingTester(self ) @property def lowercase_ (self : str ) -> Union[str, Any]: """simple docstring""" return self.image_processor_tester.prepare_image_processor_dict() def lowercase_ (self : str ) -> int: """simple docstring""" UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) self.assertTrue(hasattr(__UpperCAmelCase , "image_mean" ) ) self.assertTrue(hasattr(__UpperCAmelCase , "image_std" ) ) self.assertTrue(hasattr(__UpperCAmelCase , "do_normalize" ) ) self.assertTrue(hasattr(__UpperCAmelCase , "do_resize" ) ) self.assertTrue(hasattr(__UpperCAmelCase , "size" ) ) def lowercase_ (self : List[str] ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict ) self.assertEqual(image_processor.size , {"height": 1_8, "width": 1_8} ) UpperCAmelCase__ = self.image_processing_class.from_dict(self.image_processor_dict , size=4_2 ) self.assertEqual(image_processor.size , {"height": 4_2, "width": 4_2} ) def lowercase_ (self : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PIL images UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase , Image.Image ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched UpperCAmelCase__ = image_processing(__UpperCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def lowercase_ (self : str ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random numpy tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCAmelCase , numpify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase , np.ndarray ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched UpperCAmelCase__ = image_processing(__UpperCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) def lowercase_ (self : Any ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.image_processing_class(**self.image_processor_dict ) # create random PyTorch tensors UpperCAmelCase__ = prepare_image_inputs(self.image_processor_tester , equal_resolution=__UpperCAmelCase , torchify=__UpperCAmelCase ) for image in image_inputs: self.assertIsInstance(__UpperCAmelCase , torch.Tensor ) # Test not batched input UpperCAmelCase__ = image_processing(image_inputs[0] , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( 1, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , ) # Test batched UpperCAmelCase__ = image_processing(__UpperCAmelCase , return_tensors="pt" ).pixel_values self.assertEqual( encoded_images.shape , ( self.image_processor_tester.batch_size, self.image_processor_tester.num_channels, self.image_processor_tester.size["height"], self.image_processor_tester.size["width"], ) , )
65
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
1
import inspect import re from hashlib import shaaaa from typing import Dict, List from .arrow import arrow from .audiofolder import audiofolder from .csv import csv from .imagefolder import imagefolder from .json import json from .pandas import pandas from .parquet import parquet from .sql import sql # noqa F401 from .text import text def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = [] for line in lines: UpperCAmelCase__ = re.sub(r"#.*", "", __A ) # remove comments if line: filtered_lines.append(__A ) UpperCAmelCase__ = "\n".join(__A ) # Make a hash from all this code UpperCAmelCase__ = full_str.encode("utf-8" ) return shaaaa(__A ).hexdigest() # get importable module names and hash for caching UpperCamelCase__ = { 'csv': (csv.__name__, _hash_python_lines(inspect.getsource(csv).splitlines())), 'json': (json.__name__, _hash_python_lines(inspect.getsource(json).splitlines())), 'pandas': (pandas.__name__, _hash_python_lines(inspect.getsource(pandas).splitlines())), 'parquet': (parquet.__name__, _hash_python_lines(inspect.getsource(parquet).splitlines())), 'arrow': (arrow.__name__, _hash_python_lines(inspect.getsource(arrow).splitlines())), 'text': (text.__name__, _hash_python_lines(inspect.getsource(text).splitlines())), 'imagefolder': (imagefolder.__name__, _hash_python_lines(inspect.getsource(imagefolder).splitlines())), 'audiofolder': (audiofolder.__name__, _hash_python_lines(inspect.getsource(audiofolder).splitlines())), } # Used to infer the module to use based on the data files extensions UpperCamelCase__ = { '.csv': ('csv', {}), '.tsv': ('csv', {'sep': '\t'}), '.json': ('json', {}), '.jsonl': ('json', {}), '.parquet': ('parquet', {}), '.arrow': ('arrow', {}), '.txt': ('text', {}), } _EXTENSION_TO_MODULE.update({ext: ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('imagefolder', {}) for ext in imagefolder.ImageFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext: ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) _EXTENSION_TO_MODULE.update({ext.upper(): ('audiofolder', {}) for ext in audiofolder.AudioFolder.EXTENSIONS}) UpperCamelCase__ = {'imagefolder', 'audiofolder'} # Used to filter data files based on extensions given a module name UpperCamelCase__ = {} for _ext, (_module, _) in _EXTENSION_TO_MODULE.items(): _MODULE_TO_EXTENSIONS.setdefault(_module, []).append(_ext) _MODULE_TO_EXTENSIONS["imagefolder"].append('.zip') _MODULE_TO_EXTENSIONS["audiofolder"].append('.zip')
65
import math import random def lowerCAmelCase_ ( __A, __A = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase__ = 0.0_2 def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' UpperCAmelCase__ = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__A ): # Forward propagation UpperCAmelCase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase__ = (expected / 100) - layer_a # Error delta UpperCAmelCase__ = layer_1_error * sigmoid_function(__A, __A ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input('Expected value: ')) UpperCamelCase__ = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
65
1
from argparse import ArgumentParser from .add_new_model import AddNewModelCommand from .add_new_model_like import AddNewModelLikeCommand from .convert import ConvertCommand from .download import DownloadCommand from .env import EnvironmentCommand from .lfs import LfsCommands from .pt_to_tf import PTtoTFCommand from .run import RunCommand from .serving import ServeCommand from .user import UserCommands def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' UpperCAmelCase__ = ArgumentParser("Transformers CLI tool", usage="transformers-cli <command> [<args>]" ) UpperCAmelCase__ = parser.add_subparsers(help="transformers-cli command helpers" ) # Register commands ConvertCommand.register_subcommand(__A ) DownloadCommand.register_subcommand(__A ) EnvironmentCommand.register_subcommand(__A ) RunCommand.register_subcommand(__A ) ServeCommand.register_subcommand(__A ) UserCommands.register_subcommand(__A ) AddNewModelCommand.register_subcommand(__A ) AddNewModelLikeCommand.register_subcommand(__A ) LfsCommands.register_subcommand(__A ) PTtoTFCommand.register_subcommand(__A ) # Let's go UpperCAmelCase__ = parser.parse_args() if not hasattr(__A, "func" ): parser.print_help() exit(1 ) # Run UpperCAmelCase__ = args.func(__A ) service.run() if __name__ == "__main__": main()
65
from __future__ import annotations class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : list[list[int]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed from a list of zero or more lists containing at " "least one and the same number of values, each of which must be of type " "int or float." ) if len(__UpperCAmelCase ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__UpperCAmelCase ) != cols: raise error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def lowercase_ (self : Any ) -> list[list[int]]: """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def lowercase_ (self : Any ) -> int: """simple docstring""" return len(self.rows ) @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" return len(self.rows[0] ) @property def lowercase_ (self : List[Any] ) -> tuple[int, int]: """simple docstring""" return (self.num_rows, self.num_columns) @property def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return self.order[0] == self.order[1] def lowercase_ (self : Any ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : int ) -> int: """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return bool(self.determinant() ) def lowercase_ (self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__UpperCAmelCase ).determinant() def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) return -1 * self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Matrix: """simple docstring""" return Matrix( [ [self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def lowercase_ (self : List[str] ) -> Matrix: """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def lowercase_ (self : Optional[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError("Only matrices with a non-zero determinant have an inverse" ) return self.adjugate() * (1 / determinant) def __repr__(self : Dict ) -> str: """simple docstring""" return str(self.rows ) def __str__(self : Optional[Any] ) -> str: """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ "[" + ". ".join([str(__UpperCAmelCase ) for value in row] ) + ".]" for row in self.rows ] ) + "]" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError("Row must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_columns: raise ValueError( "Row must be equal in length to the other rows in the matrix" ) if position is None: self.rows.append(__UpperCAmelCase ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError( "Column must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in column: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_rows: raise ValueError( "Column must be equal in length to the other columns in the matrix" ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__(self : Any , __UpperCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return NotImplemented return self.rows == other.rows def __ne__(self : int , __UpperCAmelCase : object ) -> bool: """simple docstring""" return not self == other def __neg__(self : Dict ) -> Matrix: """simple docstring""" return self * -1 def __add__(self : Dict , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Addition requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__(self : Optional[Any] , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Subtraction requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__(self : Tuple , __UpperCAmelCase : Matrix | int | float ) -> Matrix: """simple docstring""" if isinstance(__UpperCAmelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): if self.num_columns != other.num_rows: raise ValueError( "The number of columns in the first matrix must " "be equal to the number of rows in the second" ) return Matrix( [ [Matrix.dot_product(__UpperCAmelCase , __UpperCAmelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( "A Matrix can only be multiplied by an int, float, or another matrix" ) def __pow__(self : List[Any] , __UpperCAmelCase : int ) -> Matrix: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("A Matrix can only be raised to the power of an int" ) if not self.is_square: raise ValueError("Only square matrices can be raised to a power" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( "Only invertable matrices can be raised to a negative power" ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : list[int] , __UpperCAmelCase : list[int] ) -> int: """simple docstring""" return sum(row[i] * column[i] for i in range(len(__UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
1
from packaging import version from .import_utils import is_accelerate_available if is_accelerate_available(): import accelerate def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not is_accelerate_available(): return method UpperCAmelCase__ = version.parse(accelerate.__version__ ).base_version if version.parse(__A ) < version.parse("0.17.0" ): return method def wrapper(self, *__A, **__A ): if hasattr(self, "_hf_hook" ) and hasattr(self._hf_hook, "pre_forward" ): self._hf_hook.pre_forward(self ) return method(self, *__A, **__A ) return wrapper
65
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
1
import copy import os from typing import TYPE_CHECKING, List, Union if TYPE_CHECKING: pass from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'kakaobrain/align-base': 'https://huggingface.co/kakaobrain/align-base/resolve/main/config.json', } class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[str] = 'align_text_model' def __init__(self : Dict , __UpperCAmelCase : List[str]=3_0_5_2_2 , __UpperCAmelCase : str=7_6_8 , __UpperCAmelCase : int=1_2 , __UpperCAmelCase : List[str]=1_2 , __UpperCAmelCase : Any=3_0_7_2 , __UpperCAmelCase : Any="gelu" , __UpperCAmelCase : Optional[int]=0.1 , __UpperCAmelCase : int=0.1 , __UpperCAmelCase : Dict=5_1_2 , __UpperCAmelCase : Dict=2 , __UpperCAmelCase : Tuple=0.02 , __UpperCAmelCase : int=1E-12 , __UpperCAmelCase : Dict=0 , __UpperCAmelCase : Optional[int]="absolute" , __UpperCAmelCase : str=True , **__UpperCAmelCase : List[Any] , ) -> Dict: """simple docstring""" super().__init__(**__UpperCAmelCase ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_act UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = position_embedding_type UpperCAmelCase__ = use_cache UpperCAmelCase__ = pad_token_id @classmethod def lowercase_ (cls : Tuple , __UpperCAmelCase : Union[str, os.PathLike] , **__UpperCAmelCase : int ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = cls.get_config_dict(__UpperCAmelCase , **__UpperCAmelCase ) # get the text config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": UpperCAmelCase__ = config_dict["text_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__UpperCAmelCase , **__UpperCAmelCase ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[Any] = 'align_vision_model' def __init__(self : Any , __UpperCAmelCase : int = 3 , __UpperCAmelCase : int = 6_0_0 , __UpperCAmelCase : float = 2.0 , __UpperCAmelCase : float = 3.1 , __UpperCAmelCase : int = 8 , __UpperCAmelCase : List[int] = [3, 3, 5, 3, 5, 5, 3] , __UpperCAmelCase : List[int] = [3_2, 1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2] , __UpperCAmelCase : List[int] = [1_6, 2_4, 4_0, 8_0, 1_1_2, 1_9_2, 3_2_0] , __UpperCAmelCase : List[int] = [] , __UpperCAmelCase : List[int] = [1, 2, 2, 2, 1, 2, 1] , __UpperCAmelCase : List[int] = [1, 2, 2, 3, 3, 4, 1] , __UpperCAmelCase : List[int] = [1, 6, 6, 6, 6, 6, 6] , __UpperCAmelCase : float = 0.25 , __UpperCAmelCase : str = "swish" , __UpperCAmelCase : int = 2_5_6_0 , __UpperCAmelCase : str = "mean" , __UpperCAmelCase : float = 0.02 , __UpperCAmelCase : float = 0.001 , __UpperCAmelCase : float = 0.99 , __UpperCAmelCase : float = 0.2 , **__UpperCAmelCase : List[Any] , ) -> Union[str, Any]: """simple docstring""" super().__init__(**__UpperCAmelCase ) UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = width_coefficient UpperCAmelCase__ = depth_coefficient UpperCAmelCase__ = depth_divisor UpperCAmelCase__ = kernel_sizes UpperCAmelCase__ = in_channels UpperCAmelCase__ = out_channels UpperCAmelCase__ = depthwise_padding UpperCAmelCase__ = strides UpperCAmelCase__ = num_block_repeats UpperCAmelCase__ = expand_ratios UpperCAmelCase__ = squeeze_expansion_ratio UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dim UpperCAmelCase__ = pooling_type UpperCAmelCase__ = initializer_range UpperCAmelCase__ = batch_norm_eps UpperCAmelCase__ = batch_norm_momentum UpperCAmelCase__ = drop_connect_rate UpperCAmelCase__ = sum(__UpperCAmelCase ) * 4 @classmethod def lowercase_ (cls : List[str] , __UpperCAmelCase : Union[str, os.PathLike] , **__UpperCAmelCase : Any ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = cls.get_config_dict(__UpperCAmelCase , **__UpperCAmelCase ) # get the vision config dict if we are loading from AlignConfig if config_dict.get("model_type" ) == "align": UpperCAmelCase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__UpperCAmelCase , **__UpperCAmelCase ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : Any = 'align' __UpperCAmelCase : str = True def __init__(self : str , __UpperCAmelCase : Any=None , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : Tuple=6_4_0 , __UpperCAmelCase : int=1.0 , __UpperCAmelCase : List[str]=0.02 , **__UpperCAmelCase : Optional[int] , ) -> Any: """simple docstring""" super().__init__(**__UpperCAmelCase ) if text_config is None: UpperCAmelCase__ = {} logger.info("text_config is None. Initializing the AlignTextConfig with default values." ) if vision_config is None: UpperCAmelCase__ = {} logger.info("vision_config is None. Initializing the AlignVisionConfig with default values." ) UpperCAmelCase__ = AlignTextConfig(**__UpperCAmelCase ) UpperCAmelCase__ = AlignVisionConfig(**__UpperCAmelCase ) UpperCAmelCase__ = projection_dim UpperCAmelCase__ = temperature_init_value UpperCAmelCase__ = initializer_range @classmethod def lowercase_ (cls : int , __UpperCAmelCase : AlignTextConfig , __UpperCAmelCase : AlignVisionConfig , **__UpperCAmelCase : int ) -> Dict: """simple docstring""" return cls(text_config=text_config.to_dict() , vision_config=vision_config.to_dict() , **__UpperCAmelCase ) def lowercase_ (self : str ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ = self.text_config.to_dict() UpperCAmelCase__ = self.vision_config.to_dict() UpperCAmelCase__ = self.__class__.model_type return output
65
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor class A ( nn.Module ): def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , __UpperCAmelCase : int=(6_4,) , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Any=3_2 , __UpperCAmelCase : str="silu" , __UpperCAmelCase : Any=True , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = torch.nn.Convad( __UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) # down UpperCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_down_block( __UpperCAmelCase , num_layers=self.layers_per_block , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) self.down_blocks.append(__UpperCAmelCase ) # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # out UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = 2 * out_channels if double_z else out_channels UpperCAmelCase__ = nn.Convad(block_out_channels[-1] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : List[Any] , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = x UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : int ): def custom_forward(*__UpperCAmelCase : Optional[Any] ): return module(*__UpperCAmelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __UpperCAmelCase ) else: # down for down_block in self.down_blocks: UpperCAmelCase__ = down_block(__UpperCAmelCase ) # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase ) # post-process UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : List[Any] , __UpperCAmelCase : str=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=("UpDecoderBlock2D",) , __UpperCAmelCase : str=(6_4,) , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : Any="silu" , __UpperCAmelCase : Any="group" , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) UpperCAmelCase__ = in_channels if norm_type == "spatial" else None # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # up UpperCAmelCase__ = list(reversed(__UpperCAmelCase ) ) UpperCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = reversed_block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_up_block( __UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , prev_output_channel=__UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , resnet_time_scale_shift=__UpperCAmelCase , ) self.up_blocks.append(__UpperCAmelCase ) UpperCAmelCase__ = output_channel # out if norm_type == "spatial": UpperCAmelCase__ = SpatialNorm(block_out_channels[0] , __UpperCAmelCase ) else: UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = nn.Convad(block_out_channels[0] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = z UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) UpperCAmelCase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : str ): def custom_forward(*__UpperCAmelCase : List[str] ): return module(*__UpperCAmelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase ) else: # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = up_block(__UpperCAmelCase , __UpperCAmelCase ) # post-process if latent_embeds is None: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) else: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Union[str, Any]="random" , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=True ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = n_e UpperCAmelCase__ = vq_embed_dim UpperCAmelCase__ = beta UpperCAmelCase__ = legacy UpperCAmelCase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) UpperCAmelCase__ = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) UpperCAmelCase__ = self.used.shape[0] UpperCAmelCase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": UpperCAmelCase__ = self.re_embed UpperCAmelCase__ = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: UpperCAmelCase__ = n_e UpperCAmelCase__ = sane_index_shape def lowercase_ (self : str , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) UpperCAmelCase__ = (inds[:, :, None] == used[None, None, ...]).long() UpperCAmelCase__ = match.argmax(-1 ) UpperCAmelCase__ = match.sum(2 ) < 1 if self.unknown_index == "random": UpperCAmelCase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: UpperCAmelCase__ = self.unknown_index return new.reshape(__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token UpperCAmelCase__ = 0 # simply set to zero UpperCAmelCase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __UpperCAmelCase ) return back.reshape(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() UpperCAmelCase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z UpperCAmelCase__ = torch.argmin(torch.cdist(__UpperCAmelCase , self.embedding.weight ) , dim=1 ) UpperCAmelCase__ = self.embedding(__UpperCAmelCase ).view(z.shape ) UpperCAmelCase__ = None UpperCAmelCase__ = None # compute loss for embedding if not self.legacy: UpperCAmelCase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: UpperCAmelCase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients UpperCAmelCase__ = z + (z_q - z).detach() # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: UpperCAmelCase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.remap_to_used(__UpperCAmelCase ) UpperCAmelCase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: UpperCAmelCase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Any: """simple docstring""" if self.remap is not None: UpperCAmelCase__ = indices.reshape(shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.unmap_to_all(__UpperCAmelCase ) UpperCAmelCase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors UpperCAmelCase__ = self.embedding(__UpperCAmelCase ) if shape is not None: UpperCAmelCase__ = z_q.view(__UpperCAmelCase ) # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parameters UpperCAmelCase__ , UpperCAmelCase__ = torch.chunk(__UpperCAmelCase , 2 , dim=1 ) UpperCAmelCase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) UpperCAmelCase__ = deterministic UpperCAmelCase__ = torch.exp(0.5 * self.logvar ) UpperCAmelCase__ = torch.exp(self.logvar ) if self.deterministic: UpperCAmelCase__ = UpperCAmelCase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = randn_tensor( self.mean.shape , generator=__UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) UpperCAmelCase__ = self.mean + self.std * sample return x def lowercase_ (self : str , __UpperCAmelCase : int=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase_ (self : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=[1, 2, 3] ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) UpperCAmelCase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" return self.mean
65
1
import os from typing import Any, Callable, Dict, List, Optional, Tuple, Union import torch from torch import nn from ...models.controlnet import ControlNetModel, ControlNetOutput from ...models.modeling_utils import ModelMixin from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): def __init__(self : List[str] , __UpperCAmelCase : Union[List[ControlNetModel], Tuple[ControlNetModel]] ) -> List[str]: """simple docstring""" super().__init__() UpperCAmelCase__ = nn.ModuleList(__UpperCAmelCase ) def lowercase_ (self : int , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Union[torch.Tensor, float, int] , __UpperCAmelCase : torch.Tensor , __UpperCAmelCase : List[torch.tensor] , __UpperCAmelCase : List[float] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[Dict[str, Any]] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : bool = True , ) -> Union[ControlNetOutput, Tuple]: """simple docstring""" for i, (image, scale, controlnet) in enumerate(zip(__UpperCAmelCase , __UpperCAmelCase , self.nets ) ): UpperCAmelCase__ , UpperCAmelCase__ = controlnet( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , ) # merge samples if i == 0: UpperCAmelCase__ , UpperCAmelCase__ = down_samples, mid_sample else: UpperCAmelCase__ = [ samples_prev + samples_curr for samples_prev, samples_curr in zip(__UpperCAmelCase , __UpperCAmelCase ) ] mid_block_res_sample += mid_sample return down_block_res_samples, mid_block_res_sample def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Union[str, os.PathLike] , __UpperCAmelCase : bool = True , __UpperCAmelCase : Callable = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[str] = None , ) -> List[str]: """simple docstring""" UpperCAmelCase__ = 0 UpperCAmelCase__ = save_directory for controlnet in self.nets: controlnet.save_pretrained( __UpperCAmelCase , is_main_process=__UpperCAmelCase , save_function=__UpperCAmelCase , safe_serialization=__UpperCAmelCase , variant=__UpperCAmelCase , ) idx += 1 UpperCAmelCase__ = model_path_to_save + f"""_{idx}""" @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : Optional[Union[str, os.PathLike]] , **__UpperCAmelCase : List[str] ) -> int: """simple docstring""" UpperCAmelCase__ = 0 UpperCAmelCase__ = [] # load controlnet and append to list until no controlnet directory exists anymore # first controlnet has to be saved under `./mydirectory/controlnet` to be compliant with `DiffusionPipeline.from_prertained` # second, third, ... controlnets have to be saved under `./mydirectory/controlnet_1`, `./mydirectory/controlnet_2`, ... UpperCAmelCase__ = pretrained_model_path while os.path.isdir(__UpperCAmelCase ): UpperCAmelCase__ = ControlNetModel.from_pretrained(__UpperCAmelCase , **__UpperCAmelCase ) controlnets.append(__UpperCAmelCase ) idx += 1 UpperCAmelCase__ = pretrained_model_path + f"""_{idx}""" logger.info(f"""{len(__UpperCAmelCase )} controlnets loaded from {pretrained_model_path}.""" ) if len(__UpperCAmelCase ) == 0: raise ValueError( f"""No ControlNets found under {os.path.dirname(__UpperCAmelCase )}. Expected at least {pretrained_model_path + "_0"}.""" ) return cls(__UpperCAmelCase )
65
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
1
from collections import defaultdict from math import gcd def lowerCAmelCase_ ( __A = 1_500_000 ) -> int: '''simple docstring''' UpperCAmelCase__ = defaultdict(__A ) UpperCAmelCase__ = 2 while 2 * euclid_m * (euclid_m + 1) <= limit: for euclid_n in range((euclid_m % 2) + 1, __A, 2 ): if gcd(__A, __A ) > 1: continue UpperCAmelCase__ = 2 * euclid_m * (euclid_m + euclid_n) for perimeter in range(__A, limit + 1, __A ): frequencies[perimeter] += 1 euclid_m += 1 return sum(1 for frequency in frequencies.values() if frequency == 1 ) if __name__ == "__main__": print(f'''{solution() = }''')
65
def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' def get_matched_characters(__A, __A ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase__ = int(max(0, i - limit ) ) UpperCAmelCase__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(__A ) UpperCAmelCase__ = f"""{_stra[0:_stra.index(__A )]} {_stra[_stra.index(__A ) + 1:]}""" return "".join(__A ) # matching characters UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = len(__A ) # transposition UpperCAmelCase__ = ( len([(ca, ca) for ca, ca in zip(__A, __A ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase__ = 0.0 else: UpperCAmelCase__ = ( 1 / 3 * ( match_count / len(__A ) + match_count / len(__A ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
65
1
from typing import Any def lowerCAmelCase_ ( __A ) -> list[Any]: '''simple docstring''' if not input_list: return [] UpperCAmelCase__ = [input_list.count(__A ) for value in input_list] UpperCAmelCase__ = max(__A ) # Gets the maximum count in the input list. # Gets values of modes return sorted({input_list[i] for i, value in enumerate(__A ) if value == y} ) if __name__ == "__main__": import doctest doctest.testmod()
65
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
1
import torch from diffusers import UnCLIPScheduler from .test_schedulers import SchedulerCommonTest class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[Any] = (UnCLIPScheduler,) def lowercase_ (self : Union[str, Any] , **__UpperCAmelCase : Optional[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = { "num_train_timesteps": 1_0_0_0, "variance_type": "fixed_small_log", "clip_sample": True, "clip_sample_range": 1.0, "prediction_type": "epsilon", } config.update(**__UpperCAmelCase ) return config def lowercase_ (self : Optional[int] ) -> Tuple: """simple docstring""" for timesteps in [1, 5, 1_0_0, 1_0_0_0]: self.check_over_configs(num_train_timesteps=__UpperCAmelCase ) def lowercase_ (self : Optional[Any] ) -> Any: """simple docstring""" for variance in ["fixed_small_log", "learned_range"]: self.check_over_configs(variance_type=__UpperCAmelCase ) def lowercase_ (self : int ) -> Union[str, Any]: """simple docstring""" for clip_sample in [True, False]: self.check_over_configs(clip_sample=__UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" for clip_sample_range in [1, 5, 1_0, 2_0]: self.check_over_configs(clip_sample_range=__UpperCAmelCase ) def lowercase_ (self : List[str] ) -> Dict: """simple docstring""" for prediction_type in ["epsilon", "sample"]: self.check_over_configs(prediction_type=__UpperCAmelCase ) def lowercase_ (self : Dict ) -> List[Any]: """simple docstring""" for time_step in [0, 5_0_0, 9_9_9]: for prev_timestep in [None, 5, 1_0_0, 2_5_0, 5_0_0, 7_5_0]: if prev_timestep is not None and prev_timestep >= time_step: continue self.check_over_forward(time_step=__UpperCAmelCase , prev_timestep=__UpperCAmelCase ) def lowercase_ (self : Dict ) -> Any: """simple docstring""" UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(variance_type="fixed_small_log" ) UpperCAmelCase__ = scheduler_class(**__UpperCAmelCase ) assert torch.sum(torch.abs(scheduler._get_variance(0 ) - 1.0000E-10 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(4_8_7 ) - 0.0549625 ) ) < 1E-5 assert torch.sum(torch.abs(scheduler._get_variance(9_9_9 ) - 0.9994987 ) ) < 1E-5 def lowercase_ (self : Dict ) -> int: """simple docstring""" UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config(variance_type="learned_range" ) UpperCAmelCase__ = scheduler_class(**__UpperCAmelCase ) UpperCAmelCase__ = 0.5 assert scheduler._get_variance(1 , predicted_variance=__UpperCAmelCase ) - -10.1712790 < 1E-5 assert scheduler._get_variance(4_8_7 , predicted_variance=__UpperCAmelCase ) - -5.7998052 < 1E-5 assert scheduler._get_variance(9_9_9 , predicted_variance=__UpperCAmelCase ) - -0.0010011 < 1E-5 def lowercase_ (self : Optional[int] ) -> Any: """simple docstring""" UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**__UpperCAmelCase ) UpperCAmelCase__ = scheduler.timesteps UpperCAmelCase__ = self.dummy_model() UpperCAmelCase__ = self.dummy_sample_deter UpperCAmelCase__ = torch.manual_seed(0 ) for i, t in enumerate(__UpperCAmelCase ): # 1. predict noise residual UpperCAmelCase__ = model(__UpperCAmelCase , __UpperCAmelCase ) # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ = scheduler.step(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , generator=__UpperCAmelCase ).prev_sample UpperCAmelCase__ = pred_prev_sample UpperCAmelCase__ = torch.sum(torch.abs(__UpperCAmelCase ) ) UpperCAmelCase__ = torch.mean(torch.abs(__UpperCAmelCase ) ) assert abs(result_sum.item() - 252.2682495 ) < 1E-2 assert abs(result_mean.item() - 0.3284743 ) < 1E-3 def lowercase_ (self : Any ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.scheduler_classes[0] UpperCAmelCase__ = self.get_scheduler_config() UpperCAmelCase__ = scheduler_class(**__UpperCAmelCase ) scheduler.set_timesteps(2_5 ) UpperCAmelCase__ = scheduler.timesteps UpperCAmelCase__ = self.dummy_model() UpperCAmelCase__ = self.dummy_sample_deter UpperCAmelCase__ = torch.manual_seed(0 ) for i, t in enumerate(__UpperCAmelCase ): # 1. predict noise residual UpperCAmelCase__ = model(__UpperCAmelCase , __UpperCAmelCase ) if i + 1 == timesteps.shape[0]: UpperCAmelCase__ = None else: UpperCAmelCase__ = timesteps[i + 1] # 2. predict previous mean of sample x_t-1 UpperCAmelCase__ = scheduler.step( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , prev_timestep=__UpperCAmelCase , generator=__UpperCAmelCase ).prev_sample UpperCAmelCase__ = pred_prev_sample UpperCAmelCase__ = torch.sum(torch.abs(__UpperCAmelCase ) ) UpperCAmelCase__ = torch.mean(torch.abs(__UpperCAmelCase ) ) assert abs(result_sum.item() - 258.2044983 ) < 1E-2 assert abs(result_mean.item() - 0.3362038 ) < 1E-3 def lowercase_ (self : Dict ) -> Optional[int]: """simple docstring""" pass def lowercase_ (self : List[Any] ) -> List[str]: """simple docstring""" pass
65
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCamelCase__ = 'base_with_context' def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["self_attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["MultiHeadDotProductAttention_0"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array, __A ) UpperCAmelCase__ = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] UpperCAmelCase__ = os.path.join(args.checkpoint_path, "..", "config.gin" ) UpperCAmelCase__ = inference.parse_training_gin_file(__A, __A ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path, __A ) UpperCAmelCase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2", variance_type="fixed_large" ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"], vocab_size=synth_model.model.module.config.vocab_size, d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims, targets_context_length=synth_model.sequence_length["targets_context"], d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims, targets_length=synth_model.sequence_length["targets_context"], max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time, d_model=synth_model.model.module.config.emb_dim, num_layers=synth_model.model.module.config.num_decoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, dropout_rate=synth_model.model.module.config.dropout_rate, ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint["target"]["token_encoder"], __A ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"], __A ) UpperCAmelCase__ = load_decoder(ta_checkpoint["target"]["decoder"], __A ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=__A, continuous_encoder=__A, decoder=__A, scheduler=__A, melgan=__A, ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help='Path to the original jax model checkpoint.', ) UpperCamelCase__ = parser.parse_args() main(args)
65
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_flax_available, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { 'configuration_distilbert': [ 'DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'DistilBertConfig', 'DistilBertOnnxConfig', ], 'tokenization_distilbert': ['DistilBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['DistilBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'DistilBertForMaskedLM', 'DistilBertForMultipleChoice', 'DistilBertForQuestionAnswering', 'DistilBertForSequenceClassification', 'DistilBertForTokenClassification', 'DistilBertModel', 'DistilBertPreTrainedModel', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFDistilBertForMaskedLM', 'TFDistilBertForMultipleChoice', 'TFDistilBertForQuestionAnswering', 'TFDistilBertForSequenceClassification', 'TFDistilBertForTokenClassification', 'TFDistilBertMainLayer', 'TFDistilBertModel', 'TFDistilBertPreTrainedModel', ] try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'FlaxDistilBertForMaskedLM', 'FlaxDistilBertForMultipleChoice', 'FlaxDistilBertForQuestionAnswering', 'FlaxDistilBertForSequenceClassification', 'FlaxDistilBertForTokenClassification', 'FlaxDistilBertModel', 'FlaxDistilBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_distilbert import ( DISTILBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, DistilBertConfig, DistilBertOnnxConfig, ) from .tokenization_distilbert import DistilBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_distilbert_fast import DistilBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_distilbert import ( DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, DistilBertForMaskedLM, DistilBertForMultipleChoice, DistilBertForQuestionAnswering, DistilBertForSequenceClassification, DistilBertForTokenClassification, DistilBertModel, DistilBertPreTrainedModel, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_distilbert import ( TF_DISTILBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFDistilBertForMaskedLM, TFDistilBertForMultipleChoice, TFDistilBertForQuestionAnswering, TFDistilBertForSequenceClassification, TFDistilBertForTokenClassification, TFDistilBertMainLayer, TFDistilBertModel, TFDistilBertPreTrainedModel, ) try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_flax_distilbert import ( FlaxDistilBertForMaskedLM, FlaxDistilBertForMultipleChoice, FlaxDistilBertForQuestionAnswering, FlaxDistilBertForSequenceClassification, FlaxDistilBertForTokenClassification, FlaxDistilBertModel, FlaxDistilBertPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
import math def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' return math.sqrt(__A ) * math.sqrt(__A ) == num def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' UpperCAmelCase__ = 0 UpperCAmelCase__ = n while left <= right: UpperCAmelCase__ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase__ = mid - 1 else: UpperCAmelCase__ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
65
1
import pandas as pd from matplotlib import pyplot as plt from sklearn.linear_model import LinearRegression # Splitting the dataset into the Training set and Test set from sklearn.model_selection import train_test_split # Fitting Polynomial Regression to the dataset from sklearn.preprocessing import PolynomialFeatures # Importing the dataset UpperCamelCase__ = pd.read_csv( 'https://s3.us-west-2.amazonaws.com/public.gamelab.fun/dataset/' 'position_salaries.csv' ) UpperCamelCase__ = dataset.iloc[:, 1:2].values UpperCamelCase__ = dataset.iloc[:, 2].values UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ , UpperCamelCase__ = train_test_split(X, y, test_size=0.2, random_state=0) UpperCamelCase__ = PolynomialFeatures(degree=4) UpperCamelCase__ = poly_reg.fit_transform(X) UpperCamelCase__ = LinearRegression() pol_reg.fit(X_poly, y) def lowerCAmelCase_ ( ) -> str: '''simple docstring''' plt.scatter(__A, __A, color="red" ) plt.plot(__A, pol_reg.predict(poly_reg.fit_transform(__A ) ), color="blue" ) plt.title("Truth or Bluff (Linear Regression)" ) plt.xlabel("Position level" ) plt.ylabel("Salary" ) plt.show() if __name__ == "__main__": viz_polymonial() # Predicting a new result with Polymonial Regression pol_reg.predict(poly_reg.fit_transform([[5.5]])) # output should be 132148.43750003
65
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor __UpperCAmelCase : Optional[torch.FloatTensor] = None def lowerCAmelCase_ ( __A, __A=0.999, __A="cosine", ) -> Tuple: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__A ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase__ = [] for i in range(__A ): UpperCAmelCase__ = i / num_diffusion_timesteps UpperCAmelCase__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__A ) / alpha_bar_fn(__A ), __A ) ) return torch.tensor(__A, dtype=torch.floataa ) class A ( UpperCAmelCase_ , UpperCAmelCase_ ): @register_to_config def __init__(self : List[str] , __UpperCAmelCase : int = 1_0_0_0 , __UpperCAmelCase : str = "fixed_small_log" , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[float] = 1.0 , __UpperCAmelCase : str = "epsilon" , __UpperCAmelCase : str = "squaredcos_cap_v2" , ) -> Optional[int]: """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) UpperCAmelCase__ = betas_for_alpha_bar(__UpperCAmelCase ) UpperCAmelCase__ = 1.0 - self.betas UpperCAmelCase__ = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase__ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase__ = 1.0 # setable values UpperCAmelCase__ = None UpperCAmelCase__ = torch.from_numpy(np.arange(0 , __UpperCAmelCase )[::-1].copy() ) UpperCAmelCase__ = variance_type def lowercase_ (self : List[str] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None ) -> torch.FloatTensor: """simple docstring""" return sample def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, torch.device] = None ) -> Any: """simple docstring""" UpperCAmelCase__ = num_inference_steps UpperCAmelCase__ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase__ = (np.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ).to(__UpperCAmelCase ) def lowercase_ (self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase__ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase__ = torch.log(torch.clamp(__UpperCAmelCase , min=1E-20 ) ) UpperCAmelCase__ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase__ = variance.log() UpperCAmelCase__ = beta.log() UpperCAmelCase__ = (predicted_variance + 1) / 2 UpperCAmelCase__ = frac * max_log + (1 - frac) * min_log return variance def lowercase_ (self : Optional[int] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : int , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : bool = True , ) -> Union[UnCLIPSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase__ , UpperCAmelCase__ = torch.split(__UpperCAmelCase , sample.shape[1] , dim=1 ) else: UpperCAmelCase__ = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] UpperCAmelCase__ = self.alphas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase__ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase__ = model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase__ = torch.clamp( __UpperCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase__ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase__ = 0 if t > 0: UpperCAmelCase__ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__UpperCAmelCase , device=model_output.device ) UpperCAmelCase__ = self._get_variance( __UpperCAmelCase , predicted_variance=__UpperCAmelCase , prev_timestep=__UpperCAmelCase , ) if self.variance_type == "fixed_small_log": UpperCAmelCase__ = variance elif self.variance_type == "learned_range": UpperCAmelCase__ = (0.5 * variance).exp() else: raise ValueError( f"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" " for the UnCLIPScheduler." ) UpperCAmelCase__ = variance * variance_noise UpperCAmelCase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__UpperCAmelCase , pred_original_sample=__UpperCAmelCase ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.IntTensor , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase__ = timesteps.to(original_samples.device ) UpperCAmelCase__ = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase__ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase__ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
65
1
import os import shutil from pathlib import Path from typing import Optional, Union import numpy as np from huggingface_hub import hf_hub_download from ..utils import ONNX_EXTERNAL_WEIGHTS_NAME, ONNX_WEIGHTS_NAME, is_onnx_available, logging if is_onnx_available(): import onnxruntime as ort UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'tensor(bool)': np.bool_, 'tensor(int8)': np.inta, 'tensor(uint8)': np.uinta, 'tensor(int16)': np.intaa, 'tensor(uint16)': np.uintaa, 'tensor(int32)': np.intaa, 'tensor(uint32)': np.uintaa, 'tensor(int64)': np.intaa, 'tensor(uint64)': np.uintaa, 'tensor(float16)': np.floataa, 'tensor(float)': np.floataa, 'tensor(double)': np.floataa, } class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : int=None , **__UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" logger.info("`diffusers.OnnxRuntimeModel` is experimental and might change in the future." ) UpperCAmelCase__ = model UpperCAmelCase__ = kwargs.get("model_save_dir" , __UpperCAmelCase ) UpperCAmelCase__ = kwargs.get("latest_model_name" , __UpperCAmelCase ) def __call__(self : Tuple , **__UpperCAmelCase : int ) -> Any: """simple docstring""" UpperCAmelCase__ = {k: np.array(__UpperCAmelCase ) for k, v in kwargs.items()} return self.model.run(__UpperCAmelCase , __UpperCAmelCase ) @staticmethod def lowercase_ (__UpperCAmelCase : Union[str, Path] , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : Dict=None ) -> List[str]: """simple docstring""" if provider is None: logger.info("No onnxruntime provider specified, using CPUExecutionProvider" ) UpperCAmelCase__ = "CPUExecutionProvider" return ort.InferenceSession(__UpperCAmelCase , providers=[provider] , sess_options=__UpperCAmelCase ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Path] , __UpperCAmelCase : Optional[str] = None , **__UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = file_name if file_name is not None else ONNX_WEIGHTS_NAME UpperCAmelCase__ = self.model_save_dir.joinpath(self.latest_model_name ) UpperCAmelCase__ = Path(__UpperCAmelCase ).joinpath(__UpperCAmelCase ) try: shutil.copyfile(__UpperCAmelCase , __UpperCAmelCase ) except shutil.SameFileError: pass # copy external weights (for models >2GB) UpperCAmelCase__ = self.model_save_dir.joinpath(__UpperCAmelCase ) if src_path.exists(): UpperCAmelCase__ = Path(__UpperCAmelCase ).joinpath(__UpperCAmelCase ) try: shutil.copyfile(__UpperCAmelCase , __UpperCAmelCase ) except shutil.SameFileError: pass def lowercase_ (self : List[str] , __UpperCAmelCase : Union[str, os.PathLike] , **__UpperCAmelCase : Tuple , ) -> Union[str, Any]: """simple docstring""" if os.path.isfile(__UpperCAmelCase ): logger.error(f"""Provided path ({save_directory}) should be a directory, not a file""" ) return os.makedirs(__UpperCAmelCase , exist_ok=__UpperCAmelCase ) # saving model weights/files self._save_pretrained(__UpperCAmelCase , **__UpperCAmelCase ) @classmethod def lowercase_ (cls : List[Any] , __UpperCAmelCase : Union[str, Path] , __UpperCAmelCase : Optional[Union[bool, str, None]] = None , __UpperCAmelCase : Optional[Union[str, None]] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[str] = None , __UpperCAmelCase : Optional[str] = None , __UpperCAmelCase : Optional[str] = None , __UpperCAmelCase : Optional["ort.SessionOptions"] = None , **__UpperCAmelCase : Union[str, Any] , ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = file_name if file_name is not None else ONNX_WEIGHTS_NAME # load model from local directory if os.path.isdir(__UpperCAmelCase ): UpperCAmelCase__ = OnnxRuntimeModel.load_model( os.path.join(__UpperCAmelCase , __UpperCAmelCase ) , provider=__UpperCAmelCase , sess_options=__UpperCAmelCase ) UpperCAmelCase__ = Path(__UpperCAmelCase ) # load model from hub else: # download model UpperCAmelCase__ = hf_hub_download( repo_id=__UpperCAmelCase , filename=__UpperCAmelCase , use_auth_token=__UpperCAmelCase , revision=__UpperCAmelCase , cache_dir=__UpperCAmelCase , force_download=__UpperCAmelCase , ) UpperCAmelCase__ = Path(__UpperCAmelCase ).parent UpperCAmelCase__ = Path(__UpperCAmelCase ).name UpperCAmelCase__ = OnnxRuntimeModel.load_model(__UpperCAmelCase , provider=__UpperCAmelCase , sess_options=__UpperCAmelCase ) return cls(model=__UpperCAmelCase , **__UpperCAmelCase ) @classmethod def lowercase_ (cls : Union[str, Any] , __UpperCAmelCase : Union[str, Path] , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[str] = None , __UpperCAmelCase : Optional[str] = None , **__UpperCAmelCase : int , ) -> int: """simple docstring""" UpperCAmelCase__ = None if len(str(__UpperCAmelCase ).split("@" ) ) == 2: UpperCAmelCase__ , UpperCAmelCase__ = model_id.split("@" ) return cls._from_pretrained( model_id=__UpperCAmelCase , revision=__UpperCAmelCase , cache_dir=__UpperCAmelCase , force_download=__UpperCAmelCase , use_auth_token=__UpperCAmelCase , **__UpperCAmelCase , )
65
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class A ( unittest.TestCase ): def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) UpperCAmelCase__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def lowercase_ (self : List[str] ) -> Any: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : str ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.operation_file_path] print(f"""Command: {cmd}""" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Tuple ) -> int: """simple docstring""" UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Dict ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices, using 2 devices only""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase__ = Accelerator() UpperCamelCase__ = (accelerator.state.process_index + 2, 1_0) UpperCamelCase__ = torch.randint(0, 1_0, shape).to(accelerator.device) UpperCamelCase__ = '' UpperCamelCase__ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." UpperCamelCase__ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." UpperCamelCase__ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
65
1
import os from shutil import copyfile from typing import List, Optional, Tuple from ...tokenization_utils import AddedToken from ...tokenization_utils_fast import PreTrainedTokenizerFast from ...utils import is_sentencepiece_available, logging if is_sentencepiece_available(): from .tokenization_camembert import CamembertTokenizer else: UpperCamelCase__ = None UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'sentencepiece.bpe.model', 'tokenizer_file': 'tokenizer.json'} UpperCamelCase__ = { 'vocab_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/sentencepiece.bpe.model', }, 'tokenizer_file': { 'camembert-base': 'https://huggingface.co/camembert-base/resolve/main/tokenizer.json', }, } UpperCamelCase__ = { 'camembert-base': 5_1_2, } UpperCamelCase__ = '▁' class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : Dict = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Optional[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : int = ['input_ids', 'attention_mask'] __UpperCAmelCase : Tuple = CamembertTokenizer def __init__(self : Optional[Any] , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Any=None , __UpperCAmelCase : Tuple="<s>" , __UpperCAmelCase : int="</s>" , __UpperCAmelCase : Union[str, Any]="</s>" , __UpperCAmelCase : Optional[int]="<s>" , __UpperCAmelCase : str="<unk>" , __UpperCAmelCase : int="<pad>" , __UpperCAmelCase : Dict="<mask>" , __UpperCAmelCase : Union[str, Any]=["<s>NOTUSED", "</s>NOTUSED"] , **__UpperCAmelCase : List[Any] , ) -> Any: """simple docstring""" UpperCAmelCase__ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token super().__init__( __UpperCAmelCase , tokenizer_file=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = vocab_file UpperCAmelCase__ = False if not self.vocab_file else True def lowercase_ (self : Any , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return [self.cls_token_id] + token_ids_a + [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] UpperCAmelCase__ = [self.sep_token_id] return cls + token_ids_a + sep + sep + token_ids_a + sep def lowercase_ (self : Any , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ (self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not self.can_save_slow_tokenizer: raise ValueError( "Your fast tokenizer does not have the necessary information to save the vocabulary for a slow " "tokenizer." ) if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ): copyfile(self.vocab_file , __UpperCAmelCase ) return (out_vocab_file,)
65
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(__A, __A ) def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = emb.weight.shape UpperCAmelCase__ = nn.Linear(__A, __A, bias=__A ) UpperCAmelCase__ = emb.weight.data return lin_layer def lowerCAmelCase_ ( __A, __A="facebook/mbart-large-en-ro", __A=False, __A=False ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = torch.load(__A, map_location="cpu" )["model"] remove_ignore_keys_(__A ) UpperCAmelCase__ = state_dict["encoder.embed_tokens.weight"].shape[0] UpperCAmelCase__ = MBartConfig.from_pretrained(__A, vocab_size=__A ) if mbart_aa and finetuned: UpperCAmelCase__ = "relu" UpperCAmelCase__ = state_dict["decoder.embed_tokens.weight"] UpperCAmelCase__ = MBartForConditionalGeneration(__A ) model.model.load_state_dict(__A ) if finetuned: UpperCAmelCase__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
65
1
import argparse import hashlib # hashlib is only used inside the Test class import struct class A : def __init__(self : Optional[Any] , __UpperCAmelCase : Optional[Any] ) -> int: """simple docstring""" UpperCAmelCase__ = data UpperCAmelCase__ = [0X67_452_301, 0XEF_CDA_B89, 0X98_BAD_CFE, 0X10_325_476, 0XC3_D2E_1F0] @staticmethod def lowercase_ (__UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str ) -> Optional[Any]: """simple docstring""" return ((n << b) | (n >> (3_2 - b))) & 0XFF_FFF_FFF def lowercase_ (self : Optional[int] ) -> Tuple: """simple docstring""" UpperCAmelCase__ = b"\x80" + b"\x00" * (6_3 - (len(self.data ) + 8) % 6_4) UpperCAmelCase__ = self.data + padding + struct.pack(">Q" , 8 * len(self.data ) ) return padded_data def lowercase_ (self : List[str] ) -> List[Any]: """simple docstring""" return [ self.padded_data[i : i + 6_4] for i in range(0 , len(self.padded_data ) , 6_4 ) ] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = list(struct.unpack(">16L" , __UpperCAmelCase ) ) + [0] * 6_4 for i in range(1_6 , 8_0 ): UpperCAmelCase__ = self.rotate((w[i - 3] ^ w[i - 8] ^ w[i - 1_4] ^ w[i - 1_6]) , 1 ) return w def lowercase_ (self : Dict ) -> Dict: """simple docstring""" UpperCAmelCase__ = self.padding() UpperCAmelCase__ = self.split_blocks() for block in self.blocks: UpperCAmelCase__ = self.expand_block(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = self.h for i in range(0 , 8_0 ): if 0 <= i < 2_0: UpperCAmelCase__ = (b & c) | ((~b) & d) UpperCAmelCase__ = 0X5A_827_999 elif 2_0 <= i < 4_0: UpperCAmelCase__ = b ^ c ^ d UpperCAmelCase__ = 0X6E_D9E_BA1 elif 4_0 <= i < 6_0: UpperCAmelCase__ = (b & c) | (b & d) | (c & d) UpperCAmelCase__ = 0X8F_1BB_CDC elif 6_0 <= i < 8_0: UpperCAmelCase__ = b ^ c ^ d UpperCAmelCase__ = 0XCA_62C_1D6 UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ , UpperCAmelCase__ = ( self.rotate(__UpperCAmelCase , 5 ) + f + e + k + expanded_block[i] & 0XFF_FFF_FFF, a, self.rotate(__UpperCAmelCase , 3_0 ), c, d, ) UpperCAmelCase__ = ( self.h[0] + a & 0XFF_FFF_FFF, self.h[1] + b & 0XFF_FFF_FFF, self.h[2] + c & 0XFF_FFF_FFF, self.h[3] + d & 0XFF_FFF_FFF, self.h[4] + e & 0XFF_FFF_FFF, ) return ("{:08x}" * 5).format(*self.h ) def lowerCAmelCase_ ( ) -> Dict: '''simple docstring''' UpperCAmelCase__ = B"Test String" assert SHAaHash(__A ).final_hash() == hashlib.shaa(__A ).hexdigest() # noqa: S324 def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser(description="Process some strings or files" ) parser.add_argument( "--string", dest="input_string", default="Hello World!! Welcome to Cryptography", help="Hash the string", ) parser.add_argument("--file", dest="input_file", help="Hash contents of a file" ) UpperCAmelCase__ = parser.parse_args() UpperCAmelCase__ = args.input_string # In any case hash input should be a bytestring if args.input_file: with open(args.input_file, "rb" ) as f: UpperCAmelCase__ = f.read() else: UpperCAmelCase__ = bytes(__A, "utf-8" ) print(SHAaHash(__A ).final_hash() ) if __name__ == "__main__": main() import doctest doctest.testmod()
65
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase__ = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def lowerCAmelCase_ ( __A, __A=None ) -> Dict: '''simple docstring''' require_version(deps[pkg], __A )
65
1
import unittest from typing import Tuple import torch from diffusers.utils import floats_tensor, randn_tensor, torch_all_close, torch_device from diffusers.utils.testing_utils import require_torch @require_torch class A : @property def lowercase_ (self : Tuple ) -> str: """simple docstring""" return self.get_dummy_input() @property def lowercase_ (self : str ) -> Tuple: """simple docstring""" if self.block_type == "down": return (4, 3_2, 1_6, 1_6) elif self.block_type == "mid": return (4, 3_2, 3_2, 3_2) elif self.block_type == "up": return (4, 3_2, 6_4, 6_4) raise ValueError(f"""'{self.block_type}' is not a supported block_type. Set it to 'up', 'mid', or 'down'.""" ) def lowercase_ (self : Dict , __UpperCAmelCase : Optional[int]=True , __UpperCAmelCase : List[Any]=False , __UpperCAmelCase : str=False , __UpperCAmelCase : Optional[Any]=False , ) -> Any: """simple docstring""" UpperCAmelCase__ = 4 UpperCAmelCase__ = 3_2 UpperCAmelCase__ = (3_2, 3_2) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = torch.device(__UpperCAmelCase ) UpperCAmelCase__ = (batch_size, num_channels) + sizes UpperCAmelCase__ = randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase , device=__UpperCAmelCase ) UpperCAmelCase__ = {"hidden_states": hidden_states} if include_temb: UpperCAmelCase__ = 1_2_8 UpperCAmelCase__ = randn_tensor((batch_size, temb_channels) , generator=__UpperCAmelCase , device=__UpperCAmelCase ) if include_res_hidden_states_tuple: UpperCAmelCase__ = torch.manual_seed(1 ) UpperCAmelCase__ = (randn_tensor(__UpperCAmelCase , generator=__UpperCAmelCase , device=__UpperCAmelCase ),) if include_encoder_hidden_states: UpperCAmelCase__ = floats_tensor((batch_size, 3_2, 3_2) ).to(__UpperCAmelCase ) if include_skip_sample: UpperCAmelCase__ = randn_tensor(((batch_size, 3) + sizes) , generator=__UpperCAmelCase , device=__UpperCAmelCase ) return dummy_input def lowercase_ (self : Optional[int] ) -> int: """simple docstring""" UpperCAmelCase__ = { "in_channels": 3_2, "out_channels": 3_2, "temb_channels": 1_2_8, } if self.block_type == "up": UpperCAmelCase__ = 3_2 if self.block_type == "mid": init_dict.pop("out_channels" ) UpperCAmelCase__ = self.dummy_input return init_dict, inputs_dict def lowercase_ (self : List[str] , __UpperCAmelCase : Any ) -> str: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.prepare_init_args_and_inputs_for_common() UpperCAmelCase__ = self.block_class(**__UpperCAmelCase ) unet_block.to(__UpperCAmelCase ) unet_block.eval() with torch.no_grad(): UpperCAmelCase__ = unet_block(**__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = output[0] self.assertEqual(output.shape , self.output_shape ) UpperCAmelCase__ = output[0, -1, -3:, -3:] UpperCAmelCase__ = torch.tensor(__UpperCAmelCase ).to(__UpperCAmelCase ) assert torch_all_close(output_slice.flatten() , __UpperCAmelCase , atol=5E-3 ) @unittest.skipIf(torch_device == "mps" , "Training is not supported in mps" ) def lowercase_ (self : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.prepare_init_args_and_inputs_for_common() UpperCAmelCase__ = self.block_class(**__UpperCAmelCase ) model.to(__UpperCAmelCase ) model.train() UpperCAmelCase__ = model(**__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = output[0] UpperCAmelCase__ = torch.device(__UpperCAmelCase ) UpperCAmelCase__ = randn_tensor(output.shape , device=__UpperCAmelCase ) UpperCAmelCase__ = torch.nn.functional.mse_loss(__UpperCAmelCase , __UpperCAmelCase ) loss.backward()
65
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path", type=__A, default="data/dump.txt", help="The path to the data." ) parser.add_argument("--tokenizer_type", type=__A, default="bert", choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name", type=__A, default="bert-base-uncased", help="The tokenizer to use." ) parser.add_argument("--dump_file", type=__A, default="data/dump", help="The dump file prefix." ) UpperCAmelCase__ = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": UpperCAmelCase__ = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `[CLS]` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCAmelCase__ = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `<s>` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": UpperCAmelCase__ = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` UpperCAmelCase__ = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path, "r", encoding="utf8" ) as fp: UpperCAmelCase__ = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__A )} examples to process.""" ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 10_000 UpperCAmelCase__ = time.time() for text in data: UpperCAmelCase__ = f"""{bos} {text.strip()} {sep}""" UpperCAmelCase__ = tokenizer.encode(__A, add_special_tokens=__A ) rslt.append(__A ) iter += 1 if iter % interval == 0: UpperCAmelCase__ = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) UpperCAmelCase__ = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__A )} examples processed.""" ) UpperCAmelCase__ = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" UpperCAmelCase__ = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCAmelCase__ = [np.uintaa(__A ) for d in rslt] else: UpperCAmelCase__ = [np.intaa(__A ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__A, "wb" ) as handle: pickle.dump(rslt_, __A, protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
65
1
from statistics import mean import numpy as np def lowerCAmelCase_ ( __A, __A, __A, __A ) -> list: '''simple docstring''' UpperCAmelCase__ = 0 # Number of processes finished UpperCAmelCase__ = 0 # Displays the finished process. # If it is 0, the performance is completed if it is 1, before the performance. UpperCAmelCase__ = [0] * no_of_process # List to include calculation results UpperCAmelCase__ = [0] * no_of_process # Sort by arrival time. UpperCAmelCase__ = [burst_time[i] for i in np.argsort(__A )] UpperCAmelCase__ = [process_name[i] for i in np.argsort(__A )] arrival_time.sort() while no_of_process > finished_process_count: UpperCAmelCase__ = 0 while finished_process[i] == 1: i += 1 if current_time < arrival_time[i]: UpperCAmelCase__ = arrival_time[i] UpperCAmelCase__ = 0 # Index showing the location of the process being performed UpperCAmelCase__ = 0 # Saves the current response ratio. UpperCAmelCase__ = 0 for i in range(0, __A ): if finished_process[i] == 0 and arrival_time[i] <= current_time: UpperCAmelCase__ = (burst_time[i] + (current_time - arrival_time[i])) / burst_time[ i ] if response_ratio < temp: UpperCAmelCase__ = temp UpperCAmelCase__ = i # Calculate the turn around time UpperCAmelCase__ = current_time + burst_time[loc] - arrival_time[loc] current_time += burst_time[loc] # Indicates that the process has been performed. UpperCAmelCase__ = 1 # Increase finished_process_count by 1 finished_process_count += 1 return turn_around_time def lowerCAmelCase_ ( __A, __A, __A, __A ) -> list: '''simple docstring''' UpperCAmelCase__ = [0] * no_of_process for i in range(0, __A ): UpperCAmelCase__ = turn_around_time[i] - burst_time[i] return waiting_time if __name__ == "__main__": UpperCamelCase__ = 5 UpperCamelCase__ = ['A', 'B', 'C', 'D', 'E'] UpperCamelCase__ = [1, 2, 3, 4, 5] UpperCamelCase__ = [1, 2, 3, 4, 5] UpperCamelCase__ = calculate_turn_around_time( process_name, arrival_time, burst_time, no_of_process ) UpperCamelCase__ = calculate_waiting_time( process_name, turn_around_time, burst_time, no_of_process ) print('Process name \tArrival time \tBurst time \tTurn around time \tWaiting time') for i in range(0, no_of_process): print( f'''{process_name[i]}\t\t{arrival_time[i]}\t\t{burst_time[i]}\t\t''' f'''{turn_around_time[i]}\t\t\t{waiting_time[i]}''' ) print(f'''average waiting time : {mean(waiting_time):.5f}''') print(f'''average turn around time : {mean(turn_around_time):.5f}''')
65
from manim import * class A ( UpperCAmelCase_ ): def lowercase_ (self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase__ = Rectangle(height=0.25 , width=0.25 ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("CPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(4 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("GPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) gpu.move_to([-1, -1, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Model" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) model.move_to([3, -1.0, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = fill.copy().set_fill(__UpperCAmelCase , opacity=0.8 ) target.move_to(__UpperCAmelCase ) model_arr.append(__UpperCAmelCase ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(__UpperCAmelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__UpperCAmelCase ) self.add(*__UpperCAmelCase , *__UpperCAmelCase ) UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Disk" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) disk.move_to([-4, -1.25, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(__UpperCAmelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase ) ) UpperCAmelCase__ = Square(0.3 ) input.set_fill(__UpperCAmelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , __UpperCAmelCase , buff=0.5 ) self.play(Write(__UpperCAmelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=__UpperCAmelCase , buff=0.02 ) self.play(MoveToTarget(__UpperCAmelCase ) ) self.play(FadeOut(__UpperCAmelCase ) ) UpperCAmelCase__ = Arrow(start=__UpperCAmelCase , end=__UpperCAmelCase , color=__UpperCAmelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , __UpperCAmelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) UpperCAmelCase__ = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) ) UpperCAmelCase__ = {"run_time": 1, "fade_in": True, "fade_out": True, "buff": 0.02} self.play( Write(__UpperCAmelCase ) , Circumscribe(model_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_cpu_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) UpperCAmelCase__ = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , __UpperCAmelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) UpperCAmelCase__ = AnimationGroup( FadeOut(__UpperCAmelCase , run_time=0.5 ) , MoveToTarget(__UpperCAmelCase , run_time=0.5 ) , FadeIn(__UpperCAmelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(__UpperCAmelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: UpperCAmelCase__ = 0.7 self.play( Circumscribe(model_arr[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_arr[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) UpperCAmelCase__ = a_c UpperCAmelCase__ = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(__UpperCAmelCase ) , FadeOut(__UpperCAmelCase , run_time=0.5 ) , ) UpperCAmelCase__ = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) , MoveToTarget(__UpperCAmelCase ) ) self.wait()
65
1
from ...configuration_utils import PretrainedConfig from ...utils import logging from ...utils.backbone_utils import BackboneConfigMixin, get_aligned_output_features_output_indices UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'shi-labs/nat-mini-in1k-224': 'https://huggingface.co/shi-labs/nat-mini-in1k-224/resolve/main/config.json', # See all Nat models at https://huggingface.co/models?filter=nat } class A ( UpperCAmelCase_ , UpperCAmelCase_ ): __UpperCAmelCase : str = 'nat' __UpperCAmelCase : Tuple = { 'num_attention_heads': 'num_heads', 'num_hidden_layers': 'num_layers', } def __init__(self : List[str] , __UpperCAmelCase : int=4 , __UpperCAmelCase : List[str]=3 , __UpperCAmelCase : str=6_4 , __UpperCAmelCase : Optional[int]=[3, 4, 6, 5] , __UpperCAmelCase : List[Any]=[2, 4, 8, 1_6] , __UpperCAmelCase : str=7 , __UpperCAmelCase : int=3.0 , __UpperCAmelCase : int=True , __UpperCAmelCase : int=0.0 , __UpperCAmelCase : Optional[Any]=0.0 , __UpperCAmelCase : Dict=0.1 , __UpperCAmelCase : int="gelu" , __UpperCAmelCase : Dict=0.02 , __UpperCAmelCase : str=1E-5 , __UpperCAmelCase : str=0.0 , __UpperCAmelCase : int=None , __UpperCAmelCase : Union[str, Any]=None , **__UpperCAmelCase : str , ) -> Union[str, Any]: """simple docstring""" super().__init__(**__UpperCAmelCase ) UpperCAmelCase__ = patch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = embed_dim UpperCAmelCase__ = depths UpperCAmelCase__ = len(__UpperCAmelCase ) UpperCAmelCase__ = num_heads UpperCAmelCase__ = kernel_size UpperCAmelCase__ = mlp_ratio UpperCAmelCase__ = qkv_bias UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = drop_path_rate UpperCAmelCase__ = hidden_act UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = initializer_range # we set the hidden_size attribute in order to make Nat work with VisionEncoderDecoderModel # this indicates the channel dimension after the last stage of the model UpperCAmelCase__ = int(embed_dim * 2 ** (len(__UpperCAmelCase ) - 1) ) UpperCAmelCase__ = layer_scale_init_value UpperCAmelCase__ = ["stem"] + [f"""stage{idx}""" for idx in range(1 , len(__UpperCAmelCase ) + 1 )] UpperCAmelCase__ , UpperCAmelCase__ = get_aligned_output_features_output_indices( out_features=__UpperCAmelCase , out_indices=__UpperCAmelCase , stage_names=self.stage_names )
65
from __future__ import annotations from scipy.special import comb # type: ignore class A : def __init__(self : List[Any] , __UpperCAmelCase : list[tuple[float, float]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. UpperCAmelCase__ = len(__UpperCAmelCase ) - 1 def lowercase_ (self : int , __UpperCAmelCase : float ) -> list[float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __UpperCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__UpperCAmelCase ) , 5 ) == 1 return output_values def lowercase_ (self : Dict , __UpperCAmelCase : float ) -> tuple[float, float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = self.basis_function(__UpperCAmelCase ) UpperCAmelCase__ = 0.0 UpperCAmelCase__ = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def lowercase_ (self : Optional[int] , __UpperCAmelCase : float = 0.01 ) -> Optional[int]: """simple docstring""" from matplotlib import pyplot as plt # type: ignore UpperCAmelCase__ = [] # x coordinates of points to plot UpperCAmelCase__ = [] # y coordinates of points to plot UpperCAmelCase__ = 0.0 while t <= 1: UpperCAmelCase__ = self.bezier_curve_function(__UpperCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size UpperCAmelCase__ = [i[0] for i in self.list_of_points] UpperCAmelCase__ = [i[1] for i in self.list_of_points] plt.plot( __UpperCAmelCase , __UpperCAmelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__UpperCAmelCase , __UpperCAmelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
65
1
import logging from pathlib import Path import numpy as np import pytorch_lightning as pl import torch from pytorch_lightning.callbacks import EarlyStopping, ModelCheckpoint from pytorch_lightning.utilities import rank_zero_only from utils_rag import save_json def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = filter(lambda __A : p.requires_grad, model.parameters() ) UpperCAmelCase__ = sum([np.prod(p.size() ) for p in model_parameters] ) return params UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( __A, __A ) -> str: '''simple docstring''' if metric == "rouge2": UpperCAmelCase__ = "{val_avg_rouge2:.4f}-{step_count}" elif metric == "bleu": UpperCAmelCase__ = "{val_avg_bleu:.4f}-{step_count}" elif metric == "em": UpperCAmelCase__ = "{val_avg_em:.4f}-{step_count}" else: raise NotImplementedError( f"""seq2seq callbacks only support rouge2 and bleu, got {metric}, You can make your own by adding to this""" " function." ) UpperCAmelCase__ = ModelCheckpoint( dirpath=__A, filename=__A, monitor=f"""val_{metric}""", mode="max", save_top_k=3, every_n_epochs=1, ) return checkpoint_callback def lowerCAmelCase_ ( __A, __A ) -> List[Any]: '''simple docstring''' return EarlyStopping( monitor=f"""val_{metric}""", mode="min" if "loss" in metric else "max", patience=__A, verbose=__A, ) class A ( pl.Callback ): def lowercase_ (self : int , __UpperCAmelCase : Any , __UpperCAmelCase : List[Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = {f"""lr_group_{i}""": param["lr"] for i, param in enumerate(pl_module.trainer.optimizers[0].param_groups )} pl_module.logger.log_metrics(__UpperCAmelCase ) @rank_zero_only def lowercase_ (self : str , __UpperCAmelCase : pl.Trainer , __UpperCAmelCase : pl.LightningModule , __UpperCAmelCase : str , __UpperCAmelCase : List[Any]=True ) -> None: """simple docstring""" logger.info(f"""***** {type_path} results at step {trainer.global_step:05d} *****""" ) UpperCAmelCase__ = trainer.callback_metrics trainer.logger.log_metrics({k: v for k, v in metrics.items() if k not in ["log", "progress_bar", "preds"]} ) # Log results UpperCAmelCase__ = Path(pl_module.hparams.output_dir ) if type_path == "test": UpperCAmelCase__ = od / "test_results.txt" UpperCAmelCase__ = od / "test_generations.txt" else: # this never gets hit. I prefer not to save intermediate generations, and results are in metrics.json # If people want this it will be easy enough to add back. UpperCAmelCase__ = od / f"""{type_path}_results/{trainer.global_step:05d}.txt""" UpperCAmelCase__ = od / f"""{type_path}_generations/{trainer.global_step:05d}.txt""" results_file.parent.mkdir(exist_ok=__UpperCAmelCase ) generations_file.parent.mkdir(exist_ok=__UpperCAmelCase ) with open(__UpperCAmelCase , "a+" ) as writer: for key in sorted(__UpperCAmelCase ): if key in ["log", "progress_bar", "preds"]: continue UpperCAmelCase__ = metrics[key] if isinstance(__UpperCAmelCase , torch.Tensor ): UpperCAmelCase__ = val.item() UpperCAmelCase__ = f"""{key}: {val:.6f}\n""" writer.write(__UpperCAmelCase ) if not save_generations: return if "preds" in metrics: UpperCAmelCase__ = "\n".join(metrics["preds"] ) generations_file.open("w+" ).write(__UpperCAmelCase ) @rank_zero_only def lowercase_ (self : List[str] , __UpperCAmelCase : Any , __UpperCAmelCase : Union[str, Any] ) -> Optional[int]: """simple docstring""" try: UpperCAmelCase__ = pl_module.model.model.num_parameters() except AttributeError: UpperCAmelCase__ = pl_module.model.num_parameters() UpperCAmelCase__ = count_trainable_parameters(__UpperCAmelCase ) # mp stands for million parameters trainer.logger.log_metrics({"n_params": npars, "mp": npars / 1E6, "grad_mp": n_trainable_pars / 1E6} ) @rank_zero_only def lowercase_ (self : List[Any] , __UpperCAmelCase : pl.Trainer , __UpperCAmelCase : pl.LightningModule ) -> Union[str, Any]: """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) return self._write_logs(__UpperCAmelCase , __UpperCAmelCase , "test" ) @rank_zero_only def lowercase_ (self : Tuple , __UpperCAmelCase : pl.Trainer , __UpperCAmelCase : Union[str, Any] ) -> int: """simple docstring""" save_json(pl_module.metrics , pl_module.metrics_save_path ) # Uncommenting this will save val generations # return self._write_logs(trainer, pl_module, "valid")
65
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
1
from numpy import exp, pi, sqrt def lowerCAmelCase_ ( __A, __A = 0.0, __A = 1.0 ) -> int: '''simple docstring''' return 1 / sqrt(2 * pi * sigma**2 ) * exp(-((x - mu) ** 2) / (2 * sigma**2) ) if __name__ == "__main__": import doctest doctest.testmod()
65
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
1
from bisect import bisect from itertools import accumulate def lowerCAmelCase_ ( __A, __A, __A, __A ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ = sorted(zip(__A, __A ), key=lambda __A : x[0] / x[1], reverse=__A ) UpperCAmelCase__ , UpperCAmelCase__ = [i[0] for i in r], [i[1] for i in r] UpperCAmelCase__ = list(accumulate(__A ) ) UpperCAmelCase__ = bisect(__A, __A ) return ( 0 if k == 0 else sum(vl[:k] ) + (w - acc[k - 1]) * (vl[k]) / (wt[k]) if k != n else sum(vl[:k] ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
from __future__ import annotations from collections import deque class A : def __init__(self : Dict , __UpperCAmelCase : list[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__UpperCAmelCase ) self.set_fail_transitions() def lowercase_ (self : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : str ) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = 0 for character in keyword: UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , __UpperCAmelCase ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase__ = len(self.adlist ) - 1 else: UpperCAmelCase__ = next_state self.adlist[current_state]["output"].append(__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = deque() for node in self.adlist[0]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = 0 while q: UpperCAmelCase__ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = self.adlist[r]["fail_state"] while ( self.find_next_state(__UpperCAmelCase , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase__ = self.adlist[state]["fail_state"] UpperCAmelCase__ = self.find_next_state( __UpperCAmelCase , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase__ = 0 UpperCAmelCase__ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> dict[str, list[int]]: """simple docstring""" UpperCAmelCase__ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase__ = 0 for i in range(len(__UpperCAmelCase ) ): while ( self.find_next_state(__UpperCAmelCase , string[i] ) is None and current_state != 0 ): UpperCAmelCase__ = self.adlist[current_state]["fail_state"] UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , string[i] ) if next_state is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase__ = [] result[key].append(i - len(__UpperCAmelCase ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
65
1
from __future__ import annotations import copy import inspect import unittest import numpy as np from transformers import is_tf_available, is_vision_available from transformers.models.auto import get_values from transformers.testing_utils import require_tf, slow from transformers.utils import cached_property from ...test_configuration_common import ConfigTester from ...test_modeling_tf_common import TFModelTesterMixin, floats_tensor, ids_tensor, random_attention_mask from ...test_pipeline_mixin import PipelineTesterMixin if is_tf_available(): import tensorflow as tf from transformers import ( TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST, TF_MODEL_FOR_MULTIPLE_CHOICE_MAPPING, TF_MODEL_FOR_QUESTION_ANSWERING_MAPPING, TF_MODEL_FOR_SEQUENCE_CLASSIFICATION_MAPPING, TF_MODEL_FOR_TOKEN_CLASSIFICATION_MAPPING, LayoutLMvaConfig, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, TFLayoutLMvaModel, ) if is_vision_available(): from PIL import Image from transformers import LayoutLMvaImageProcessor class A : def __init__(self : int , __UpperCAmelCase : Any , __UpperCAmelCase : str=2 , __UpperCAmelCase : str=3 , __UpperCAmelCase : Tuple=4 , __UpperCAmelCase : List[Any]=2 , __UpperCAmelCase : int=7 , __UpperCAmelCase : Dict=True , __UpperCAmelCase : List[Any]=True , __UpperCAmelCase : Tuple=True , __UpperCAmelCase : Union[str, Any]=True , __UpperCAmelCase : Dict=9_9 , __UpperCAmelCase : List[str]=3_6 , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : int=4 , __UpperCAmelCase : Optional[Any]=3_7 , __UpperCAmelCase : List[str]="gelu" , __UpperCAmelCase : Dict=0.1 , __UpperCAmelCase : Any=0.1 , __UpperCAmelCase : int=5_1_2 , __UpperCAmelCase : str=1_6 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Any=0.02 , __UpperCAmelCase : str=6 , __UpperCAmelCase : str=6 , __UpperCAmelCase : Any=3 , __UpperCAmelCase : List[Any]=4 , __UpperCAmelCase : Dict=None , __UpperCAmelCase : List[Any]=1_0_0_0 , ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = parent UpperCAmelCase__ = batch_size UpperCAmelCase__ = num_channels UpperCAmelCase__ = image_size UpperCAmelCase__ = patch_size UpperCAmelCase__ = is_training UpperCAmelCase__ = use_input_mask UpperCAmelCase__ = use_token_type_ids UpperCAmelCase__ = use_labels UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_act UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = type_vocab_size UpperCAmelCase__ = type_sequence_label_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = coordinate_size UpperCAmelCase__ = shape_size UpperCAmelCase__ = num_labels UpperCAmelCase__ = num_choices UpperCAmelCase__ = scope UpperCAmelCase__ = range_bbox # LayoutLMv3's sequence length equals the number of text tokens + number of patches + 1 (we add 1 for the CLS token) UpperCAmelCase__ = text_seq_length UpperCAmelCase__ = (image_size // patch_size) ** 2 + 1 UpperCAmelCase__ = self.text_seq_length + self.image_seq_length def lowercase_ (self : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = ids_tensor([self.batch_size, self.text_seq_length] , self.vocab_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.text_seq_length, 4] , self.range_bbox ) UpperCAmelCase__ = bbox.numpy() # Ensure that bbox is legal for i in range(bbox.shape[0] ): for j in range(bbox.shape[1] ): if bbox[i, j, 3] < bbox[i, j, 1]: UpperCAmelCase__ = bbox[i, j, 3] UpperCAmelCase__ = bbox[i, j, 1] UpperCAmelCase__ = tmp_coordinate if bbox[i, j, 2] < bbox[i, j, 0]: UpperCAmelCase__ = bbox[i, j, 2] UpperCAmelCase__ = bbox[i, j, 0] UpperCAmelCase__ = tmp_coordinate UpperCAmelCase__ = tf.constant(__UpperCAmelCase ) UpperCAmelCase__ = floats_tensor([self.batch_size, self.num_channels, self.image_size, self.image_size] ) UpperCAmelCase__ = None if self.use_input_mask: UpperCAmelCase__ = random_attention_mask([self.batch_size, self.text_seq_length] ) UpperCAmelCase__ = None if self.use_token_type_ids: UpperCAmelCase__ = ids_tensor([self.batch_size, self.text_seq_length] , self.type_vocab_size ) UpperCAmelCase__ = None UpperCAmelCase__ = None if self.use_labels: UpperCAmelCase__ = ids_tensor([self.batch_size] , self.type_sequence_label_size ) UpperCAmelCase__ = ids_tensor([self.batch_size, self.text_seq_length] , self.num_labels ) UpperCAmelCase__ = LayoutLMvaConfig( vocab_size=self.vocab_size , hidden_size=self.hidden_size , num_hidden_layers=self.num_hidden_layers , num_attention_heads=self.num_attention_heads , intermediate_size=self.intermediate_size , hidden_act=self.hidden_act , hidden_dropout_prob=self.hidden_dropout_prob , attention_probs_dropout_prob=self.attention_probs_dropout_prob , max_position_embeddings=self.max_position_embeddings , type_vocab_size=self.type_vocab_size , initializer_range=self.initializer_range , coordinate_size=self.coordinate_size , shape_size=self.shape_size , input_size=self.image_size , patch_size=self.patch_size , ) return config, input_ids, bbox, pixel_values, token_type_ids, input_mask, sequence_labels, token_labels def lowercase_ (self : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = TFLayoutLMvaModel(config=__UpperCAmelCase ) # text + image UpperCAmelCase__ = model(__UpperCAmelCase , pixel_values=__UpperCAmelCase , training=__UpperCAmelCase ) UpperCAmelCase__ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , training=__UpperCAmelCase , ) UpperCAmelCase__ = model(__UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , training=__UpperCAmelCase ) self.parent.assertEqual(result.last_hidden_state.shape , (self.batch_size, self.seq_length, self.hidden_size) ) # text only UpperCAmelCase__ = model(__UpperCAmelCase , training=__UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.text_seq_length, self.hidden_size) ) # image only UpperCAmelCase__ = model({"pixel_values": pixel_values} , training=__UpperCAmelCase ) self.parent.assertEqual( result.last_hidden_state.shape , (self.batch_size, self.image_seq_length, self.hidden_size) ) def lowercase_ (self : List[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict , __UpperCAmelCase : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFLayoutLMvaForSequenceClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , training=__UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.num_labels) ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, Any] ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.num_labels UpperCAmelCase__ = TFLayoutLMvaForTokenClassification(config=__UpperCAmelCase ) UpperCAmelCase__ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , labels=__UpperCAmelCase , training=__UpperCAmelCase , ) self.parent.assertEqual(result.logits.shape , (self.batch_size, self.text_seq_length, self.num_labels) ) def lowercase_ (self : str , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = 2 UpperCAmelCase__ = TFLayoutLMvaForQuestionAnswering(config=__UpperCAmelCase ) UpperCAmelCase__ = model( __UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , start_positions=__UpperCAmelCase , end_positions=__UpperCAmelCase , training=__UpperCAmelCase , ) self.parent.assertEqual(result.start_logits.shape , (self.batch_size, self.seq_length) ) self.parent.assertEqual(result.end_logits.shape , (self.batch_size, self.seq_length) ) def lowercase_ (self : Tuple ) -> str: """simple docstring""" UpperCAmelCase__ = self.prepare_config_and_inputs() ((UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__) , (UpperCAmelCase__)) = config_and_inputs UpperCAmelCase__ = { "input_ids": input_ids, "bbox": bbox, "pixel_values": pixel_values, "token_type_ids": token_type_ids, "attention_mask": input_mask, } return config, inputs_dict @require_tf class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Optional[Any] = ( ( TFLayoutLMvaModel, TFLayoutLMvaForQuestionAnswering, TFLayoutLMvaForSequenceClassification, TFLayoutLMvaForTokenClassification, ) if is_tf_available() else () ) __UpperCAmelCase : Tuple = ( {'document-question-answering': TFLayoutLMvaForQuestionAnswering, 'feature-extraction': TFLayoutLMvaModel} if is_tf_available() else {} ) __UpperCAmelCase : Union[str, Any] = False __UpperCAmelCase : Optional[int] = False __UpperCAmelCase : int = False def lowercase_ (self : int , __UpperCAmelCase : Any , __UpperCAmelCase : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : int ) -> List[str]: """simple docstring""" return True def lowercase_ (self : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : Optional[Any] , __UpperCAmelCase : List[str]=False ) -> dict: """simple docstring""" UpperCAmelCase__ = copy.deepcopy(__UpperCAmelCase ) if model_class in get_values(__UpperCAmelCase ): UpperCAmelCase__ = { k: tf.tile(tf.expand_dims(__UpperCAmelCase , 1 ) , (1, self.model_tester.num_choices) + (1,) * (v.ndim - 1) ) if isinstance(__UpperCAmelCase , tf.Tensor ) and v.ndim > 0 else v for k, v in inputs_dict.items() } if return_labels: if model_class in get_values(__UpperCAmelCase ): UpperCAmelCase__ = tf.ones(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(__UpperCAmelCase ): UpperCAmelCase__ = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) UpperCAmelCase__ = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(__UpperCAmelCase ): UpperCAmelCase__ = tf.zeros(self.model_tester.batch_size , dtype=tf.intaa ) elif model_class in get_values(__UpperCAmelCase ): UpperCAmelCase__ = tf.zeros( (self.model_tester.batch_size, self.model_tester.text_seq_length) , dtype=tf.intaa ) return inputs_dict def lowercase_ (self : int ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TFLayoutLMvaModelTester(self ) UpperCAmelCase__ = ConfigTester(self , config_class=__UpperCAmelCase , hidden_size=3_7 ) def lowercase_ (self : Tuple ) -> Optional[int]: """simple docstring""" self.config_tester.run_common_tests() def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.model_tester.prepare_config_and_inputs_for_common() for model_class in self.all_model_classes: UpperCAmelCase__ = model_class(__UpperCAmelCase ) if getattr(__UpperCAmelCase , "hf_compute_loss" , __UpperCAmelCase ): # The number of elements in the loss should be the same as the number of elements in the label UpperCAmelCase__ = self._prepare_for_class(inputs_dict.copy() , __UpperCAmelCase , return_labels=__UpperCAmelCase ) UpperCAmelCase__ = prepared_for_class[ sorted(prepared_for_class.keys() - inputs_dict.keys() , reverse=__UpperCAmelCase )[0] ] UpperCAmelCase__ = added_label.shape.as_list()[:1] # Test that model correctly compute the loss with kwargs UpperCAmelCase__ = self._prepare_for_class(inputs_dict.copy() , __UpperCAmelCase , return_labels=__UpperCAmelCase ) UpperCAmelCase__ = prepared_for_class.pop("input_ids" ) UpperCAmelCase__ = model(__UpperCAmelCase , **__UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss when we mask some positions UpperCAmelCase__ = self._prepare_for_class(inputs_dict.copy() , __UpperCAmelCase , return_labels=__UpperCAmelCase ) UpperCAmelCase__ = prepared_for_class.pop("input_ids" ) if "labels" in prepared_for_class: UpperCAmelCase__ = prepared_for_class["labels"].numpy() if len(labels.shape ) > 1 and labels.shape[1] != 1: UpperCAmelCase__ = -1_0_0 UpperCAmelCase__ = tf.convert_to_tensor(__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase , **__UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) self.assertTrue(not np.any(np.isnan(loss.numpy() ) ) ) # Test that model correctly compute the loss with a dict UpperCAmelCase__ = self._prepare_for_class(inputs_dict.copy() , __UpperCAmelCase , return_labels=__UpperCAmelCase ) UpperCAmelCase__ = model(__UpperCAmelCase )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) # Test that model correctly compute the loss with a tuple UpperCAmelCase__ = self._prepare_for_class(inputs_dict.copy() , __UpperCAmelCase , return_labels=__UpperCAmelCase ) # Get keys that were added with the _prepare_for_class function UpperCAmelCase__ = prepared_for_class.keys() - inputs_dict.keys() UpperCAmelCase__ = inspect.signature(model.call ).parameters UpperCAmelCase__ = list(signature.keys() ) # Create a dictionary holding the location of the tensors in the tuple UpperCAmelCase__ = {0: "input_ids"} for label_key in label_keys: UpperCAmelCase__ = signature_names.index(__UpperCAmelCase ) UpperCAmelCase__ = label_key UpperCAmelCase__ = sorted(tuple_index_mapping.items() ) # Initialize a list with their default values, update the values and convert to a tuple UpperCAmelCase__ = [] for name in signature_names: if name != "kwargs": list_input.append(signature[name].default ) for index, value in sorted_tuple_index_mapping: UpperCAmelCase__ = prepared_for_class[value] UpperCAmelCase__ = tuple(__UpperCAmelCase ) # Send to model UpperCAmelCase__ = model(tuple_input[:-1] )[0] self.assertTrue(loss.shape.as_list() == expected_loss_size or loss.shape.as_list() == [1] ) def lowercase_ (self : Optional[int] ) -> Optional[Any]: """simple docstring""" ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = self.model_tester.prepare_config_and_inputs() for type in ["absolute", "relative_key", "relative_key_query"]: UpperCAmelCase__ = type self.model_tester.create_and_check_model(__UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_sequence_classification( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> List[Any]: """simple docstring""" ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_token_classification( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : int ) -> List[Any]: """simple docstring""" ( ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ( UpperCAmelCase__ ) , ) = self.model_tester.prepare_config_and_inputs() self.model_tester.create_and_check_for_question_answering( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase ) @slow def lowercase_ (self : Any ) -> str: """simple docstring""" for model_name in TF_LAYOUTLMV3_PRETRAINED_MODEL_ARCHIVE_LIST[:1]: UpperCAmelCase__ = TFLayoutLMvaModel.from_pretrained(__UpperCAmelCase ) self.assertIsNotNone(__UpperCAmelCase ) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = Image.open("./tests/fixtures/tests_samples/COCO/000000039769.png" ) return image @require_tf class A ( unittest.TestCase ): @cached_property def lowercase_ (self : int ) -> int: """simple docstring""" return LayoutLMvaImageProcessor(apply_ocr=__UpperCAmelCase ) if is_vision_available() else None @slow def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = TFLayoutLMvaModel.from_pretrained("microsoft/layoutlmv3-base" ) UpperCAmelCase__ = self.default_image_processor UpperCAmelCase__ = prepare_img() UpperCAmelCase__ = image_processor(images=__UpperCAmelCase , return_tensors="tf" ).pixel_values UpperCAmelCase__ = tf.constant([[1, 2]] ) UpperCAmelCase__ = tf.expand_dims(tf.constant([[1, 2, 3, 4], [5, 6, 7, 8]] ) , axis=0 ) # forward pass UpperCAmelCase__ = model(input_ids=__UpperCAmelCase , bbox=__UpperCAmelCase , pixel_values=__UpperCAmelCase , training=__UpperCAmelCase ) # verify the logits UpperCAmelCase__ = (1, 1_9_9, 7_6_8) self.assertEqual(outputs.last_hidden_state.shape , __UpperCAmelCase ) UpperCAmelCase__ = tf.constant( [[-0.0529, 0.3618, 0.1632], [-0.1587, -0.1667, -0.0400], [-0.1557, -0.1671, -0.0505]] ) self.assertTrue(np.allclose(outputs.last_hidden_state[0, :3, :3] , __UpperCAmelCase , atol=1E-4 ) )
65
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ['input_values', 'attention_mask'] def __init__(self : Any , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 1_6_0_0_0 , __UpperCAmelCase : float = 0.0 , __UpperCAmelCase : bool = False , __UpperCAmelCase : int = 8_0 , __UpperCAmelCase : int = 1_6 , __UpperCAmelCase : int = 6_4 , __UpperCAmelCase : str = "hann_window" , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : float = 8_0 , __UpperCAmelCase : float = 7_6_0_0 , __UpperCAmelCase : float = 1E-10 , __UpperCAmelCase : int = 2 , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Any , ) -> str: """simple docstring""" super().__init__(feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = do_normalize UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = frame_signal_scale UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = reduction_factor UpperCAmelCase__ = win_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = hop_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = optimal_fft_length(self.sample_size ) UpperCAmelCase__ = (self.n_fft // 2) + 1 UpperCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCAmelCase ) UpperCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowercase_ (__UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase , np.intaa ) UpperCAmelCase__ = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): UpperCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: UpperCAmelCase__ = padding_value normed_input_values.append(__UpperCAmelCase ) else: UpperCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowercase_ (self : Optional[int] , __UpperCAmelCase : np.ndarray , ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ = spectrogram( __UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__(self : Any , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[int] = None , **__UpperCAmelCase : str , ) -> BatchFeature: """simple docstring""" if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) else: UpperCAmelCase__ = None if audio_target is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) if inputs is None: return inputs_target else: UpperCAmelCase__ = inputs_target["input_values"] UpperCAmelCase__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: UpperCAmelCase__ = decoder_attention_mask return inputs def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCAmelCase : bool = False , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : Any , ) -> BatchFeature: """simple docstring""" UpperCAmelCase__ = isinstance(__UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): UpperCAmelCase__ = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs UpperCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: UpperCAmelCase__ = [self._extract_mel_features(__UpperCAmelCase ) for waveform in speech] UpperCAmelCase__ = BatchFeature({"input_values": features} ) UpperCAmelCase__ = self.num_mel_bins else: UpperCAmelCase__ = BatchFeature({"input_values": speech} ) UpperCAmelCase__ = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = feature_size_hack # convert input values to correct format UpperCAmelCase__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(__UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): UpperCAmelCase__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(__UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = input_values.astype(np.floataa ) # convert attention_mask to correct format UpperCAmelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: UpperCAmelCase__ = ( attention_mask if self._get_padding_strategies(__UpperCAmelCase , max_length=__UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=__UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def lowercase_ (self : Tuple ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. UpperCAmelCase__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
65
1
import builtins import sys from ...utils.imports import _is_package_available from . import cursor, input from .helpers import Direction, clear_line, forceWrite, linebreak, move_cursor, reset_cursor, writeColor from .keymap import KEYMAP UpperCamelCase__ = False try: UpperCamelCase__ = _is_package_available('google.colab') except ModuleNotFoundError: pass @input.register class A : def __init__(self : Optional[int] , __UpperCAmelCase : str = None , __UpperCAmelCase : list = [] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = 0 UpperCAmelCase__ = choices UpperCAmelCase__ = prompt if sys.platform == "win32": UpperCAmelCase__ = "*" else: UpperCAmelCase__ = "➔ " def lowercase_ (self : Optional[int] , __UpperCAmelCase : List[Any] , __UpperCAmelCase : str = "" ) -> Optional[int]: """simple docstring""" if sys.platform != "win32": writeColor(self.choices[index] , 3_2 , __UpperCAmelCase ) else: forceWrite(self.choices[index] , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : int ) -> Dict: """simple docstring""" if index == self.position: forceWrite(f""" {self.arrow_char} """ ) self.write_choice(__UpperCAmelCase ) else: forceWrite(f""" {self.choices[index]}""" ) reset_cursor() def lowercase_ (self : Dict , __UpperCAmelCase : Direction , __UpperCAmelCase : int = 1 ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = self.position if direction == Direction.DOWN: if self.position + 1 >= len(self.choices ): return self.position += num_spaces else: if self.position - 1 < 0: return self.position -= num_spaces clear_line() self.print_choice(__UpperCAmelCase ) move_cursor(__UpperCAmelCase , direction.name ) self.print_choice(self.position ) @input.mark(KEYMAP["up"] ) def lowercase_ (self : Tuple ) -> Optional[int]: """simple docstring""" self.move_direction(Direction.UP ) @input.mark(KEYMAP["down"] ) def lowercase_ (self : Any ) -> Optional[int]: """simple docstring""" self.move_direction(Direction.DOWN ) @input.mark(KEYMAP["newline"] ) def lowercase_ (self : List[str] ) -> Optional[Any]: """simple docstring""" move_cursor(len(self.choices ) - self.position , "DOWN" ) return self.position @input.mark(KEYMAP["interrupt"] ) def lowercase_ (self : Tuple ) -> Union[str, Any]: """simple docstring""" move_cursor(len(self.choices ) - self.position , "DOWN" ) raise KeyboardInterrupt @input.mark_multiple(*[KEYMAP[str(__UpperCAmelCase )] for number in range(1_0 )] ) def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = int(chr(self.current_selection ) ) UpperCAmelCase__ = index - self.position if index == self.position: return if index < len(self.choices ): if self.position > index: self.move_direction(Direction.UP , -movement ) elif self.position < index: self.move_direction(Direction.DOWN , __UpperCAmelCase ) else: return else: return def lowercase_ (self : Tuple , __UpperCAmelCase : int = 0 ) -> Union[str, Any]: """simple docstring""" if self.prompt: linebreak() forceWrite(self.prompt , "\n" ) if in_colab: forceWrite("Please input a choice index (starting from 0), and press enter" , "\n" ) else: forceWrite("Please select a choice using the arrow or number keys, and selecting with enter" , "\n" ) UpperCAmelCase__ = default_choice for i in range(len(self.choices ) ): self.print_choice(__UpperCAmelCase ) forceWrite("\n" ) move_cursor(len(self.choices ) - self.position , "UP" ) with cursor.hide(): while True: if in_colab: try: UpperCAmelCase__ = int(builtins.input() ) except ValueError: UpperCAmelCase__ = default_choice else: UpperCAmelCase__ = self.handle_input() if choice is not None: reset_cursor() for _ in range(len(self.choices ) + 1 ): move_cursor(1 , "UP" ) clear_line() self.write_choice(__UpperCAmelCase , "\n" ) return choice
65
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[torch.FloatTensor] = None __UpperCAmelCase : torch.FloatTensor = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Union[str, Any]=5_1_2 , __UpperCAmelCase : List[str]="cls" , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=True , **__UpperCAmelCase : str , ) -> int: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = project_dim UpperCAmelCase__ = pooler_fn UpperCAmelCase__ = learn_encoder UpperCAmelCase__ = use_attention_mask class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = [r'pooler', r'logit_scale'] __UpperCAmelCase : int = [r'position_ids', r'predictions.decoder.bias'] __UpperCAmelCase : Any = 'roberta' __UpperCAmelCase : List[str] = RobertaSeriesConfig def __init__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = XLMRobertaModel(__UpperCAmelCase ) UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = getattr(__UpperCAmelCase , "has_pre_transformation" , __UpperCAmelCase ) if self.has_pre_transformation: UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.base_model( input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , position_ids=__UpperCAmelCase , head_mask=__UpperCAmelCase , inputs_embeds=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__UpperCAmelCase , ) if self.has_pre_transformation: UpperCAmelCase__ = outputs["hidden_states"][-2] UpperCAmelCase__ = self.pre_LN(__UpperCAmelCase ) UpperCAmelCase__ = self.transformation_pre(__UpperCAmelCase ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: UpperCAmelCase__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
65
1
import importlib import shutil import threading import warnings from typing import List import fsspec import fsspec.asyn from . import compression from .hffilesystem import HfFileSystem UpperCamelCase__ = importlib.util.find_spec('s3fs') is not None if _has_safs: from .safilesystem import SaFileSystem # noqa: F401 UpperCamelCase__ = [ compression.BzaFileSystem, compression.GzipFileSystem, compression.LzaFileSystem, compression.XzFileSystem, compression.ZstdFileSystem, ] # Register custom filesystems for fs_class in COMPRESSION_FILESYSTEMS + [HfFileSystem]: if fs_class.protocol in fsspec.registry and fsspec.registry[fs_class.protocol] is not fs_class: warnings.warn(f'''A filesystem protocol was already set for {fs_class.protocol} and will be overwritten.''') fsspec.register_implementation(fs_class.protocol, fs_class, clobber=True) def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' if "://" in dataset_path: UpperCAmelCase__ = dataset_path.split("://" )[1] return dataset_path def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' if fs is not None and fs.protocol != "file": return True else: return False def lowerCAmelCase_ ( __A, __A, __A ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = not is_remote_filesystem(__A ) if is_local: # LocalFileSystem.mv does copy + rm, it is more efficient to simply move a local directory shutil.move(fs._strip_protocol(__A ), fs._strip_protocol(__A ) ) else: fs.mv(__A, __A, recursive=__A ) def lowerCAmelCase_ ( ) -> None: '''simple docstring''' if hasattr(fsspec.asyn, "reset_lock" ): # for future fsspec>2022.05.0 fsspec.asyn.reset_lock() else: UpperCAmelCase__ = None UpperCAmelCase__ = None UpperCAmelCase__ = threading.Lock()
65
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
1
from __future__ import annotations def lowerCAmelCase_ ( __A = 4 ) -> list[list[int]]: '''simple docstring''' UpperCAmelCase__ = abs(__A ) or 4 return [[1 + x + y * row_size for x in range(__A )] for y in range(__A )] def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' return reverse_row(transpose(__A ) ) # OR.. transpose(reverse_column(matrix)) def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' return reverse_row(reverse_column(__A ) ) # OR.. reverse_column(reverse_row(matrix)) def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' return reverse_column(transpose(__A ) ) # OR.. transpose(reverse_row(matrix)) def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' UpperCAmelCase__ = [list(__A ) for x in zip(*__A )] return matrix def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' UpperCAmelCase__ = matrix[::-1] return matrix def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' UpperCAmelCase__ = [x[::-1] for x in matrix] return matrix def lowerCAmelCase_ ( __A ) -> None: '''simple docstring''' for i in matrix: print(*__A ) if __name__ == "__main__": UpperCamelCase__ = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 90 counterclockwise:\n') print_matrix(rotate_aa(matrix)) UpperCamelCase__ = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 180:\n') print_matrix(rotate_aaa(matrix)) UpperCamelCase__ = make_matrix() print('\norigin:\n') print_matrix(matrix) print('\nrotate 270 counterclockwise:\n') print_matrix(rotate_aaa(matrix))
65
import math import random def lowerCAmelCase_ ( __A, __A = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase__ = 0.0_2 def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' UpperCAmelCase__ = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__A ): # Forward propagation UpperCAmelCase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase__ = (expected / 100) - layer_a # Error delta UpperCAmelCase__ = layer_1_error * sigmoid_function(__A, __A ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input('Expected value: ')) UpperCamelCase__ = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
65
1
import argparse import io import requests import torch from omegaconf import OmegaConf from diffusers import AutoencoderKL from diffusers.pipelines.stable_diffusion.convert_from_ckpt import ( assign_to_checkpoint, conv_attn_to_linear, create_vae_diffusers_config, renew_vae_attention_paths, renew_vae_resnet_paths, ) def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ = checkpoint UpperCAmelCase__ = {} UpperCAmelCase__ = vae_state_dict["encoder.conv_in.weight"] UpperCAmelCase__ = vae_state_dict["encoder.conv_in.bias"] UpperCAmelCase__ = vae_state_dict["encoder.conv_out.weight"] UpperCAmelCase__ = vae_state_dict["encoder.conv_out.bias"] UpperCAmelCase__ = vae_state_dict["encoder.norm_out.weight"] UpperCAmelCase__ = vae_state_dict["encoder.norm_out.bias"] UpperCAmelCase__ = vae_state_dict["decoder.conv_in.weight"] UpperCAmelCase__ = vae_state_dict["decoder.conv_in.bias"] UpperCAmelCase__ = vae_state_dict["decoder.conv_out.weight"] UpperCAmelCase__ = vae_state_dict["decoder.conv_out.bias"] UpperCAmelCase__ = vae_state_dict["decoder.norm_out.weight"] UpperCAmelCase__ = vae_state_dict["decoder.norm_out.bias"] UpperCAmelCase__ = vae_state_dict["quant_conv.weight"] UpperCAmelCase__ = vae_state_dict["quant_conv.bias"] UpperCAmelCase__ = vae_state_dict["post_quant_conv.weight"] UpperCAmelCase__ = vae_state_dict["post_quant_conv.bias"] # Retrieves the keys for the encoder down blocks only UpperCAmelCase__ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "encoder.down" in layer} ) UpperCAmelCase__ = { layer_id: [key for key in vae_state_dict if f"""down.{layer_id}""" in key] for layer_id in range(__A ) } # Retrieves the keys for the decoder up blocks only UpperCAmelCase__ = len({".".join(layer.split("." )[:3] ) for layer in vae_state_dict if "decoder.up" in layer} ) UpperCAmelCase__ = { layer_id: [key for key in vae_state_dict if f"""up.{layer_id}""" in key] for layer_id in range(__A ) } for i in range(__A ): UpperCAmelCase__ = [key for key in down_blocks[i] if f"""down.{i}""" in key and f"""down.{i}.downsample""" not in key] if f"""encoder.down.{i}.downsample.conv.weight""" in vae_state_dict: UpperCAmelCase__ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.weight""" ) UpperCAmelCase__ = vae_state_dict.pop( f"""encoder.down.{i}.downsample.conv.bias""" ) UpperCAmelCase__ = renew_vae_resnet_paths(__A ) UpperCAmelCase__ = {"old": f"""down.{i}.block""", "new": f"""down_blocks.{i}.resnets"""} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) UpperCAmelCase__ = [key for key in vae_state_dict if "encoder.mid.block" in key] UpperCAmelCase__ = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase__ = [key for key in mid_resnets if f"""encoder.mid.block_{i}""" in key] UpperCAmelCase__ = renew_vae_resnet_paths(__A ) UpperCAmelCase__ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) UpperCAmelCase__ = [key for key in vae_state_dict if "encoder.mid.attn" in key] UpperCAmelCase__ = renew_vae_attention_paths(__A ) UpperCAmelCase__ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) conv_attn_to_linear(__A ) for i in range(__A ): UpperCAmelCase__ = num_up_blocks - 1 - i UpperCAmelCase__ = [ key for key in up_blocks[block_id] if f"""up.{block_id}""" in key and f"""up.{block_id}.upsample""" not in key ] if f"""decoder.up.{block_id}.upsample.conv.weight""" in vae_state_dict: UpperCAmelCase__ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.weight""" ] UpperCAmelCase__ = vae_state_dict[ f"""decoder.up.{block_id}.upsample.conv.bias""" ] UpperCAmelCase__ = renew_vae_resnet_paths(__A ) UpperCAmelCase__ = {"old": f"""up.{block_id}.block""", "new": f"""up_blocks.{i}.resnets"""} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) UpperCAmelCase__ = [key for key in vae_state_dict if "decoder.mid.block" in key] UpperCAmelCase__ = 2 for i in range(1, num_mid_res_blocks + 1 ): UpperCAmelCase__ = [key for key in mid_resnets if f"""decoder.mid.block_{i}""" in key] UpperCAmelCase__ = renew_vae_resnet_paths(__A ) UpperCAmelCase__ = {"old": f"""mid.block_{i}""", "new": f"""mid_block.resnets.{i - 1}"""} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) UpperCAmelCase__ = [key for key in vae_state_dict if "decoder.mid.attn" in key] UpperCAmelCase__ = renew_vae_attention_paths(__A ) UpperCAmelCase__ = {"old": "mid.attn_1", "new": "mid_block.attentions.0"} assign_to_checkpoint(__A, __A, __A, additional_replacements=[meta_path], config=__A ) conv_attn_to_linear(__A ) return new_checkpoint def lowerCAmelCase_ ( __A, __A, ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = requests.get( " https://raw.githubusercontent.com/CompVis/stable-diffusion/main/configs/stable-diffusion/v1-inference.yaml" ) UpperCAmelCase__ = io.BytesIO(r.content ) UpperCAmelCase__ = OmegaConf.load(__A ) UpperCAmelCase__ = 512 UpperCAmelCase__ = "cuda" if torch.cuda.is_available() else "cpu" if checkpoint_path.endswith("safetensors" ): from safetensors import safe_open UpperCAmelCase__ = {} with safe_open(__A, framework="pt", device="cpu" ) as f: for key in f.keys(): UpperCAmelCase__ = f.get_tensor(__A ) else: UpperCAmelCase__ = torch.load(__A, map_location=__A )["state_dict"] # Convert the VAE model. UpperCAmelCase__ = create_vae_diffusers_config(__A, image_size=__A ) UpperCAmelCase__ = custom_convert_ldm_vae_checkpoint(__A, __A ) UpperCAmelCase__ = AutoencoderKL(**__A ) vae.load_state_dict(__A ) vae.save_pretrained(__A ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--vae_pt_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') parser.add_argument('--dump_path', default=None, type=str, required=True, help='Path to the VAE.pt to convert.') UpperCamelCase__ = parser.parse_args() vae_pt_to_vae_diffuser(args.vae_pt_path, args.dump_path)
65
from __future__ import annotations class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : list[list[int]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed from a list of zero or more lists containing at " "least one and the same number of values, each of which must be of type " "int or float." ) if len(__UpperCAmelCase ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__UpperCAmelCase ) != cols: raise error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def lowercase_ (self : Any ) -> list[list[int]]: """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def lowercase_ (self : Any ) -> int: """simple docstring""" return len(self.rows ) @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" return len(self.rows[0] ) @property def lowercase_ (self : List[Any] ) -> tuple[int, int]: """simple docstring""" return (self.num_rows, self.num_columns) @property def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return self.order[0] == self.order[1] def lowercase_ (self : Any ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : int ) -> int: """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return bool(self.determinant() ) def lowercase_ (self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__UpperCAmelCase ).determinant() def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) return -1 * self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Matrix: """simple docstring""" return Matrix( [ [self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def lowercase_ (self : List[str] ) -> Matrix: """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def lowercase_ (self : Optional[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError("Only matrices with a non-zero determinant have an inverse" ) return self.adjugate() * (1 / determinant) def __repr__(self : Dict ) -> str: """simple docstring""" return str(self.rows ) def __str__(self : Optional[Any] ) -> str: """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ "[" + ". ".join([str(__UpperCAmelCase ) for value in row] ) + ".]" for row in self.rows ] ) + "]" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError("Row must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_columns: raise ValueError( "Row must be equal in length to the other rows in the matrix" ) if position is None: self.rows.append(__UpperCAmelCase ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError( "Column must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in column: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_rows: raise ValueError( "Column must be equal in length to the other columns in the matrix" ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__(self : Any , __UpperCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return NotImplemented return self.rows == other.rows def __ne__(self : int , __UpperCAmelCase : object ) -> bool: """simple docstring""" return not self == other def __neg__(self : Dict ) -> Matrix: """simple docstring""" return self * -1 def __add__(self : Dict , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Addition requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__(self : Optional[Any] , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Subtraction requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__(self : Tuple , __UpperCAmelCase : Matrix | int | float ) -> Matrix: """simple docstring""" if isinstance(__UpperCAmelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): if self.num_columns != other.num_rows: raise ValueError( "The number of columns in the first matrix must " "be equal to the number of rows in the second" ) return Matrix( [ [Matrix.dot_product(__UpperCAmelCase , __UpperCAmelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( "A Matrix can only be multiplied by an int, float, or another matrix" ) def __pow__(self : List[Any] , __UpperCAmelCase : int ) -> Matrix: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("A Matrix can only be raised to the power of an int" ) if not self.is_square: raise ValueError("Only square matrices can be raised to a power" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( "Only invertable matrices can be raised to a negative power" ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : list[int] , __UpperCAmelCase : list[int] ) -> int: """simple docstring""" return sum(row[i] * column[i] for i in range(len(__UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
1
from typing import TYPE_CHECKING from ...utils import ( OptionalDependencyNotAvailable, _LazyModule, is_tf_available, is_tokenizers_available, is_torch_available, ) UpperCamelCase__ = { 'configuration_convbert': ['CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'ConvBertConfig', 'ConvBertOnnxConfig'], 'tokenization_convbert': ['ConvBertTokenizer'], } try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = ['ConvBertTokenizerFast'] try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'ConvBertForMaskedLM', 'ConvBertForMultipleChoice', 'ConvBertForQuestionAnswering', 'ConvBertForSequenceClassification', 'ConvBertForTokenClassification', 'ConvBertLayer', 'ConvBertModel', 'ConvBertPreTrainedModel', 'load_tf_weights_in_convbert', ] try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST', 'TFConvBertForMaskedLM', 'TFConvBertForMultipleChoice', 'TFConvBertForQuestionAnswering', 'TFConvBertForSequenceClassification', 'TFConvBertForTokenClassification', 'TFConvBertLayer', 'TFConvBertModel', 'TFConvBertPreTrainedModel', ] if TYPE_CHECKING: from .configuration_convbert import CONVBERT_PRETRAINED_CONFIG_ARCHIVE_MAP, ConvBertConfig, ConvBertOnnxConfig from .tokenization_convbert import ConvBertTokenizer try: if not is_tokenizers_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .tokenization_convbert_fast import ConvBertTokenizerFast try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_convbert import ( CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, ConvBertForMaskedLM, ConvBertForMultipleChoice, ConvBertForQuestionAnswering, ConvBertForSequenceClassification, ConvBertForTokenClassification, ConvBertLayer, ConvBertModel, ConvBertPreTrainedModel, load_tf_weights_in_convbert, ) try: if not is_tf_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_tf_convbert import ( TF_CONVBERT_PRETRAINED_MODEL_ARCHIVE_LIST, TFConvBertForMaskedLM, TFConvBertForMultipleChoice, TFConvBertForQuestionAnswering, TFConvBertForSequenceClassification, TFConvBertForTokenClassification, TFConvBertLayer, TFConvBertModel, TFConvBertPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
1
import unittest from pathlib import Path from tempfile import TemporaryDirectory from transformers import AutoConfig, TFAutoModel, is_tensorflow_text_available, is_tf_available from transformers.models.bert.tokenization_bert import BertTokenizer from transformers.testing_utils import require_tensorflow_text, require_tf, slow if is_tf_available(): import tensorflow as tf if is_tensorflow_text_available(): from transformers.models.bert import TFBertTokenizer UpperCamelCase__ = ['bert-base-uncased', 'bert-base-cased'] UpperCamelCase__ = 'hf-internal-testing/tiny-bert-tf-only' if is_tf_available(): class A ( tf.keras.Model ): def __init__(self : Any , __UpperCAmelCase : Tuple ) -> Optional[Any]: """simple docstring""" super().__init__() UpperCAmelCase__ = tokenizer UpperCAmelCase__ = AutoConfig.from_pretrained(__UpperCAmelCase ) UpperCAmelCase__ = TFAutoModel.from_config(__UpperCAmelCase ) def lowercase_ (self : int , __UpperCAmelCase : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = self.tokenizer(__UpperCAmelCase ) UpperCAmelCase__ = self.bert(**__UpperCAmelCase ) return out["pooler_output"] @require_tf @require_tensorflow_text class A ( unittest.TestCase ): def lowercase_ (self : List[str] ) -> List[str]: """simple docstring""" super().setUp() UpperCAmelCase__ = [ BertTokenizer.from_pretrained(__UpperCAmelCase ) for checkpoint in (TOKENIZER_CHECKPOINTS * 2) ] # repeat for when fast_bert_tokenizer=false UpperCAmelCase__ = [TFBertTokenizer.from_pretrained(__UpperCAmelCase ) for checkpoint in TOKENIZER_CHECKPOINTS] + [ TFBertTokenizer.from_pretrained(__UpperCAmelCase , use_fast_bert_tokenizer=__UpperCAmelCase ) for checkpoint in TOKENIZER_CHECKPOINTS ] assert len(self.tokenizers ) == len(self.tf_tokenizers ) UpperCAmelCase__ = [ "This is a straightforward English test sentence.", "This one has some weird characters\rto\nsee\r\nif those\u00E9break things.", "Now we're going to add some Chinese: 一 二 三 一二三", "And some much more rare Chinese: 齉 堃 齉堃", "Je vais aussi écrire en français pour tester les accents", "Classical Irish also has some unusual characters, so in they go: Gaelaċ, ꝼ", ] UpperCAmelCase__ = list(zip(self.test_sentences , self.test_sentences[::-1] ) ) def lowercase_ (self : Any ) -> int: """simple docstring""" for tokenizer, tf_tokenizer in zip(self.tokenizers , self.tf_tokenizers ): for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase__ = tokenizer(__UpperCAmelCase , return_tensors="tf" , padding="longest" ) UpperCAmelCase__ = tf_tokenizer(__UpperCAmelCase ) for key in python_outputs.keys(): self.assertTrue(tf.reduce_all(python_outputs[key].shape == tf_outputs[key].shape ) ) self.assertTrue(tf.reduce_all(tf.cast(python_outputs[key] , tf.intaa ) == tf_outputs[key] ) ) @slow def lowercase_ (self : List[Any] ) -> List[Any]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = tf_tokenizer(self.paired_sentences ) UpperCAmelCase__ = tf_tokenizer( text=[sentence[0] for sentence in self.paired_sentences] , text_pair=[sentence[1] for sentence in self.paired_sentences] , ) for key in merged_outputs.keys(): self.assertTrue(tf.reduce_all(tf.cast(merged_outputs[key] , tf.intaa ) == separated_outputs[key] ) ) @slow def lowercase_ (self : Union[str, Any] ) -> Optional[int]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = tf.function(__UpperCAmelCase ) for test_inputs in (self.test_sentences, self.paired_sentences): UpperCAmelCase__ = tf.constant(__UpperCAmelCase ) UpperCAmelCase__ = compiled_tokenizer(__UpperCAmelCase ) UpperCAmelCase__ = tf_tokenizer(__UpperCAmelCase ) for key in eager_outputs.keys(): self.assertTrue(tf.reduce_all(eager_outputs[key] == compiled_outputs[key] ) ) @slow def lowercase_ (self : str ) -> List[str]: """simple docstring""" for tf_tokenizer in self.tf_tokenizers: UpperCAmelCase__ = ModelToSave(tokenizer=__UpperCAmelCase ) UpperCAmelCase__ = tf.convert_to_tensor(self.test_sentences ) UpperCAmelCase__ = model(__UpperCAmelCase ) # Build model with some sample inputs with TemporaryDirectory() as tempdir: UpperCAmelCase__ = Path(__UpperCAmelCase ) / "saved.model" model.save(__UpperCAmelCase ) UpperCAmelCase__ = tf.keras.models.load_model(__UpperCAmelCase ) UpperCAmelCase__ = loaded_model(__UpperCAmelCase ) # We may see small differences because the loaded model is compiled, so we need an epsilon for the test self.assertLessEqual(tf.reduce_max(tf.abs(out - loaded_output ) ) , 1E-5 )
65
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor class A ( nn.Module ): def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , __UpperCAmelCase : int=(6_4,) , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Any=3_2 , __UpperCAmelCase : str="silu" , __UpperCAmelCase : Any=True , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = torch.nn.Convad( __UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) # down UpperCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_down_block( __UpperCAmelCase , num_layers=self.layers_per_block , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) self.down_blocks.append(__UpperCAmelCase ) # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # out UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = 2 * out_channels if double_z else out_channels UpperCAmelCase__ = nn.Convad(block_out_channels[-1] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : List[Any] , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = x UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : int ): def custom_forward(*__UpperCAmelCase : Optional[Any] ): return module(*__UpperCAmelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __UpperCAmelCase ) else: # down for down_block in self.down_blocks: UpperCAmelCase__ = down_block(__UpperCAmelCase ) # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase ) # post-process UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : List[Any] , __UpperCAmelCase : str=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=("UpDecoderBlock2D",) , __UpperCAmelCase : str=(6_4,) , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : Any="silu" , __UpperCAmelCase : Any="group" , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) UpperCAmelCase__ = in_channels if norm_type == "spatial" else None # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # up UpperCAmelCase__ = list(reversed(__UpperCAmelCase ) ) UpperCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = reversed_block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_up_block( __UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , prev_output_channel=__UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , resnet_time_scale_shift=__UpperCAmelCase , ) self.up_blocks.append(__UpperCAmelCase ) UpperCAmelCase__ = output_channel # out if norm_type == "spatial": UpperCAmelCase__ = SpatialNorm(block_out_channels[0] , __UpperCAmelCase ) else: UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = nn.Convad(block_out_channels[0] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = z UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) UpperCAmelCase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : str ): def custom_forward(*__UpperCAmelCase : List[str] ): return module(*__UpperCAmelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase ) else: # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = up_block(__UpperCAmelCase , __UpperCAmelCase ) # post-process if latent_embeds is None: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) else: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Union[str, Any]="random" , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=True ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = n_e UpperCAmelCase__ = vq_embed_dim UpperCAmelCase__ = beta UpperCAmelCase__ = legacy UpperCAmelCase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) UpperCAmelCase__ = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) UpperCAmelCase__ = self.used.shape[0] UpperCAmelCase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": UpperCAmelCase__ = self.re_embed UpperCAmelCase__ = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: UpperCAmelCase__ = n_e UpperCAmelCase__ = sane_index_shape def lowercase_ (self : str , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) UpperCAmelCase__ = (inds[:, :, None] == used[None, None, ...]).long() UpperCAmelCase__ = match.argmax(-1 ) UpperCAmelCase__ = match.sum(2 ) < 1 if self.unknown_index == "random": UpperCAmelCase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: UpperCAmelCase__ = self.unknown_index return new.reshape(__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token UpperCAmelCase__ = 0 # simply set to zero UpperCAmelCase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __UpperCAmelCase ) return back.reshape(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() UpperCAmelCase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z UpperCAmelCase__ = torch.argmin(torch.cdist(__UpperCAmelCase , self.embedding.weight ) , dim=1 ) UpperCAmelCase__ = self.embedding(__UpperCAmelCase ).view(z.shape ) UpperCAmelCase__ = None UpperCAmelCase__ = None # compute loss for embedding if not self.legacy: UpperCAmelCase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: UpperCAmelCase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients UpperCAmelCase__ = z + (z_q - z).detach() # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: UpperCAmelCase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.remap_to_used(__UpperCAmelCase ) UpperCAmelCase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: UpperCAmelCase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Any: """simple docstring""" if self.remap is not None: UpperCAmelCase__ = indices.reshape(shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.unmap_to_all(__UpperCAmelCase ) UpperCAmelCase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors UpperCAmelCase__ = self.embedding(__UpperCAmelCase ) if shape is not None: UpperCAmelCase__ = z_q.view(__UpperCAmelCase ) # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parameters UpperCAmelCase__ , UpperCAmelCase__ = torch.chunk(__UpperCAmelCase , 2 , dim=1 ) UpperCAmelCase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) UpperCAmelCase__ = deterministic UpperCAmelCase__ = torch.exp(0.5 * self.logvar ) UpperCAmelCase__ = torch.exp(self.logvar ) if self.deterministic: UpperCAmelCase__ = UpperCAmelCase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = randn_tensor( self.mean.shape , generator=__UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) UpperCAmelCase__ = self.mean + self.std * sample return x def lowercase_ (self : str , __UpperCAmelCase : int=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase_ (self : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=[1, 2, 3] ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) UpperCAmelCase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" return self.mean
65
1
from math import ceil def lowerCAmelCase_ ( __A = 1_001 ) -> int: '''simple docstring''' UpperCAmelCase__ = 1 for i in range(1, int(ceil(n / 2.0 ) ) ): UpperCAmelCase__ = 2 * i + 1 UpperCAmelCase__ = 2 * i UpperCAmelCase__ = total + 4 * odd**2 - 6 * even return total if __name__ == "__main__": import sys if len(sys.argv) == 1: print(solution()) else: try: UpperCamelCase__ = int(sys.argv[1]) print(solution(n)) except ValueError: print('Invalid entry - please enter a number')
65
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
1
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' def get_matched_characters(__A, __A ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase__ = int(max(0, i - limit ) ) UpperCAmelCase__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(__A ) UpperCAmelCase__ = f"""{_stra[0:_stra.index(__A )]} {_stra[_stra.index(__A ) + 1:]}""" return "".join(__A ) # matching characters UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = len(__A ) # transposition UpperCAmelCase__ = ( len([(ca, ca) for ca, ca in zip(__A, __A ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase__ = 0.0 else: UpperCAmelCase__ = ( 1 / 3 * ( match_count / len(__A ) + match_count / len(__A ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
65
1
from dataclasses import asdict, dataclass from typing import Optional from ...configuration_utils import PretrainedConfig from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) # TODO Update this UpperCamelCase__ = { 'facebook/esm-1b': 'https://huggingface.co/facebook/esm-1b/resolve/main/config.json', # See all ESM models at https://huggingface.co/models?filter=esm } class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = 'esm' def __init__(self : List[str] , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : str=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=7_6_8 , __UpperCAmelCase : List[str]=1_2 , __UpperCAmelCase : Any=1_2 , __UpperCAmelCase : List[str]=3_0_7_2 , __UpperCAmelCase : Any=0.1 , __UpperCAmelCase : Any=0.1 , __UpperCAmelCase : int=1_0_2_6 , __UpperCAmelCase : Optional[int]=0.02 , __UpperCAmelCase : Optional[Any]=1E-12 , __UpperCAmelCase : int="absolute" , __UpperCAmelCase : Optional[Any]=True , __UpperCAmelCase : Optional[Any]=None , __UpperCAmelCase : str=False , __UpperCAmelCase : Dict=False , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Optional[Any]=None , **__UpperCAmelCase : Union[str, Any] , ) -> Any: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , mask_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = position_embedding_type UpperCAmelCase__ = use_cache UpperCAmelCase__ = emb_layer_norm_before UpperCAmelCase__ = token_dropout UpperCAmelCase__ = is_folding_model if is_folding_model: if esmfold_config is None: logger.info("No esmfold_config supplied for folding model, using default values." ) UpperCAmelCase__ = EsmFoldConfig() elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): UpperCAmelCase__ = EsmFoldConfig(**__UpperCAmelCase ) UpperCAmelCase__ = esmfold_config if vocab_list is None: logger.warning("No vocab_list supplied for folding model, assuming the ESM-2 vocabulary!" ) UpperCAmelCase__ = get_default_vocab_list() else: UpperCAmelCase__ = vocab_list else: UpperCAmelCase__ = None UpperCAmelCase__ = None if self.esmfold_config is not None and getattr(self.esmfold_config , "use_esm_attn_map" , __UpperCAmelCase ): raise ValueError("The HuggingFace port of ESMFold does not support use_esm_attn_map at this time!" ) def lowercase_ (self : Tuple ) -> Dict: """simple docstring""" UpperCAmelCase__ = super().to_dict() if isinstance(self.esmfold_config , __UpperCAmelCase ): UpperCAmelCase__ = self.esmfold_config.to_dict() return output @dataclass class A : __UpperCAmelCase : str = None __UpperCAmelCase : bool = True __UpperCAmelCase : bool = False __UpperCAmelCase : bool = False __UpperCAmelCase : bool = False __UpperCAmelCase : float = 0 __UpperCAmelCase : bool = True __UpperCAmelCase : bool = False __UpperCAmelCase : int = 1_28 __UpperCAmelCase : "TrunkConfig" = None def lowercase_ (self : Optional[Any] ) -> List[Any]: """simple docstring""" if self.trunk is None: UpperCAmelCase__ = TrunkConfig() elif isinstance(self.trunk , __UpperCAmelCase ): UpperCAmelCase__ = TrunkConfig(**self.trunk ) def lowercase_ (self : str ) -> Tuple: """simple docstring""" UpperCAmelCase__ = asdict(self ) UpperCAmelCase__ = self.trunk.to_dict() return output @dataclass class A : __UpperCAmelCase : int = 48 __UpperCAmelCase : int = 10_24 __UpperCAmelCase : int = 1_28 __UpperCAmelCase : int = 32 __UpperCAmelCase : int = 32 __UpperCAmelCase : int = 32 __UpperCAmelCase : float = 0 __UpperCAmelCase : float = 0 __UpperCAmelCase : bool = False __UpperCAmelCase : int = 4 __UpperCAmelCase : Optional[int] = 1_28 __UpperCAmelCase : "StructureModuleConfig" = None def lowercase_ (self : Optional[int] ) -> Dict: """simple docstring""" if self.structure_module is None: UpperCAmelCase__ = StructureModuleConfig() elif isinstance(self.structure_module , __UpperCAmelCase ): UpperCAmelCase__ = StructureModuleConfig(**self.structure_module ) if self.max_recycles <= 0: raise ValueError(f"""`max_recycles` should be positive, got {self.max_recycles}.""" ) if self.sequence_state_dim % self.sequence_state_dim != 0: raise ValueError( "`sequence_state_dim` should be a round multiple of `sequence_state_dim`, got" f""" {self.sequence_state_dim} and {self.sequence_state_dim}.""" ) if self.pairwise_state_dim % self.pairwise_state_dim != 0: raise ValueError( "`pairwise_state_dim` should be a round multiple of `pairwise_state_dim`, got" f""" {self.pairwise_state_dim} and {self.pairwise_state_dim}.""" ) UpperCAmelCase__ = self.sequence_state_dim // self.sequence_head_width UpperCAmelCase__ = self.pairwise_state_dim // self.pairwise_head_width if self.sequence_state_dim != sequence_num_heads * self.sequence_head_width: raise ValueError( "`sequence_state_dim` should be equal to `sequence_num_heads * sequence_head_width, got" f""" {self.sequence_state_dim} != {sequence_num_heads} * {self.sequence_head_width}.""" ) if self.pairwise_state_dim != pairwise_num_heads * self.pairwise_head_width: raise ValueError( "`pairwise_state_dim` should be equal to `pairwise_num_heads * pairwise_head_width, got" f""" {self.pairwise_state_dim} != {pairwise_num_heads} * {self.pairwise_head_width}.""" ) if self.pairwise_state_dim % 2 != 0: raise ValueError(f"""`pairwise_state_dim` should be even, got {self.pairwise_state_dim}.""" ) if self.dropout >= 0.4: raise ValueError(f"""`dropout` should not be greater than 0.4, got {self.dropout}.""" ) def lowercase_ (self : Any ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = asdict(self ) UpperCAmelCase__ = self.structure_module.to_dict() return output @dataclass class A : __UpperCAmelCase : int = 3_84 __UpperCAmelCase : int = 1_28 __UpperCAmelCase : int = 16 __UpperCAmelCase : int = 1_28 __UpperCAmelCase : int = 12 __UpperCAmelCase : int = 4 __UpperCAmelCase : int = 8 __UpperCAmelCase : float = 0.1 __UpperCAmelCase : int = 8 __UpperCAmelCase : int = 1 __UpperCAmelCase : int = 2 __UpperCAmelCase : int = 7 __UpperCAmelCase : int = 10 __UpperCAmelCase : float = 1E-8 __UpperCAmelCase : float = 1E5 def lowercase_ (self : Optional[Any] ) -> str: """simple docstring""" return asdict(self ) def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' return ( "<cls>", "<pad>", "<eos>", "<unk>", "L", "A", "G", "V", "S", "E", "R", "T", "I", "D", "P", "K", "Q", "N", "F", "Y", "M", "H", "W", "C", "X", "B", "U", "Z", "O", ".", "-", "<null_1>", "<mask>", )
65
def lowerCAmelCase_ ( __A, __A ) -> None: '''simple docstring''' UpperCAmelCase__ = len(__A ) print("The following activities are selected:" ) # The first activity is always selected UpperCAmelCase__ = 0 print(__A, end="," ) # Consider rest of the activities for j in range(__A ): # If this activity has start time greater than # or equal to the finish time of previously # selected activity, then select it if start[j] >= finish[i]: print(__A, end="," ) UpperCAmelCase__ = j if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = [1, 3, 0, 5, 8, 5] UpperCamelCase__ = [2, 4, 6, 7, 9, 9] print_max_activities(start, finish)
65
1
# Copyright 2022 The HuggingFace Team and The OpenBMB Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from typing import TYPE_CHECKING # rely on isort to merge the imports from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase__ = { 'configuration_cpmant': ['CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'CpmAntConfig'], 'tokenization_cpmant': ['CpmAntTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST', 'CpmAntForCausalLM', 'CpmAntModel', 'CpmAntPreTrainedModel', ] if TYPE_CHECKING: from .configuration_cpmant import CPMANT_PRETRAINED_CONFIG_ARCHIVE_MAP, CpmAntConfig from .tokenization_cpmant import CpmAntTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_cpmant import ( CPMANT_PRETRAINED_MODEL_ARCHIVE_LIST, CpmAntForCausalLM, CpmAntModel, CpmAntPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
import argparse import os import jax as jnp import numpy as onp import torch import torch.nn as nn from music_spectrogram_diffusion import inference from tax import checkpoints from diffusers import DDPMScheduler, OnnxRuntimeModel, SpectrogramDiffusionPipeline from diffusers.pipelines.spectrogram_diffusion import SpectrogramContEncoder, SpectrogramNotesEncoder, TaFilmDecoder UpperCamelCase__ = 'base_with_context' def lowerCAmelCase_ ( __A, __A ) -> int: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["token_embedder"]["embedding"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["input_proj"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) for lyr_num, lyr in enumerate(model.encoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = ly_weight["attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["encoder_norm"]["scale"] ) ) return model def lowerCAmelCase_ ( __A, __A ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["time_emb_dense1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["Embed_0"]["embedding"] ), requires_grad=__A ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(weights["continuous_inputs_projection"]["kernel"].T ) ) for lyr_num, lyr in enumerate(model.decoders ): UpperCAmelCase__ = weights[f"""layers_{lyr_num}"""] UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_self_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_0"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["self_attention"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = ly_weight["MultiHeadDotProductAttention_0"] UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["query"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["key"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["value"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(attention_weights["out"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["pre_cross_attention_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["pre_mlp_layer_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter( torch.FloatTensor(ly_weight["FiLMLayer_1"]["DenseGeneral_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_0"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wi_1"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(ly_weight["mlp"]["wo"]["kernel"].T ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["decoder_norm"]["scale"] ) ) UpperCAmelCase__ = nn.Parameter(torch.FloatTensor(weights["spec_out_dense"]["kernel"].T ) ) return model def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = checkpoints.load_tax_checkpoint(args.checkpoint_path ) UpperCAmelCase__ = jnp.tree_util.tree_map(onp.array, __A ) UpperCAmelCase__ = [ "from __gin__ import dynamic_registration", "from music_spectrogram_diffusion.models.diffusion import diffusion_utils", "diffusion_utils.ClassifierFreeGuidanceConfig.eval_condition_weight = 2.0", "diffusion_utils.DiffusionConfig.classifier_free_guidance = @diffusion_utils.ClassifierFreeGuidanceConfig()", ] UpperCAmelCase__ = os.path.join(args.checkpoint_path, "..", "config.gin" ) UpperCAmelCase__ = inference.parse_training_gin_file(__A, __A ) UpperCAmelCase__ = inference.InferenceModel(args.checkpoint_path, __A ) UpperCAmelCase__ = DDPMScheduler(beta_schedule="squaredcos_cap_v2", variance_type="fixed_large" ) UpperCAmelCase__ = SpectrogramNotesEncoder( max_length=synth_model.sequence_length["inputs"], vocab_size=synth_model.model.module.config.vocab_size, d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = SpectrogramContEncoder( input_dims=synth_model.audio_codec.n_dims, targets_context_length=synth_model.sequence_length["targets_context"], d_model=synth_model.model.module.config.emb_dim, dropout_rate=synth_model.model.module.config.dropout_rate, num_layers=synth_model.model.module.config.num_encoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, feed_forward_proj="gated-gelu", ) UpperCAmelCase__ = TaFilmDecoder( input_dims=synth_model.audio_codec.n_dims, targets_length=synth_model.sequence_length["targets_context"], max_decoder_noise_time=synth_model.model.module.config.max_decoder_noise_time, d_model=synth_model.model.module.config.emb_dim, num_layers=synth_model.model.module.config.num_decoder_layers, num_heads=synth_model.model.module.config.num_heads, d_kv=synth_model.model.module.config.head_dim, d_ff=synth_model.model.module.config.mlp_dim, dropout_rate=synth_model.model.module.config.dropout_rate, ) UpperCAmelCase__ = load_notes_encoder(ta_checkpoint["target"]["token_encoder"], __A ) UpperCAmelCase__ = load_continuous_encoder(ta_checkpoint["target"]["continuous_encoder"], __A ) UpperCAmelCase__ = load_decoder(ta_checkpoint["target"]["decoder"], __A ) UpperCAmelCase__ = OnnxRuntimeModel.from_pretrained("kashif/soundstream_mel_decoder" ) UpperCAmelCase__ = SpectrogramDiffusionPipeline( notes_encoder=__A, continuous_encoder=__A, decoder=__A, scheduler=__A, melgan=__A, ) if args.save: pipe.save_pretrained(args.output_path ) if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument('--output_path', default=None, type=str, required=True, help='Path to the converted model.') parser.add_argument( '--save', default=True, type=bool, required=False, help='Whether to save the converted model or not.' ) parser.add_argument( '--checkpoint_path', default=f'''{MODEL}/checkpoint_500000''', type=str, required=False, help='Path to the original jax model checkpoint.', ) UpperCamelCase__ = parser.parse_args() main(args)
65
1
from __future__ import annotations import math def lowerCAmelCase_ ( __A ) -> list[int]: '''simple docstring''' if num <= 0: UpperCAmelCase__ = f"""{num}: Invalid input, please enter a positive integer.""" raise ValueError(__A ) UpperCAmelCase__ = [True] * (num + 1) UpperCAmelCase__ = [] UpperCAmelCase__ = 2 UpperCAmelCase__ = int(math.sqrt(__A ) ) while start <= end: # If start is a prime if sieve[start] is True: prime.append(__A ) # Set multiples of start be False for i in range(start * start, num + 1, __A ): if sieve[i] is True: UpperCAmelCase__ = False start += 1 for j in range(end + 1, num + 1 ): if sieve[j] is True: prime.append(__A ) return prime if __name__ == "__main__": print(prime_sieve(int(input('Enter a positive integer: ').strip())))
65
import math def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' return math.sqrt(__A ) * math.sqrt(__A ) == num def lowerCAmelCase_ ( __A ) -> bool: '''simple docstring''' UpperCAmelCase__ = 0 UpperCAmelCase__ = n while left <= right: UpperCAmelCase__ = (left + right) // 2 if mid**2 == n: return True elif mid**2 > n: UpperCAmelCase__ = mid - 1 else: UpperCAmelCase__ = mid + 1 return False if __name__ == "__main__": import doctest doctest.testmod()
65
1
import argparse import json import os from pathlib import Path import requests import torch from transformers import JukeboxConfig, JukeboxModel from transformers.utils import logging logging.set_verbosity_info() UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = 'https://openaipublic.azureedge.net/jukebox/models/' UpperCamelCase__ = { 'jukebox-1b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '1b_lyrics/prior_level_2.pth.tar', ], 'jukebox-5b-lyrics': [ '5b/vqvae.pth.tar', '5b/prior_level_0.pth.tar', '5b/prior_level_1.pth.tar', '5b_lyrics/prior_level_2.pth.tar', ], } def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if key.endswith(".model.1.bias" ) and len(key.split("." ) ) > 10: UpperCAmelCase__ = key.replace(".model.1.bias", ".conv1d_1.bias" ) elif key.endswith(".model.1.weight" ) and len(key.split("." ) ) > 10: UpperCAmelCase__ = key.replace(".model.1.weight", ".conv1d_1.weight" ) elif key.endswith(".model.3.bias" ) and len(key.split("." ) ) > 10: UpperCAmelCase__ = key.replace(".model.3.bias", ".conv1d_2.bias" ) elif key.endswith(".model.3.weight" ) and len(key.split("." ) ) > 10: UpperCAmelCase__ = key.replace(".model.3.weight", ".conv1d_2.weight" ) if "conditioner_blocks.0." in key: UpperCAmelCase__ = key.replace("conditioner_blocks.0", "conditioner_blocks" ) if "prime_prior" in key: UpperCAmelCase__ = key.replace("prime_prior", "encoder" ) if ".emb." in key and "total" not in key and "absolute" not in key and "relative" not in key: UpperCAmelCase__ = key.replace(".emb.", "." ) if key.endswith("k" ): # replace vqvae.X.k with vqvae.X.codebook return key.replace(".k", ".codebook" ) if "y_emb." in key: return key.replace("y_emb.", "metadata_embedding." ) if "x_emb.emb." in key: UpperCAmelCase__ = key.replace("0.x_emb.emb", "embed_tokens" ) if "prime_state_ln" in key: return key.replace("prime_state_ln", "encoder.final_layer_norm" ) if ".ln" in key: return key.replace(".ln", ".layer_norm" ) if "_ln" in key: return key.replace("_ln", "_layer_norm" ) if "prime_state_proj" in key: return key.replace("prime_state_proj", "encoder.proj_in" ) if "prime_x_out" in key: return key.replace("prime_x_out", "encoder.lm_head" ) if "prior.x_out" in key: return key.replace("x_out", "fc_proj_out" ) if "x_emb" in key: return key.replace("x_emb", "embed_tokens" ) return key def lowerCAmelCase_ ( __A, __A, __A, __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = {} import re UpperCAmelCase__ = re.compile(r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase__ = re.compile( r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase__ = re.compile(r"encoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase__ = re.compile(r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase__ = re.compile( r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase__ = re.compile(r"decoders.(\d*).level_blocks.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase__ = re.compile(r"conditioner_blocks.(\d*).cond.model.(\d*).(\d).(bias|weight)" ) UpperCAmelCase__ = re.compile( r"conditioner_blocks.(\d*).cond.model.(\d*).(\d).model.(\d*).model.(\d*).(bias|weight)" ) UpperCAmelCase__ = re.compile(r"conditioner_blocks.(\d*).cond.model.(\d*).(bias|weight)" ) for original_key, value in state_dict.items(): # rename vqvae.encoder keys if re_encoder_block_conv_in.fullmatch(__A ): UpperCAmelCase__ = re_encoder_block_conv_in.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) UpperCAmelCase__ = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.{groups[-1]}""" UpperCAmelCase__ = re_encoder_block_conv_in.sub(__A, __A ) elif re_encoder_block_resnet.fullmatch(__A ): UpperCAmelCase__ = re_encoder_block_resnet.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) UpperCAmelCase__ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase__ = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.downsample_block.{block_index}.""" UpperCAmelCase__ = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" UpperCAmelCase__ = prefix + resnet_block UpperCAmelCase__ = re_encoder_block_resnet.sub(__A, __A ) elif re_encoder_block_proj_out.fullmatch(__A ): UpperCAmelCase__ = re_encoder_block_proj_out.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = f"""encoders.{groups[0]}.level_blocks.{groups[1]}.proj_out.{groups[-1]}""" UpperCAmelCase__ = re_encoder_block_proj_out.sub(__A, __A ) # rename vqvae.decoder keys elif re_decoder_block_conv_out.fullmatch(__A ): UpperCAmelCase__ = re_decoder_block_conv_out.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) - 2 UpperCAmelCase__ = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.{groups[-1]}""" UpperCAmelCase__ = re_decoder_block_conv_out.sub(__A, __A ) elif re_decoder_block_resnet.fullmatch(__A ): UpperCAmelCase__ = re_decoder_block_resnet.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[2] ) * 2 + int(groups[3] ) - 2 UpperCAmelCase__ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase__ = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.upsample_block.{block_index}.""" UpperCAmelCase__ = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" UpperCAmelCase__ = prefix + resnet_block UpperCAmelCase__ = re_decoder_block_resnet.sub(__A, __A ) elif re_decoder_block_proj_in.fullmatch(__A ): UpperCAmelCase__ = re_decoder_block_proj_in.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = f"""decoders.{groups[0]}.level_blocks.{groups[1]}.proj_in.{groups[-1]}""" UpperCAmelCase__ = re_decoder_block_proj_in.sub(__A, __A ) # rename prior cond.model to upsampler.upsample_block and resnet elif re_prior_cond_conv_out.fullmatch(__A ): UpperCAmelCase__ = re_prior_cond_conv_out.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[1] ) * 2 + int(groups[2] ) - 2 UpperCAmelCase__ = f"""conditioner_blocks.upsampler.upsample_block.{block_index}.{groups[-1]}""" UpperCAmelCase__ = re_prior_cond_conv_out.sub(__A, __A ) elif re_prior_cond_resnet.fullmatch(__A ): UpperCAmelCase__ = re_prior_cond_resnet.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = int(groups[1] ) * 2 + int(groups[2] ) - 2 UpperCAmelCase__ = {"1": 1, "3": 2}[groups[-2]] UpperCAmelCase__ = f"""conditioner_blocks.upsampler.upsample_block.{block_index}.""" UpperCAmelCase__ = f"""resnet_block.{groups[-3]}.conv1d_{conv_index}.{groups[-1]}""" UpperCAmelCase__ = prefix + resnet_block UpperCAmelCase__ = re_prior_cond_resnet.sub(__A, __A ) elif re_prior_cond_proj_in.fullmatch(__A ): UpperCAmelCase__ = re_prior_cond_proj_in.match(__A ) UpperCAmelCase__ = regex_match.groups() UpperCAmelCase__ = f"""conditioner_blocks.upsampler.proj_in.{groups[-1]}""" UpperCAmelCase__ = re_prior_cond_proj_in.sub(__A, __A ) # keep original key else: UpperCAmelCase__ = original_key UpperCAmelCase__ = replace_key(__A ) if f"""{key_prefix}.{key}""" not in model_state_dict or key is None: print(f"""failed converting {original_key} to {key}, does not match""" ) # handle missmatched shape elif value.shape != model_state_dict[f"""{key_prefix}.{key}"""].shape: UpperCAmelCase__ = model_state_dict[f"""{key_prefix}.{key}"""] print(f"""{original_key}-> {key} : \nshape {val.shape} and { value.shape}, do not match""" ) UpperCAmelCase__ = original_key UpperCAmelCase__ = original_key UpperCAmelCase__ = value return new_dict @torch.no_grad() def lowerCAmelCase_ ( __A=None, __A=None ) -> Optional[Any]: '''simple docstring''' for file in MODEL_MAPPING[model_name]: if not os.path.isfile(f"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""" ): UpperCAmelCase__ = requests.get(f"""{PREFIX}{file}""", allow_redirects=__A ) os.makedirs(f"""{pytorch_dump_folder_path}/""", exist_ok=__A ) open(f"""{pytorch_dump_folder_path}/{file.split("/" )[-1]}""", "wb" ).write(r.content ) UpperCAmelCase__ = MODEL_MAPPING[model_name.split("/" )[-1]] UpperCAmelCase__ = JukeboxConfig.from_pretrained(__A ) UpperCAmelCase__ = JukeboxModel(__A ) UpperCAmelCase__ = [] UpperCAmelCase__ = {} for i, dict_name in enumerate(__A ): UpperCAmelCase__ = torch.load(f"""{pytorch_dump_folder_path}/{dict_name.split("/" )[-1]}""" )["model"] UpperCAmelCase__ = {} for k in old_dic.keys(): if k.endswith(".b" ): UpperCAmelCase__ = old_dic[k] elif k.endswith(".w" ): UpperCAmelCase__ = old_dic[k] elif "level_2" not in dict_name and "cond.model." in k: UpperCAmelCase__ = old_dic[k] else: UpperCAmelCase__ = old_dic[k] UpperCAmelCase__ = "vqvae" if i == 0 else f"""priors.{3 - i}""" UpperCAmelCase__ = fix_jukebox_keys(__A, model.state_dict(), __A, __A ) weight_dict.append(__A ) UpperCAmelCase__ = weight_dict.pop(0 ) model.vqvae.load_state_dict(__A ) for i in range(len(__A ) ): model.priors[i].load_state_dict(weight_dict[2 - i] ) Path(__A ).mkdir(exist_ok=__A ) with open(f"""{pytorch_dump_folder_path}/mapping.json""", "w" ) as txtfile: json.dump(__A, __A ) print(f"""Saving model {model_name} to {pytorch_dump_folder_path}""" ) model.save_pretrained(__A ) return weight_dict if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( '--model_name', default='jukebox-5b-lyrics', type=str, help='Name of the model you\'d like to convert.', ) parser.add_argument( '--pytorch_dump_folder_path', default='jukebox-5b-lyrics-converted', type=str, help='Path to the output PyTorch model directory.', ) UpperCamelCase__ = parser.parse_args() convert_openai_checkpoint(args.model_name, args.pytorch_dump_folder_path)
65
import math from dataclasses import dataclass from typing import Optional, Tuple, Union import numpy as np import torch from ..configuration_utils import ConfigMixin, register_to_config from ..utils import BaseOutput, randn_tensor from .scheduling_utils import SchedulerMixin @dataclass # Copied from diffusers.schedulers.scheduling_ddpm.DDPMSchedulerOutput with DDPM->UnCLIP class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor __UpperCAmelCase : Optional[torch.FloatTensor] = None def lowerCAmelCase_ ( __A, __A=0.999, __A="cosine", ) -> Tuple: '''simple docstring''' if alpha_transform_type == "cosine": def alpha_bar_fn(__A ): return math.cos((t + 0.008) / 1.008 * math.pi / 2 ) ** 2 elif alpha_transform_type == "exp": def alpha_bar_fn(__A ): return math.exp(t * -12.0 ) else: raise ValueError(f"""Unsupported alpha_tranform_type: {alpha_transform_type}""" ) UpperCAmelCase__ = [] for i in range(__A ): UpperCAmelCase__ = i / num_diffusion_timesteps UpperCAmelCase__ = (i + 1) / num_diffusion_timesteps betas.append(min(1 - alpha_bar_fn(__A ) / alpha_bar_fn(__A ), __A ) ) return torch.tensor(__A, dtype=torch.floataa ) class A ( UpperCAmelCase_ , UpperCAmelCase_ ): @register_to_config def __init__(self : List[str] , __UpperCAmelCase : int = 1_0_0_0 , __UpperCAmelCase : str = "fixed_small_log" , __UpperCAmelCase : bool = True , __UpperCAmelCase : Optional[float] = 1.0 , __UpperCAmelCase : str = "epsilon" , __UpperCAmelCase : str = "squaredcos_cap_v2" , ) -> Optional[int]: """simple docstring""" if beta_schedule != "squaredcos_cap_v2": raise ValueError("UnCLIPScheduler only supports `beta_schedule`: 'squaredcos_cap_v2'" ) UpperCAmelCase__ = betas_for_alpha_bar(__UpperCAmelCase ) UpperCAmelCase__ = 1.0 - self.betas UpperCAmelCase__ = torch.cumprod(self.alphas , dim=0 ) UpperCAmelCase__ = torch.tensor(1.0 ) # standard deviation of the initial noise distribution UpperCAmelCase__ = 1.0 # setable values UpperCAmelCase__ = None UpperCAmelCase__ = torch.from_numpy(np.arange(0 , __UpperCAmelCase )[::-1].copy() ) UpperCAmelCase__ = variance_type def lowercase_ (self : List[str] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None ) -> torch.FloatTensor: """simple docstring""" return sample def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : Union[str, torch.device] = None ) -> Any: """simple docstring""" UpperCAmelCase__ = num_inference_steps UpperCAmelCase__ = (self.config.num_train_timesteps - 1) / (self.num_inference_steps - 1) UpperCAmelCase__ = (np.arange(0 , __UpperCAmelCase ) * step_ratio).round()[::-1].copy().astype(np.intaa ) UpperCAmelCase__ = torch.from_numpy(__UpperCAmelCase ).to(__UpperCAmelCase ) def lowercase_ (self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : Optional[int]=None , __UpperCAmelCase : Tuple=None , __UpperCAmelCase : List[str]=None ) -> Tuple: """simple docstring""" if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev # For t > 0, compute predicted variance βt (see formula (6) and (7) from https://arxiv.org/pdf/2006.11239.pdf) # and sample from it to get previous sample # x_{t-1} ~ N(pred_prev_sample, variance) == add variance to pred_sample UpperCAmelCase__ = beta_prod_t_prev / beta_prod_t * beta if variance_type is None: UpperCAmelCase__ = self.config.variance_type # hacks - were probably added for training stability if variance_type == "fixed_small_log": UpperCAmelCase__ = torch.log(torch.clamp(__UpperCAmelCase , min=1E-20 ) ) UpperCAmelCase__ = torch.exp(0.5 * variance ) elif variance_type == "learned_range": # NOTE difference with DDPM scheduler UpperCAmelCase__ = variance.log() UpperCAmelCase__ = beta.log() UpperCAmelCase__ = (predicted_variance + 1) / 2 UpperCAmelCase__ = frac * max_log + (1 - frac) * min_log return variance def lowercase_ (self : Optional[int] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : int , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : bool = True , ) -> Union[UnCLIPSchedulerOutput, Tuple]: """simple docstring""" UpperCAmelCase__ = timestep if model_output.shape[1] == sample.shape[1] * 2 and self.variance_type == "learned_range": UpperCAmelCase__ , UpperCAmelCase__ = torch.split(__UpperCAmelCase , sample.shape[1] , dim=1 ) else: UpperCAmelCase__ = None # 1. compute alphas, betas if prev_timestep is None: UpperCAmelCase__ = t - 1 UpperCAmelCase__ = self.alphas_cumprod[t] UpperCAmelCase__ = self.alphas_cumprod[prev_timestep] if prev_timestep >= 0 else self.one UpperCAmelCase__ = 1 - alpha_prod_t UpperCAmelCase__ = 1 - alpha_prod_t_prev if prev_timestep == t - 1: UpperCAmelCase__ = self.betas[t] UpperCAmelCase__ = self.alphas[t] else: UpperCAmelCase__ = 1 - alpha_prod_t / alpha_prod_t_prev UpperCAmelCase__ = 1 - beta # 2. compute predicted original sample from predicted noise also called # "predicted x_0" of formula (15) from https://arxiv.org/pdf/2006.11239.pdf if self.config.prediction_type == "epsilon": UpperCAmelCase__ = (sample - beta_prod_t ** 0.5 * model_output) / alpha_prod_t ** 0.5 elif self.config.prediction_type == "sample": UpperCAmelCase__ = model_output else: raise ValueError( f"""prediction_type given as {self.config.prediction_type} must be one of `epsilon` or `sample`""" " for the UnCLIPScheduler." ) # 3. Clip "predicted x_0" if self.config.clip_sample: UpperCAmelCase__ = torch.clamp( __UpperCAmelCase , -self.config.clip_sample_range , self.config.clip_sample_range ) # 4. Compute coefficients for pred_original_sample x_0 and current sample x_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = (alpha_prod_t_prev ** 0.5 * beta) / beta_prod_t UpperCAmelCase__ = alpha ** 0.5 * beta_prod_t_prev / beta_prod_t # 5. Compute predicted previous sample µ_t # See formula (7) from https://arxiv.org/pdf/2006.11239.pdf UpperCAmelCase__ = pred_original_sample_coeff * pred_original_sample + current_sample_coeff * sample # 6. Add noise UpperCAmelCase__ = 0 if t > 0: UpperCAmelCase__ = randn_tensor( model_output.shape , dtype=model_output.dtype , generator=__UpperCAmelCase , device=model_output.device ) UpperCAmelCase__ = self._get_variance( __UpperCAmelCase , predicted_variance=__UpperCAmelCase , prev_timestep=__UpperCAmelCase , ) if self.variance_type == "fixed_small_log": UpperCAmelCase__ = variance elif self.variance_type == "learned_range": UpperCAmelCase__ = (0.5 * variance).exp() else: raise ValueError( f"""variance_type given as {self.variance_type} must be one of `fixed_small_log` or `learned_range`""" " for the UnCLIPScheduler." ) UpperCAmelCase__ = variance * variance_noise UpperCAmelCase__ = pred_prev_sample + variance if not return_dict: return (pred_prev_sample,) return UnCLIPSchedulerOutput(prev_sample=__UpperCAmelCase , pred_original_sample=__UpperCAmelCase ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.FloatTensor , __UpperCAmelCase : torch.IntTensor , ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = self.alphas_cumprod.to(device=original_samples.device , dtype=original_samples.dtype ) UpperCAmelCase__ = timesteps.to(original_samples.device ) UpperCAmelCase__ = alphas_cumprod[timesteps] ** 0.5 UpperCAmelCase__ = sqrt_alpha_prod.flatten() while len(sqrt_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = (1 - alphas_cumprod[timesteps]) ** 0.5 UpperCAmelCase__ = sqrt_one_minus_alpha_prod.flatten() while len(sqrt_one_minus_alpha_prod.shape ) < len(original_samples.shape ): UpperCAmelCase__ = sqrt_one_minus_alpha_prod.unsqueeze(-1 ) UpperCAmelCase__ = sqrt_alpha_prod * original_samples + sqrt_one_minus_alpha_prod * noise return noisy_samples
65
1
from __future__ import annotations UpperCamelCase__ = 'Muhammad Umer Farooq' UpperCamelCase__ = 'MIT' UpperCamelCase__ = '1.0.0' UpperCamelCase__ = 'Muhammad Umer Farooq' UpperCamelCase__ = 'contact@muhammadumerfarooq.me' UpperCamelCase__ = 'Alpha' import re from html.parser import HTMLParser from urllib import parse import requests class A ( UpperCAmelCase_ ): def __init__(self : Optional[int] , __UpperCAmelCase : str ) -> None: """simple docstring""" super().__init__() UpperCAmelCase__ = [] UpperCAmelCase__ = domain def lowercase_ (self : List[Any] , __UpperCAmelCase : str , __UpperCAmelCase : list[tuple[str, str | None]] ) -> None: """simple docstring""" if tag == "a": # Check the list of defined attributes. for name, value in attrs: # If href is defined, and not empty nor # print it. if name == "href" and value != "#" and value != "": # If not already in urls. if value not in self.urls: UpperCAmelCase__ = parse.urljoin(self.domain , __UpperCAmelCase ) self.urls.append(__UpperCAmelCase ) def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' return ".".join(get_sub_domain_name(__A ).split("." )[-2:] ) def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' return parse.urlparse(__A ).netloc def lowerCAmelCase_ ( __A = "https://github.com" ) -> list[str]: '''simple docstring''' UpperCAmelCase__ = get_domain_name(__A ) # Initialize the parser UpperCAmelCase__ = Parser(__A ) try: # Open URL UpperCAmelCase__ = requests.get(__A ) # pass the raw HTML to the parser to get links parser.feed(r.text ) # Get links and loop through UpperCAmelCase__ = set() for link in parser.urls: # open URL. # read = requests.get(link) try: UpperCAmelCase__ = requests.get(__A ) # Get the valid email. UpperCAmelCase__ = re.findall("[a-zA-Z0-9]+@" + domain, read.text ) # If not in list then append it. for email in emails: valid_emails.add(__A ) except ValueError: pass except ValueError: raise SystemExit(1 ) # Finally return a sorted list of email addresses with no duplicates. return sorted(__A ) if __name__ == "__main__": UpperCamelCase__ = emails_from_url('https://github.com') print(f'''{len(emails)} emails found:''') print('\n'.join(sorted(emails)))
65
import inspect import os import unittest import torch import accelerate from accelerate import Accelerator from accelerate.test_utils import execute_subprocess_async, require_multi_gpu from accelerate.utils import patch_environment class A ( unittest.TestCase ): def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = inspect.getfile(accelerate.test_utils ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_script.py"] ) UpperCAmelCase__ = os.path.sep.join( mod_file.split(os.path.sep )[:-1] + ["scripts", "test_distributed_data_loop.py"] ) UpperCAmelCase__ = os.path.sep.join(mod_file.split(os.path.sep )[:-1] + ["scripts", "test_ops.py"] ) @require_multi_gpu def lowercase_ (self : List[str] ) -> Any: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.test_file_path] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : str ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices.""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.operation_file_path] print(f"""Command: {cmd}""" ) with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Tuple ) -> int: """simple docstring""" UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", inspect.getfile(self.__class__ )] with patch_environment(omp_num_threads=1 ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) @require_multi_gpu def lowercase_ (self : Dict ) -> str: """simple docstring""" print(f"""Found {torch.cuda.device_count()} devices, using 2 devices only""" ) UpperCAmelCase__ = ["torchrun", f"""--nproc_per_node={torch.cuda.device_count()}""", self.data_loop_file_path] with patch_environment(omp_num_threads=1 , cuda_visible_devices="0,1" ): execute_subprocess_async(__UpperCAmelCase , env=os.environ.copy() ) if __name__ == "__main__": UpperCamelCase__ = Accelerator() UpperCamelCase__ = (accelerator.state.process_index + 2, 1_0) UpperCamelCase__ = torch.randint(0, 1_0, shape).to(accelerator.device) UpperCamelCase__ = '' UpperCamelCase__ = accelerator.pad_across_processes(tensor) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." if not torch.equal(tensora[: accelerator.state.process_index + 2], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[accelerator.state.process_index + 2 :] == 0): error_msg += "Padding was not done with the right value (0)." UpperCamelCase__ = accelerator.pad_across_processes(tensor, pad_first=True) if tensora.shape[0] != accelerator.state.num_processes + 1: error_msg += f"Found shape {tensora.shape} but should have {accelerator.state.num_processes + 1} at dim 0." UpperCamelCase__ = accelerator.state.num_processes - accelerator.state.process_index - 1 if not torch.equal(tensora[index:], tensor): error_msg += "Tensors have different values." if not torch.all(tensora[:index] == 0): error_msg += "Padding was not done with the right value (0)." # Raise error at the end to make sure we don't stop at the first failure. if len(error_msg) > 0: raise ValueError(error_msg)
65
1
from math import ceil def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = list(range(0, __A ) ) UpperCAmelCase__ = [item for sublist in list(device_map.values() ) for item in sublist] # Duplicate check UpperCAmelCase__ = [] for i in device_map_blocks: if device_map_blocks.count(__A ) > 1 and i not in duplicate_blocks: duplicate_blocks.append(__A ) # Missing blocks UpperCAmelCase__ = [i for i in blocks if i not in device_map_blocks] UpperCAmelCase__ = [i for i in device_map_blocks if i not in blocks] if len(__A ) != 0: raise ValueError( "Duplicate attention blocks specified in device_map. Attention blocks must be specified to one device." " These attention blocks were specified more than once: " + str(__A ) ) if len(__A ) != 0: raise ValueError( "There are attention blocks for this model that are not specified in the device_map. Add these attention " "blocks to a device on the device_map: " + str(__A ) ) if len(__A ) != 0: raise ValueError( "The device_map contains more attention blocks than this model has. Remove these from the device_map:" + str(__A ) ) def lowerCAmelCase_ ( __A, __A ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ = list(range(__A ) ) UpperCAmelCase__ = int(ceil(n_layers / len(__A ) ) ) UpperCAmelCase__ = [layers[i : i + n_blocks] for i in range(0, __A, __A )] return dict(zip(__A, __A ) )
65
import argparse import torch from torch import nn from transformers import MBartConfig, MBartForConditionalGeneration def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = [ "encoder.version", "decoder.version", "model.encoder.version", "model.decoder.version", "_float_tensor", "decoder.output_projection.weight", ] for k in ignore_keys: state_dict.pop(__A, __A ) def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = emb.weight.shape UpperCAmelCase__ = nn.Linear(__A, __A, bias=__A ) UpperCAmelCase__ = emb.weight.data return lin_layer def lowerCAmelCase_ ( __A, __A="facebook/mbart-large-en-ro", __A=False, __A=False ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = torch.load(__A, map_location="cpu" )["model"] remove_ignore_keys_(__A ) UpperCAmelCase__ = state_dict["encoder.embed_tokens.weight"].shape[0] UpperCAmelCase__ = MBartConfig.from_pretrained(__A, vocab_size=__A ) if mbart_aa and finetuned: UpperCAmelCase__ = "relu" UpperCAmelCase__ = state_dict["decoder.embed_tokens.weight"] UpperCAmelCase__ = MBartForConditionalGeneration(__A ) model.model.load_state_dict(__A ) if finetuned: UpperCAmelCase__ = make_linear_from_emb(model.model.shared ) return model if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() # Required parameters parser.add_argument( 'fairseq_path', type=str, help='bart.large, bart.large.cnn or a path to a model.pt on local filesystem.' ) parser.add_argument('pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model.') parser.add_argument( '--hf_config', default='facebook/mbart-large-cc25', type=str, help='Which huggingface architecture to use: mbart-large', ) parser.add_argument('--mbart_50', action='store_true', help='whether the model is mMART-50 checkpoint') parser.add_argument('--finetuned', action='store_true', help='whether the model is a fine-tuned checkpoint') UpperCamelCase__ = parser.parse_args() UpperCamelCase__ = convert_fairseq_mbart_checkpoint_from_disk( args.fairseq_path, hf_config_path=args.hf_config, finetuned=args.finetuned, mbart_aa=args.mbart_aa ) model.save_pretrained(args.pytorch_dump_folder_path)
65
1
def lowerCAmelCase_ ( __A ) -> list[list[int]]: '''simple docstring''' UpperCAmelCase__ = [] if len(__A ) == 1: return [nums.copy()] for _ in range(len(__A ) ): UpperCAmelCase__ = nums.pop(0 ) UpperCAmelCase__ = permute(__A ) for perm in permutations: perm.append(__A ) result.extend(__A ) nums.append(__A ) return result def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' def backtrack(__A ): if start == len(__A ) - 1: output.append(nums[:] ) else: for i in range(__A, len(__A ) ): UpperCAmelCase__ , UpperCAmelCase__ = nums[i], nums[start] backtrack(start + 1 ) UpperCAmelCase__ , UpperCAmelCase__ = nums[i], nums[start] # backtrack UpperCAmelCase__ = [] backtrack(0 ) return output if __name__ == "__main__": import doctest # use res to print the data in permute2 function UpperCamelCase__ = permutea([1, 2, 3]) print(res) doctest.testmod()
65
from .dependency_versions_table import deps from .utils.versions import require_version, require_version_core # define which module versions we always want to check at run time # (usually the ones defined in `install_requires` in setup.py) # # order specific notes: # - tqdm must be checked before tokenizers UpperCamelCase__ = [ 'python', 'tqdm', 'regex', 'requests', 'packaging', 'filelock', 'numpy', 'tokenizers', 'huggingface-hub', 'safetensors', 'accelerate', 'pyyaml', ] for pkg in pkgs_to_check_at_runtime: if pkg in deps: if pkg == "tokenizers": # must be loaded here, or else tqdm check may fail from .utils import is_tokenizers_available if not is_tokenizers_available(): continue # not required, check version only if installed elif pkg == "accelerate": # must be loaded here, or else tqdm check may fail from .utils import is_accelerate_available # Maybe switch to is_torch_available in the future here so that Accelerate is hard dep of # Transformers with PyTorch if not is_accelerate_available(): continue # not required, check version only if installed require_version_core(deps[pkg]) else: raise ValueError(f'''can\'t find {pkg} in {deps.keys()}, check dependency_versions_table.py''') def lowerCAmelCase_ ( __A, __A=None ) -> Dict: '''simple docstring''' require_version(deps[pkg], __A )
65
1
from typing import TYPE_CHECKING from ...utils import OptionalDependencyNotAvailable, _LazyModule, is_tokenizers_available, is_torch_available UpperCamelCase__ = { 'configuration_biogpt': ['BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP', 'BioGptConfig'], 'tokenization_biogpt': ['BioGptTokenizer'], } try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: UpperCamelCase__ = [ 'BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST', 'BioGptForCausalLM', 'BioGptForTokenClassification', 'BioGptForSequenceClassification', 'BioGptModel', 'BioGptPreTrainedModel', ] if TYPE_CHECKING: from .configuration_biogpt import BIOGPT_PRETRAINED_CONFIG_ARCHIVE_MAP, BioGptConfig from .tokenization_biogpt import BioGptTokenizer try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: pass else: from .modeling_biogpt import ( BIOGPT_PRETRAINED_MODEL_ARCHIVE_LIST, BioGptForCausalLM, BioGptForSequenceClassification, BioGptForTokenClassification, BioGptModel, BioGptPreTrainedModel, ) else: import sys UpperCamelCase__ = _LazyModule(__name__, globals()['__file__'], _import_structure, module_spec=__spec__)
65
import argparse import logging import pickle import random import time import numpy as np from transformers import BertTokenizer, GPTaTokenizer, RobertaTokenizer logging.basicConfig( format='%(asctime)s - %(levelname)s - %(name)s - %(message)s', datefmt='%m/%d/%Y %H:%M:%S', level=logging.INFO ) UpperCamelCase__ = logging.getLogger(__name__) def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = argparse.ArgumentParser( description="Preprocess the data to avoid re-doing it several times by (tokenization + token_to_ids)." ) parser.add_argument("--file_path", type=__A, default="data/dump.txt", help="The path to the data." ) parser.add_argument("--tokenizer_type", type=__A, default="bert", choices=["bert", "roberta", "gpt2"] ) parser.add_argument("--tokenizer_name", type=__A, default="bert-base-uncased", help="The tokenizer to use." ) parser.add_argument("--dump_file", type=__A, default="data/dump", help="The dump file prefix." ) UpperCAmelCase__ = parser.parse_args() logger.info(f"""Loading Tokenizer ({args.tokenizer_name})""" ) if args.tokenizer_type == "bert": UpperCAmelCase__ = BertTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `[CLS]` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `[SEP]` elif args.tokenizer_type == "roberta": UpperCAmelCase__ = RobertaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["cls_token"] # `<s>` UpperCAmelCase__ = tokenizer.special_tokens_map["sep_token"] # `</s>` elif args.tokenizer_type == "gpt2": UpperCAmelCase__ = GPTaTokenizer.from_pretrained(args.tokenizer_name ) UpperCAmelCase__ = tokenizer.special_tokens_map["bos_token"] # `<|endoftext|>` UpperCAmelCase__ = tokenizer.special_tokens_map["eos_token"] # `<|endoftext|>` logger.info(f"""Loading text from {args.file_path}""" ) with open(args.file_path, "r", encoding="utf8" ) as fp: UpperCAmelCase__ = fp.readlines() logger.info("Start encoding" ) logger.info(f"""{len(__A )} examples to process.""" ) UpperCAmelCase__ = [] UpperCAmelCase__ = 0 UpperCAmelCase__ = 10_000 UpperCAmelCase__ = time.time() for text in data: UpperCAmelCase__ = f"""{bos} {text.strip()} {sep}""" UpperCAmelCase__ = tokenizer.encode(__A, add_special_tokens=__A ) rslt.append(__A ) iter += 1 if iter % interval == 0: UpperCAmelCase__ = time.time() logger.info(f"""{iter} examples processed. - {(end-start):.2f}s/{interval}expl""" ) UpperCAmelCase__ = time.time() logger.info("Finished binarization" ) logger.info(f"""{len(__A )} examples processed.""" ) UpperCAmelCase__ = f"""{args.dump_file}.{args.tokenizer_name}.pickle""" UpperCAmelCase__ = tokenizer.vocab_size if vocab_size < (1 << 16): UpperCAmelCase__ = [np.uintaa(__A ) for d in rslt] else: UpperCAmelCase__ = [np.intaa(__A ) for d in rslt] random.shuffle(rslt_ ) logger.info(f"""Dump to {dp_file}""" ) with open(__A, "wb" ) as handle: pickle.dump(rslt_, __A, protocol=pickle.HIGHEST_PROTOCOL ) if __name__ == "__main__": main()
65
1
import gc import random import unittest import numpy as np import torch from transformers import CLIPTextConfig, CLIPTextModel, CLIPTokenizer from diffusers import AutoencoderKL, CycleDiffusionPipeline, DDIMScheduler, UNetaDConditionModel from diffusers.utils import floats_tensor, load_image, load_numpy, slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import ( IMAGE_TO_IMAGE_IMAGE_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS, TEXT_GUIDED_IMAGE_VARIATION_PARAMS, ) from ..test_pipelines_common import PipelineLatentTesterMixin, PipelineTesterMixin enable_full_determinism() class A ( UpperCAmelCase_ , UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Optional[int] = CycleDiffusionPipeline __UpperCAmelCase : Union[str, Any] = TEXT_GUIDED_IMAGE_VARIATION_PARAMS - { 'negative_prompt', 'height', 'width', 'negative_prompt_embeds', } __UpperCAmelCase : Union[str, Any] = PipelineTesterMixin.required_optional_params - {'latents'} __UpperCAmelCase : List[Any] = TEXT_GUIDED_IMAGE_VARIATION_BATCH_PARAMS.union({'source_prompt'} ) __UpperCAmelCase : Optional[int] = IMAGE_TO_IMAGE_IMAGE_PARAMS __UpperCAmelCase : str = IMAGE_TO_IMAGE_IMAGE_PARAMS def lowercase_ (self : str ) -> List[Any]: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDConditionModel( block_out_channels=(3_2, 6_4) , layers_per_block=2 , sample_size=3_2 , in_channels=4 , out_channels=4 , down_block_types=("DownBlock2D", "CrossAttnDownBlock2D") , up_block_types=("CrossAttnUpBlock2D", "UpBlock2D") , cross_attention_dim=3_2 , ) UpperCAmelCase__ = DDIMScheduler( beta_start=0.00085 , beta_end=0.012 , beta_schedule="scaled_linear" , num_train_timesteps=1_0_0_0 , clip_sample=__UpperCAmelCase , set_alpha_to_one=__UpperCAmelCase , ) torch.manual_seed(0 ) UpperCAmelCase__ = AutoencoderKL( block_out_channels=[3_2, 6_4] , in_channels=3 , out_channels=3 , down_block_types=["DownEncoderBlock2D", "DownEncoderBlock2D"] , up_block_types=["UpDecoderBlock2D", "UpDecoderBlock2D"] , latent_channels=4 , ) torch.manual_seed(0 ) UpperCAmelCase__ = CLIPTextConfig( bos_token_id=0 , eos_token_id=2 , hidden_size=3_2 , intermediate_size=3_7 , layer_norm_eps=1E-05 , num_attention_heads=4 , num_hidden_layers=5 , pad_token_id=1 , vocab_size=1_0_0_0 , ) UpperCAmelCase__ = CLIPTextModel(__UpperCAmelCase ) UpperCAmelCase__ = CLIPTokenizer.from_pretrained("hf-internal-testing/tiny-random-clip" ) UpperCAmelCase__ = { "unet": unet, "scheduler": scheduler, "vae": vae, "text_encoder": text_encoder, "tokenizer": tokenizer, "safety_checker": None, "feature_extractor": None, } return components def lowercase_ (self : str , __UpperCAmelCase : Union[str, Any] , __UpperCAmelCase : List[Any]=0 ) -> int: """simple docstring""" UpperCAmelCase__ = floats_tensor((1, 3, 3_2, 3_2) , rng=random.Random(__UpperCAmelCase ) ).to(__UpperCAmelCase ) UpperCAmelCase__ = image / 2 + 0.5 if str(__UpperCAmelCase ).startswith("mps" ): UpperCAmelCase__ = torch.manual_seed(__UpperCAmelCase ) else: UpperCAmelCase__ = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) UpperCAmelCase__ = { "prompt": "An astronaut riding an elephant", "source_prompt": "An astronaut riding a horse", "image": image, "generator": generator, "num_inference_steps": 2, "eta": 0.1, "strength": 0.8, "guidance_scale": 3, "source_guidance_scale": 1, "output_type": "numpy", } return inputs def lowercase_ (self : List[Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = CycleDiffusionPipeline(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase ) UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.images UpperCAmelCase__ = images[0, -3:, -3:, -1] assert images.shape == (1, 3_2, 3_2, 3) UpperCAmelCase__ = np.array([0.4459, 0.4943, 0.4544, 0.6643, 0.5474, 0.4327, 0.5701, 0.5959, 0.5179] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @unittest.skipIf(torch_device != "cuda" , "This test requires a GPU" ) def lowercase_ (self : int ) -> int: """simple docstring""" UpperCAmelCase__ = self.get_dummy_components() for name, module in components.items(): if hasattr(__UpperCAmelCase , "half" ): UpperCAmelCase__ = module.half() UpperCAmelCase__ = CycleDiffusionPipeline(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase ) UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.images UpperCAmelCase__ = images[0, -3:, -3:, -1] assert images.shape == (1, 3_2, 3_2, 3) UpperCAmelCase__ = np.array([0.3506, 0.4543, 0.446, 0.4575, 0.5195, 0.4155, 0.5273, 0.518, 0.4116] ) assert np.abs(image_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def lowercase_ (self : Optional[int] ) -> Tuple: """simple docstring""" return super().test_save_load_local() @unittest.skip("non-deterministic pipeline" ) def lowercase_ (self : Any ) -> str: """simple docstring""" return super().test_inference_batch_single_identical() @skip_mps def lowercase_ (self : List[str] ) -> Any: """simple docstring""" return super().test_dict_tuple_outputs_equivalent() @skip_mps def lowercase_ (self : Optional[int] ) -> Any: """simple docstring""" return super().test_save_load_optional_components() @skip_mps def lowercase_ (self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" return super().test_attention_slicing_forward_pass() @slow @require_torch_gpu class A ( unittest.TestCase ): def lowercase_ (self : Dict ) -> Optional[Any]: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ (self : Union[str, Any] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/cycle-diffusion/black_colored_car.png" ) UpperCAmelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car_fp16.npy" ) UpperCAmelCase__ = init_image.resize((5_1_2, 5_1_2) ) UpperCAmelCase__ = "CompVis/stable-diffusion-v1-4" UpperCAmelCase__ = DDIMScheduler.from_pretrained(__UpperCAmelCase , subfolder="scheduler" ) UpperCAmelCase__ = CycleDiffusionPipeline.from_pretrained( __UpperCAmelCase , scheduler=__UpperCAmelCase , safety_checker=__UpperCAmelCase , torch_dtype=torch.floataa , revision="fp16" ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) pipe.enable_attention_slicing() UpperCAmelCase__ = "A black colored car" UpperCAmelCase__ = "A blue colored car" UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( prompt=__UpperCAmelCase , source_prompt=__UpperCAmelCase , image=__UpperCAmelCase , num_inference_steps=1_0_0 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=__UpperCAmelCase , output_type="np" , ) UpperCAmelCase__ = output.images # the values aren't exactly equal, but the images look the same visually assert np.abs(image - expected_image ).max() < 5E-1 def lowercase_ (self : Any ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = load_image( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main" "/cycle-diffusion/black_colored_car.png" ) UpperCAmelCase__ = load_numpy( "https://huggingface.co/datasets/hf-internal-testing/diffusers-images/resolve/main/cycle-diffusion/blue_colored_car.npy" ) UpperCAmelCase__ = init_image.resize((5_1_2, 5_1_2) ) UpperCAmelCase__ = "CompVis/stable-diffusion-v1-4" UpperCAmelCase__ = DDIMScheduler.from_pretrained(__UpperCAmelCase , subfolder="scheduler" ) UpperCAmelCase__ = CycleDiffusionPipeline.from_pretrained(__UpperCAmelCase , scheduler=__UpperCAmelCase , safety_checker=__UpperCAmelCase ) pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) pipe.enable_attention_slicing() UpperCAmelCase__ = "A black colored car" UpperCAmelCase__ = "A blue colored car" UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe( prompt=__UpperCAmelCase , source_prompt=__UpperCAmelCase , image=__UpperCAmelCase , num_inference_steps=1_0_0 , eta=0.1 , strength=0.85 , guidance_scale=3 , source_guidance_scale=1 , generator=__UpperCAmelCase , output_type="np" , ) UpperCAmelCase__ = output.images assert np.abs(image - expected_image ).max() < 2E-2
65
from manim import * class A ( UpperCAmelCase_ ): def lowercase_ (self : Union[str, Any] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = Rectangle(height=0.5 , width=0.5 ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0 ) UpperCAmelCase__ = Rectangle(height=0.25 , width=0.25 ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("CPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) cpu.move_to([-2.5, -0.5, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(4 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("GPU" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) gpu.move_to([-1, -1, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Model" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) model.move_to([3, -1.0, 0] ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = [] UpperCAmelCase__ = [] for i, rect in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = fill.copy().set_fill(__UpperCAmelCase , opacity=0.8 ) target.move_to(__UpperCAmelCase ) model_arr.append(__UpperCAmelCase ) UpperCAmelCase__ = Rectangle(height=0.46 , width=0.46 ).set_stroke(width=0.0 ).set_fill(__UpperCAmelCase , opacity=0.8 ) cpu_target.move_to(cpu_left_col_base[i] ) model_cpu_arr.append(__UpperCAmelCase ) self.add(*__UpperCAmelCase , *__UpperCAmelCase ) UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = [meta_mem.copy() for i in range(6 )] UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(*__UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = VGroup(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0 ) UpperCAmelCase__ = Text("Disk" , font_size=2_4 ) UpperCAmelCase__ = Group(__UpperCAmelCase , __UpperCAmelCase ).arrange(__UpperCAmelCase , buff=0.5 , aligned_edge=__UpperCAmelCase ) disk.move_to([-4, -1.25, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = Square(side_length=2.2 ) key.move_to([-5, 2, 0] ) UpperCAmelCase__ = MarkupText( f"""<b>Key:</b>\n\n<span fgcolor='{YELLOW}'>●</span> Empty Model""" , font_size=1_8 , ) key_text.move_to([-5, 2.4, 0] ) self.add(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""<span fgcolor='{BLUE}'>●</span> Checkpoint""" , font_size=1_8 , ) blue_text.next_to(__UpperCAmelCase , DOWN * 2.4 , aligned_edge=key_text.get_left() ) self.add(__UpperCAmelCase ) UpperCAmelCase__ = MarkupText( f"""Now watch as an input is passed through the model\nand how the memory is utilized and handled.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase ) ) UpperCAmelCase__ = Square(0.3 ) input.set_fill(__UpperCAmelCase , opacity=1.0 ) input.set_stroke(width=0.0 ) input.next_to(model_base[0] , __UpperCAmelCase , buff=0.5 ) self.play(Write(__UpperCAmelCase ) ) input.generate_target() input.target.next_to(model_arr[0] , direction=__UpperCAmelCase , buff=0.02 ) self.play(MoveToTarget(__UpperCAmelCase ) ) self.play(FadeOut(__UpperCAmelCase ) ) UpperCAmelCase__ = Arrow(start=__UpperCAmelCase , end=__UpperCAmelCase , color=__UpperCAmelCase , buff=0.5 ) a.next_to(model_arr[0].get_left() , __UpperCAmelCase , buff=0.2 ) model_cpu_arr[0].generate_target() model_cpu_arr[0].target.move_to(gpu_rect[0] ) UpperCAmelCase__ = MarkupText( f"""As the input reaches a layer, the hook triggers\nand weights are moved from the CPU\nto the GPU and back.""" , font_size=2_4 , ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) ) UpperCAmelCase__ = {"run_time": 1, "fade_in": True, "fade_out": True, "buff": 0.02} self.play( Write(__UpperCAmelCase ) , Circumscribe(model_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_cpu_arr[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[0] ) ) UpperCAmelCase__ = a.copy() for i in range(6 ): a_c.next_to(model_arr[i].get_right() + 0.02 , __UpperCAmelCase , buff=0.2 ) input.generate_target() input.target.move_to(model_arr[i].get_right() + 0.02 ) UpperCAmelCase__ = AnimationGroup( FadeOut(__UpperCAmelCase , run_time=0.5 ) , MoveToTarget(__UpperCAmelCase , run_time=0.5 ) , FadeIn(__UpperCAmelCase , run_time=0.5 ) , lag_ratio=0.2 ) self.play(__UpperCAmelCase ) model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[i] ) if i < 5: model_cpu_arr[i + 1].generate_target() model_cpu_arr[i + 1].target.move_to(gpu_rect[0] ) if i >= 1: UpperCAmelCase__ = 0.7 self.play( Circumscribe(model_arr[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i] , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(model_arr[i + 1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) if i < 1: self.play( MoveToTarget(model_cpu_arr[i] ) , MoveToTarget(model_cpu_arr[i + 1] ) , ) else: self.play( MoveToTarget(model_cpu_arr[i] , run_time=0.7 ) , MoveToTarget(model_cpu_arr[i + 1] , run_time=0.7 ) , ) else: model_cpu_arr[i].generate_target() model_cpu_arr[i].target.move_to(cpu_left_col_base[-1] ) input.generate_target() input.target.next_to(model_arr[-1].get_right() , RIGHT + 0.02 , buff=0.2 ) self.play( Circumscribe(model_arr[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(cpu_left_col_base[-1] , color=__UpperCAmelCase , **__UpperCAmelCase ) , Circumscribe(gpu_rect[0] , color=__UpperCAmelCase , **__UpperCAmelCase ) , ) self.play(MoveToTarget(model_cpu_arr[i] ) ) UpperCAmelCase__ = a_c UpperCAmelCase__ = a_c.copy() input.generate_target() input.target.next_to(model_base[-1] , RIGHT + 0.02 , buff=0.5 ) self.play( FadeOut(__UpperCAmelCase ) , FadeOut(__UpperCAmelCase , run_time=0.5 ) , ) UpperCAmelCase__ = MarkupText(f"""Inference on a model too large for GPU memory\nis successfully completed.""" , font_size=2_4 ) step_a.move_to([2, 2, 0] ) self.play(Write(__UpperCAmelCase , run_time=3 ) , MoveToTarget(__UpperCAmelCase ) ) self.wait()
65
1
from collections import defaultdict from graphs.minimum_spanning_tree_prims import prisms_algorithm as mst def lowerCAmelCase_ ( ) -> str: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = 9, 14 # noqa: F841 UpperCAmelCase__ = [ [0, 1, 4], [0, 7, 8], [1, 2, 8], [7, 8, 7], [7, 6, 1], [2, 8, 2], [8, 6, 6], [2, 3, 7], [2, 5, 4], [6, 5, 2], [3, 5, 14], [3, 4, 9], [5, 4, 10], [1, 7, 11], ] UpperCAmelCase__ = defaultdict(__A ) for nodea, nodea, cost in edges: adjancency[nodea].append([nodea, cost] ) adjancency[nodea].append([nodea, cost] ) UpperCAmelCase__ = mst(__A ) UpperCAmelCase__ = [ [7, 6, 1], [2, 8, 2], [6, 5, 2], [0, 1, 4], [2, 5, 4], [2, 3, 7], [0, 7, 8], [3, 4, 9], ] for answer in expected: UpperCAmelCase__ = tuple(answer[:2] ) UpperCAmelCase__ = tuple(edge[::-1] ) assert edge in result or reverse in result
65
from __future__ import annotations from scipy.special import comb # type: ignore class A : def __init__(self : List[Any] , __UpperCAmelCase : list[tuple[float, float]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = list_of_points # Degree determines the flexibility of the curve. # Degree = 1 will produce a straight line. UpperCAmelCase__ = len(__UpperCAmelCase ) - 1 def lowercase_ (self : int , __UpperCAmelCase : float ) -> list[float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = [] for i in range(len(self.list_of_points ) ): # basis function for each i output_values.append( comb(self.degree , __UpperCAmelCase ) * ((1 - t) ** (self.degree - i)) * (t**i) ) # the basis must sum up to 1 for it to produce a valid Bezier curve. assert round(sum(__UpperCAmelCase ) , 5 ) == 1 return output_values def lowercase_ (self : Dict , __UpperCAmelCase : float ) -> tuple[float, float]: """simple docstring""" assert 0 <= t <= 1, "Time t must be between 0 and 1." UpperCAmelCase__ = self.basis_function(__UpperCAmelCase ) UpperCAmelCase__ = 0.0 UpperCAmelCase__ = 0.0 for i in range(len(self.list_of_points ) ): # For all points, sum up the product of i-th basis function and i-th point. x += basis_function[i] * self.list_of_points[i][0] y += basis_function[i] * self.list_of_points[i][1] return (x, y) def lowercase_ (self : Optional[int] , __UpperCAmelCase : float = 0.01 ) -> Optional[int]: """simple docstring""" from matplotlib import pyplot as plt # type: ignore UpperCAmelCase__ = [] # x coordinates of points to plot UpperCAmelCase__ = [] # y coordinates of points to plot UpperCAmelCase__ = 0.0 while t <= 1: UpperCAmelCase__ = self.bezier_curve_function(__UpperCAmelCase ) to_plot_x.append(value[0] ) to_plot_y.append(value[1] ) t += step_size UpperCAmelCase__ = [i[0] for i in self.list_of_points] UpperCAmelCase__ = [i[1] for i in self.list_of_points] plt.plot( __UpperCAmelCase , __UpperCAmelCase , color="blue" , label="Curve of Degree " + str(self.degree ) , ) plt.scatter(__UpperCAmelCase , __UpperCAmelCase , color="red" , label="Control Points" ) plt.legend() plt.show() if __name__ == "__main__": import doctest doctest.testmod() BezierCurve([(1, 2), (3, 5)]).plot_curve() # degree 1 BezierCurve([(0, 0), (5, 5), (5, 0)]).plot_curve() # degree 2 BezierCurve([(0, 0), (5, 5), (5, 0), (2.5, -2.5)]).plot_curve() # degree 3
65
1
from ..utils import ( OptionalDependencyNotAvailable, is_flax_available, is_scipy_available, is_torch_available, is_torchsde_available, ) try: if not is_torch_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_pt_objects import * # noqa F403 else: from .scheduling_consistency_models import CMStochasticIterativeScheduler from .scheduling_ddim import DDIMScheduler from .scheduling_ddim_inverse import DDIMInverseScheduler from .scheduling_ddim_parallel import DDIMParallelScheduler from .scheduling_ddpm import DDPMScheduler from .scheduling_ddpm_parallel import DDPMParallelScheduler from .scheduling_deis_multistep import DEISMultistepScheduler from .scheduling_dpmsolver_multistep import DPMSolverMultistepScheduler from .scheduling_dpmsolver_multistep_inverse import DPMSolverMultistepInverseScheduler from .scheduling_dpmsolver_singlestep import DPMSolverSinglestepScheduler from .scheduling_euler_ancestral_discrete import EulerAncestralDiscreteScheduler from .scheduling_euler_discrete import EulerDiscreteScheduler from .scheduling_heun_discrete import HeunDiscreteScheduler from .scheduling_ipndm import IPNDMScheduler from .scheduling_k_dpm_2_ancestral_discrete import KDPMaAncestralDiscreteScheduler from .scheduling_k_dpm_2_discrete import KDPMaDiscreteScheduler from .scheduling_karras_ve import KarrasVeScheduler from .scheduling_pndm import PNDMScheduler from .scheduling_repaint import RePaintScheduler from .scheduling_sde_ve import ScoreSdeVeScheduler from .scheduling_sde_vp import ScoreSdeVpScheduler from .scheduling_unclip import UnCLIPScheduler from .scheduling_unipc_multistep import UniPCMultistepScheduler from .scheduling_utils import KarrasDiffusionSchedulers, SchedulerMixin from .scheduling_vq_diffusion import VQDiffusionScheduler try: if not is_flax_available(): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_flax_objects import * # noqa F403 else: from .scheduling_ddim_flax import FlaxDDIMScheduler from .scheduling_ddpm_flax import FlaxDDPMScheduler from .scheduling_dpmsolver_multistep_flax import FlaxDPMSolverMultistepScheduler from .scheduling_karras_ve_flax import FlaxKarrasVeScheduler from .scheduling_lms_discrete_flax import FlaxLMSDiscreteScheduler from .scheduling_pndm_flax import FlaxPNDMScheduler from .scheduling_sde_ve_flax import FlaxScoreSdeVeScheduler from .scheduling_utils_flax import ( FlaxKarrasDiffusionSchedulers, FlaxSchedulerMixin, FlaxSchedulerOutput, broadcast_to_shape_from_left, ) try: if not (is_torch_available() and is_scipy_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_scipy_objects import * # noqa F403 else: from .scheduling_lms_discrete import LMSDiscreteScheduler try: if not (is_torch_available() and is_torchsde_available()): raise OptionalDependencyNotAvailable() except OptionalDependencyNotAvailable: from ..utils.dummy_torch_and_torchsde_objects import * # noqa F403 else: from .scheduling_dpmsolver_sde import DPMSolverSDEScheduler
65
import os import tempfile import unittest from transformers.models.marian.convert_marian_tatoeba_to_pytorch import DEFAULT_REPO, TatoebaConverter from transformers.testing_utils import slow from transformers.utils import cached_property @unittest.skipUnless(os.path.exists(UpperCAmelCase_ ) , 'Tatoeba directory does not exist.' ) class A ( unittest.TestCase ): @cached_property def lowercase_ (self : Optional[int] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = tempfile.mkdtemp() return TatoebaConverter(save_dir=__UpperCAmelCase ) @slow def lowercase_ (self : List[Any] ) -> Optional[int]: """simple docstring""" self.resolver.convert_models(["heb-eng"] ) @slow def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ , UpperCAmelCase__ = self.resolver.write_model_card("opus-mt-he-en" , dry_run=__UpperCAmelCase ) assert mmeta["long_pair"] == "heb-eng"
65
1
import pprint import requests UpperCamelCase__ = 'https://zenquotes.io/api' def lowerCAmelCase_ ( ) -> list: '''simple docstring''' return requests.get(API_ENDPOINT_URL + "/today" ).json() def lowerCAmelCase_ ( ) -> list: '''simple docstring''' return requests.get(API_ENDPOINT_URL + "/random" ).json() if __name__ == "__main__": UpperCamelCase__ = random_quotes() pprint.pprint(response)
65
import numpy as np import skfuzzy as fuzz if __name__ == "__main__": # Create universe of discourse in Python using linspace () UpperCamelCase__ = np.linspace(start=0, stop=7_5, num=7_5, endpoint=True, retstep=False) # Create two fuzzy sets by defining any membership function # (trapmf(), gbellmf(), gaussmf(), etc). UpperCamelCase__ = [0, 2_5, 5_0] UpperCamelCase__ = [2_5, 5_0, 7_5] UpperCamelCase__ = fuzz.membership.trimf(X, abca) UpperCamelCase__ = fuzz.membership.trimf(X, abca) # Compute the different operations using inbuilt functions. UpperCamelCase__ = np.ones(7_5) UpperCamelCase__ = np.zeros((7_5,)) # 1. Union = max(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_or(X, young, X, middle_aged)[1] # 2. Intersection = min(µA(x), µB(x)) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, middle_aged)[1] # 3. Complement (A) = (1- min(µA(x)) UpperCamelCase__ = fuzz.fuzzy_not(young) # 4. Difference (A/B) = min(µA(x),(1- µB(x))) UpperCamelCase__ = fuzz.fuzzy_and(X, young, X, fuzz.fuzzy_not(middle_aged)[1])[1] # 5. Algebraic Sum = [µA(x) + µB(x) – (µA(x) * µB(x))] UpperCamelCase__ = young + middle_aged - (young * middle_aged) # 6. Algebraic Product = (µA(x) * µB(x)) UpperCamelCase__ = young * middle_aged # 7. Bounded Sum = min[1,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_and(X, one, X, young + middle_aged)[1] # 8. Bounded difference = min[0,(µA(x), µB(x))] UpperCamelCase__ = fuzz.fuzzy_or(X, zero, X, young - middle_aged)[1] # max-min composition # max-product composition # Plot each set A, set B and each operation result using plot() and subplot(). from matplotlib import pyplot as plt plt.figure() plt.subplot(4, 3, 1) plt.plot(X, young) plt.title('Young') plt.grid(True) plt.subplot(4, 3, 2) plt.plot(X, middle_aged) plt.title('Middle aged') plt.grid(True) plt.subplot(4, 3, 3) plt.plot(X, union) plt.title('union') plt.grid(True) plt.subplot(4, 3, 4) plt.plot(X, intersection) plt.title('intersection') plt.grid(True) plt.subplot(4, 3, 5) plt.plot(X, complement_a) plt.title('complement_a') plt.grid(True) plt.subplot(4, 3, 6) plt.plot(X, difference) plt.title('difference a/b') plt.grid(True) plt.subplot(4, 3, 7) plt.plot(X, alg_sum) plt.title('alg_sum') plt.grid(True) plt.subplot(4, 3, 8) plt.plot(X, alg_product) plt.title('alg_product') plt.grid(True) plt.subplot(4, 3, 9) plt.plot(X, bdd_sum) plt.title('bdd_sum') plt.grid(True) plt.subplot(4, 3, 1_0) plt.plot(X, bdd_difference) plt.title('bdd_difference') plt.grid(True) plt.subplots_adjust(hspace=0.5) plt.show()
65
1
import os import unicodedata from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, PreTrainedTokenizer from ...utils import SPIECE_UNDERLINE, logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = {'vocab_file': 'spiece.model'} UpperCamelCase__ = { 'vocab_file': { 'TsinghuaAI/CPM-Generate': 'https://huggingface.co/TsinghuaAI/CPM-Generate/resolve/main/spiece.model', } } class A ( UpperCAmelCase_ ): def __init__(self : Tuple , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Any=False , __UpperCAmelCase : Dict=True , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : int="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : str="<sep>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : List[str]="<cls>" , __UpperCAmelCase : Dict="<mask>" , __UpperCAmelCase : int=["<eop>", "<eod>"] , __UpperCAmelCase : Optional[Dict[str, Any]] = None , **__UpperCAmelCase : Dict , ) -> None: """simple docstring""" UpperCAmelCase__ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( do_lower_case=__UpperCAmelCase , remove_space=__UpperCAmelCase , keep_accents=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) UpperCAmelCase__ = 3 UpperCAmelCase__ = do_lower_case UpperCAmelCase__ = remove_space UpperCAmelCase__ = keep_accents UpperCAmelCase__ = vocab_file UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(__UpperCAmelCase ) try: import jieba except ModuleNotFoundError as error: raise error.__class__( "You need to install jieba to use CpmTokenizer or CpmTokenizerFast. " "See https://pypi.org/project/jieba/ for installation." ) UpperCAmelCase__ = jieba UpperCAmelCase__ = str.maketrans(" \n" , "\u2582\u2583" ) @property # Copied from transformers.models.xlnet.tokenization_xlnet.XLNetTokenizer.vocab_size def lowercase_ (self : Optional[Any] ) -> List[Any]: """simple docstring""" return len(self.sp_model ) def lowercase_ (self : int ) -> Any: """simple docstring""" UpperCAmelCase__ = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def __getstate__(self : int ) -> Tuple: """simple docstring""" UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None return state def __setstate__(self : List[Any] , __UpperCAmelCase : Dict ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(self.vocab_file ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[Any] ) -> int: """simple docstring""" if self.remove_space: UpperCAmelCase__ = " ".join(inputs.strip().split() ) else: UpperCAmelCase__ = inputs UpperCAmelCase__ = outputs.replace("``" , "\"" ).replace("''" , "\"" ) if not self.keep_accents: UpperCAmelCase__ = unicodedata.normalize("NFKD" , __UpperCAmelCase ) UpperCAmelCase__ = "".join([c for c in outputs if not unicodedata.combining(__UpperCAmelCase )] ) if self.do_lower_case: UpperCAmelCase__ = outputs.lower() return outputs def lowercase_ (self : Any , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = self.preprocess_text(__UpperCAmelCase ) UpperCAmelCase__ = self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) UpperCAmelCase__ = [] for piece in pieces: if len(__UpperCAmelCase ) > 1 and piece[-1] == str("," ) and piece[-2].isdigit(): UpperCAmelCase__ = self.sp_model.EncodeAsPieces(piece[:-1].replace(__UpperCAmelCase , "" ) ) if piece[0] != SPIECE_UNDERLINE and cur_pieces[0][0] == SPIECE_UNDERLINE: if len(cur_pieces[0] ) == 1: UpperCAmelCase__ = cur_pieces[1:] else: UpperCAmelCase__ = cur_pieces[0][1:] cur_pieces.append(piece[-1] ) new_pieces.extend(__UpperCAmelCase ) else: new_pieces.append(__UpperCAmelCase ) return new_pieces def lowercase_ (self : Optional[Any] , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" return self.sp_model.PieceToId(__UpperCAmelCase ) def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> Optional[Any]: """simple docstring""" return self.sp_model.IdToPiece(__UpperCAmelCase ) def lowercase_ (self : str , __UpperCAmelCase : Optional[int] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = "".join(__UpperCAmelCase ).replace(__UpperCAmelCase , " " ).strip() return out_string def lowercase_ (self : Dict , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return token_ids_a + sep + cls return token_ids_a + sep + token_ids_a + sep + cls def lowercase_ (self : Tuple , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) if token_ids_a is not None: return ([0] * len(__UpperCAmelCase )) + [1] + ([0] * len(__UpperCAmelCase )) + [1, 1] return ([0] * len(__UpperCAmelCase )) + [1, 1] def lowercase_ (self : Optional[Any] , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [2] if token_ids_a is None: return len(token_ids_a + sep ) * [0] + cls_segment_id return len(token_ids_a + sep ) * [0] + len(token_ids_a + sep ) * [1] + cls_segment_id def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , "wb" ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,) def lowercase_ (self : List[Any] , *__UpperCAmelCase : Tuple , **__UpperCAmelCase : List[Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = super()._decode(*__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = text.replace(" " , "" ).replace("\u2582" , " " ).replace("\u2583" , "\n" ) return text
65
from __future__ import annotations from collections import deque class A : def __init__(self : Dict , __UpperCAmelCase : list[str] ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = [] self.adlist.append( {"value": "", "next_states": [], "fail_state": 0, "output": []} ) for keyword in keywords: self.add_keyword(__UpperCAmelCase ) self.set_fail_transitions() def lowercase_ (self : Tuple , __UpperCAmelCase : int , __UpperCAmelCase : str ) -> int | None: """simple docstring""" for state in self.adlist[current_state]["next_states"]: if char == self.adlist[state]["value"]: return state return None def lowercase_ (self : Dict , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = 0 for character in keyword: UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , __UpperCAmelCase ) if next_state is None: self.adlist.append( { "value": character, "next_states": [], "fail_state": 0, "output": [], } ) self.adlist[current_state]["next_states"].append(len(self.adlist ) - 1 ) UpperCAmelCase__ = len(self.adlist ) - 1 else: UpperCAmelCase__ = next_state self.adlist[current_state]["output"].append(__UpperCAmelCase ) def lowercase_ (self : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = deque() for node in self.adlist[0]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = 0 while q: UpperCAmelCase__ = q.popleft() for child in self.adlist[r]["next_states"]: q.append(__UpperCAmelCase ) UpperCAmelCase__ = self.adlist[r]["fail_state"] while ( self.find_next_state(__UpperCAmelCase , self.adlist[child]["value"] ) is None and state != 0 ): UpperCAmelCase__ = self.adlist[state]["fail_state"] UpperCAmelCase__ = self.find_next_state( __UpperCAmelCase , self.adlist[child]["value"] ) if self.adlist[child]["fail_state"] is None: UpperCAmelCase__ = 0 UpperCAmelCase__ = ( self.adlist[child]["output"] + self.adlist[self.adlist[child]["fail_state"]]["output"] ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> dict[str, list[int]]: """simple docstring""" UpperCAmelCase__ = {} # returns a dict with keywords and list of its occurrences UpperCAmelCase__ = 0 for i in range(len(__UpperCAmelCase ) ): while ( self.find_next_state(__UpperCAmelCase , string[i] ) is None and current_state != 0 ): UpperCAmelCase__ = self.adlist[current_state]["fail_state"] UpperCAmelCase__ = self.find_next_state(__UpperCAmelCase , string[i] ) if next_state is None: UpperCAmelCase__ = 0 else: UpperCAmelCase__ = next_state for key in self.adlist[current_state]["output"]: if key not in result: UpperCAmelCase__ = [] result[key].append(i - len(__UpperCAmelCase ) + 1 ) return result if __name__ == "__main__": import doctest doctest.testmod()
65
1
import math import tensorflow as tf from packaging import version def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' UpperCAmelCase__ = tf.convert_to_tensor(__A ) UpperCAmelCase__ = 0.5 * (1.0 + tf.math.erf(x / tf.cast(tf.sqrt(2.0 ), x.dtype ) )) return x * cdf def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = tf.convert_to_tensor(__A ) UpperCAmelCase__ = tf.cast(math.pi, x.dtype ) UpperCAmelCase__ = tf.cast(0.044715, x.dtype ) UpperCAmelCase__ = 0.5 * (1.0 + tf.tanh(tf.sqrt(2.0 / pi ) * (x + coeff * tf.pow(__A, 3 )) )) return x * cdf def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' UpperCAmelCase__ = tf.convert_to_tensor(__A ) return x * tf.tanh(tf.math.softplus(__A ) ) def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = tf.convert_to_tensor(__A ) UpperCAmelCase__ = tf.cast(0.044715, x.dtype ) UpperCAmelCase__ = tf.cast(0.7978845608, x.dtype ) return 0.5 * x * (1.0 + tf.tanh(x * coeffa * (1.0 + coeffa * x * x) )) def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = tf.convert_to_tensor(__A ) UpperCAmelCase__ = tf.cast(1.702, x.dtype ) return x * tf.math.sigmoid(coeff * x ) def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' return tf.clip_by_value(_gelu(__A ), -10, 10 ) def lowerCAmelCase_ ( __A, __A=-1 ) -> Dict: '''simple docstring''' UpperCAmelCase__ , UpperCAmelCase__ = tf.split(__A, 2, axis=__A ) return a * tf.math.sigmoid(__A ) if version.parse(tf.version.VERSION) >= version.parse('2.4'): def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' return tf.keras.activations.gelu(__A, approximate=__A ) UpperCamelCase__ = tf.keras.activations.gelu UpperCamelCase__ = approximate_gelu_wrap else: UpperCamelCase__ = _gelu UpperCamelCase__ = _gelu_new UpperCamelCase__ = { 'gelu': gelu, 'gelu_10': gelu_aa, 'gelu_fast': gelu_fast, 'gelu_new': gelu_new, 'glu': glu, 'mish': mish, 'quick_gelu': quick_gelu, 'relu': tf.keras.activations.relu, 'sigmoid': tf.keras.activations.sigmoid, 'silu': tf.keras.activations.swish, 'swish': tf.keras.activations.swish, 'tanh': tf.keras.activations.tanh, } def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' if activation_string in ACTaFN: return ACTaFN[activation_string] else: raise KeyError(f"""function {activation_string} not found in ACT2FN mapping {list(ACTaFN.keys() )}""" )
65
import warnings from typing import Any, Dict, List, Optional, Union import numpy as np from ...audio_utils import mel_filter_bank, optimal_fft_length, spectrogram, window_function from ...feature_extraction_sequence_utils import SequenceFeatureExtractor from ...feature_extraction_utils import BatchFeature from ...utils import PaddingStrategy, TensorType, logging UpperCamelCase__ = logging.get_logger(__name__) class A ( UpperCAmelCase_ ): __UpperCAmelCase : int = ['input_values', 'attention_mask'] def __init__(self : Any , __UpperCAmelCase : int = 1 , __UpperCAmelCase : int = 1_6_0_0_0 , __UpperCAmelCase : float = 0.0 , __UpperCAmelCase : bool = False , __UpperCAmelCase : int = 8_0 , __UpperCAmelCase : int = 1_6 , __UpperCAmelCase : int = 6_4 , __UpperCAmelCase : str = "hann_window" , __UpperCAmelCase : float = 1.0 , __UpperCAmelCase : float = 8_0 , __UpperCAmelCase : float = 7_6_0_0 , __UpperCAmelCase : float = 1E-10 , __UpperCAmelCase : int = 2 , __UpperCAmelCase : bool = True , **__UpperCAmelCase : Any , ) -> str: """simple docstring""" super().__init__(feature_size=__UpperCAmelCase , sampling_rate=__UpperCAmelCase , padding_value=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = do_normalize UpperCAmelCase__ = return_attention_mask UpperCAmelCase__ = num_mel_bins UpperCAmelCase__ = hop_length UpperCAmelCase__ = win_length UpperCAmelCase__ = win_function UpperCAmelCase__ = frame_signal_scale UpperCAmelCase__ = fmin UpperCAmelCase__ = fmax UpperCAmelCase__ = mel_floor UpperCAmelCase__ = reduction_factor UpperCAmelCase__ = win_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = hop_length * sampling_rate // 1_0_0_0 UpperCAmelCase__ = optimal_fft_length(self.sample_size ) UpperCAmelCase__ = (self.n_fft // 2) + 1 UpperCAmelCase__ = window_function(window_length=self.sample_size , name=self.win_function , periodic=__UpperCAmelCase ) UpperCAmelCase__ = mel_filter_bank( num_frequency_bins=self.n_freqs , num_mel_filters=self.num_mel_bins , min_frequency=self.fmin , max_frequency=self.fmax , sampling_rate=self.sampling_rate , norm="slaney" , mel_scale="slaney" , ) if frame_signal_scale != 1.0: warnings.warn( "The argument `frame_signal_scale` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) if reduction_factor != 2.0: warnings.warn( "The argument `reduction_factor` is deprecated and will be removed in version 4.30.0 of Transformers" , __UpperCAmelCase , ) @staticmethod # Copied from transformers.models.wav2vec2.feature_extraction_wav2vec2.Wav2Vec2FeatureExtractor.zero_mean_unit_var_norm def lowercase_ (__UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : List[np.ndarray] , __UpperCAmelCase : float = 0.0 ) -> List[np.ndarray]: """simple docstring""" if attention_mask is not None: UpperCAmelCase__ = np.array(__UpperCAmelCase , np.intaa ) UpperCAmelCase__ = [] for vector, length in zip(__UpperCAmelCase , attention_mask.sum(-1 ) ): UpperCAmelCase__ = (vector - vector[:length].mean()) / np.sqrt(vector[:length].var() + 1E-7 ) if length < normed_slice.shape[0]: UpperCAmelCase__ = padding_value normed_input_values.append(__UpperCAmelCase ) else: UpperCAmelCase__ = [(x - x.mean()) / np.sqrt(x.var() + 1E-7 ) for x in input_values] return normed_input_values def lowercase_ (self : Optional[int] , __UpperCAmelCase : np.ndarray , ) -> np.ndarray: """simple docstring""" UpperCAmelCase__ = spectrogram( __UpperCAmelCase , window=self.window , frame_length=self.sample_size , hop_length=self.sample_stride , fft_length=self.n_fft , mel_filters=self.mel_filters , mel_floor=self.mel_floor , log_mel="log10" , ) return log_mel_spec.T def __call__(self : Any , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Optional[Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]]] = None , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , __UpperCAmelCase : Optional[int] = None , **__UpperCAmelCase : str , ) -> BatchFeature: """simple docstring""" if audio is None and audio_target is None: raise ValueError("You must provide either `audio` or `audio_target` values." ) if sampling_rate is not None: if sampling_rate != self.sampling_rate: raise ValueError( f"""The model corresponding to this feature extractor: {self} was trained using a sampling rate of""" f""" {self.sampling_rate}. Please make sure that the provided audio input was sampled with""" f""" {self.sampling_rate} and not {sampling_rate}.""" ) else: logger.warning( "It is strongly recommended to pass the ``sampling_rate`` argument to this function. " "Failing to do so can result in silent errors that might be hard to debug." ) if audio is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) else: UpperCAmelCase__ = None if audio_target is not None: UpperCAmelCase__ = self._process_audio( __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase , ) if inputs is None: return inputs_target else: UpperCAmelCase__ = inputs_target["input_values"] UpperCAmelCase__ = inputs_target.get("attention_mask" ) if decoder_attention_mask is not None: UpperCAmelCase__ = decoder_attention_mask return inputs def lowercase_ (self : Optional[int] , __UpperCAmelCase : Union[np.ndarray, List[float], List[np.ndarray], List[List[float]]] , __UpperCAmelCase : bool = False , __UpperCAmelCase : Union[bool, str, PaddingStrategy] = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : bool = False , __UpperCAmelCase : Optional[int] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[Union[str, TensorType]] = None , **__UpperCAmelCase : Any , ) -> BatchFeature: """simple docstring""" UpperCAmelCase__ = isinstance(__UpperCAmelCase , np.ndarray ) and len(speech.shape ) > 1 if is_batched_numpy and len(speech.shape ) > 2: raise ValueError(f"""Only mono-channel audio is supported for input to {self}""" ) UpperCAmelCase__ = is_batched_numpy or ( isinstance(__UpperCAmelCase , (list, tuple) ) and (isinstance(speech[0] , (np.ndarray, tuple, list) )) ) if is_batched: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for speech in speech] elif not is_batched and not isinstance(__UpperCAmelCase , np.ndarray ): UpperCAmelCase__ = np.asarray(__UpperCAmelCase , dtype=np.floataa ) elif isinstance(__UpperCAmelCase , np.ndarray ) and speech.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = speech.astype(np.floataa ) # always return batch if not is_batched: UpperCAmelCase__ = [speech] # needed to make pad() work on spectrogram inputs UpperCAmelCase__ = self.feature_size # convert into correct format for padding if is_target: UpperCAmelCase__ = [self._extract_mel_features(__UpperCAmelCase ) for waveform in speech] UpperCAmelCase__ = BatchFeature({"input_values": features} ) UpperCAmelCase__ = self.num_mel_bins else: UpperCAmelCase__ = BatchFeature({"input_values": speech} ) UpperCAmelCase__ = self.pad( __UpperCAmelCase , padding=__UpperCAmelCase , max_length=__UpperCAmelCase , truncation=__UpperCAmelCase , pad_to_multiple_of=__UpperCAmelCase , return_attention_mask=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = feature_size_hack # convert input values to correct format UpperCAmelCase__ = padded_inputs["input_values"] if not isinstance(input_values[0] , np.ndarray ): UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.floataa ) for array in input_values] elif ( not isinstance(__UpperCAmelCase , np.ndarray ) and isinstance(input_values[0] , np.ndarray ) and input_values[0].dtype is np.dtype(np.floataa ) ): UpperCAmelCase__ = [array.astype(np.floataa ) for array in input_values] elif isinstance(__UpperCAmelCase , np.ndarray ) and input_values.dtype is np.dtype(np.floataa ): UpperCAmelCase__ = input_values.astype(np.floataa ) # convert attention_mask to correct format UpperCAmelCase__ = padded_inputs.get("attention_mask" ) if attention_mask is not None: UpperCAmelCase__ = [np.asarray(__UpperCAmelCase , dtype=np.intaa ) for array in attention_mask] # zero-mean and unit-variance normalization if not is_target and self.do_normalize: UpperCAmelCase__ = ( attention_mask if self._get_padding_strategies(__UpperCAmelCase , max_length=__UpperCAmelCase ) is not PaddingStrategy.DO_NOT_PAD else None ) UpperCAmelCase__ = self.zero_mean_unit_var_norm( padded_inputs["input_values"] , attention_mask=__UpperCAmelCase , padding_value=self.padding_value ) if return_tensors is not None: UpperCAmelCase__ = padded_inputs.convert_to_tensors(__UpperCAmelCase ) return padded_inputs def lowercase_ (self : Tuple ) -> Dict[str, Any]: """simple docstring""" UpperCAmelCase__ = super().to_dict() # Don't serialize these as they are derived from the other properties. UpperCAmelCase__ = ["window", "mel_filters", "sample_size", "sample_stride", "n_fft", "n_freqs"] for name in names: if name in output: del output[name] return output
65
1
import sys UpperCamelCase__ = ( '73167176531330624919225119674426574742355349194934' '96983520312774506326239578318016984801869478851843' '85861560789112949495459501737958331952853208805511' '12540698747158523863050715693290963295227443043557' '66896648950445244523161731856403098711121722383113' '62229893423380308135336276614282806444486645238749' '30358907296290491560440772390713810515859307960866' '70172427121883998797908792274921901699720888093776' '65727333001053367881220235421809751254540594752243' '52584907711670556013604839586446706324415722155397' '53697817977846174064955149290862569321978468622482' '83972241375657056057490261407972968652414535100474' '82166370484403199890008895243450658541227588666881' '16427171479924442928230863465674813919123162824586' '17866458359124566529476545682848912883142607690042' '24219022671055626321111109370544217506941658960408' '07198403850962455444362981230987879927244284909188' '84580156166097919133875499200524063689912560717606' '05886116467109405077541002256983155200055935729725' '71636269561882670428252483600823257530420752963450' ) def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = 1 for digit in s: product *= int(__A ) return product def lowerCAmelCase_ ( __A = N ) -> int: '''simple docstring''' UpperCAmelCase__ = -sys.maxsize - 1 UpperCAmelCase__ = n[:13] UpperCAmelCase__ = 13 while cur_index < len(__A ) - 13: if int(n[cur_index] ) >= int(substr[0] ): UpperCAmelCase__ = substr[1:] + n[cur_index] cur_index += 1 else: UpperCAmelCase__ = max(__A, str_eval(__A ) ) UpperCAmelCase__ = n[cur_index : cur_index + 13] cur_index += 13 return largest_product if __name__ == "__main__": print(f'''{solution() = }''')
65
from dataclasses import dataclass from typing import Optional, Tuple import torch from torch import nn from transformers import RobertaPreTrainedModel, XLMRobertaConfig, XLMRobertaModel from transformers.utils import ModelOutput @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Optional[torch.FloatTensor] = None __UpperCAmelCase : torch.FloatTensor = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None __UpperCAmelCase : Optional[Tuple[torch.FloatTensor]] = None class A ( UpperCAmelCase_ ): def __init__(self : Union[str, Any] , __UpperCAmelCase : Tuple=1 , __UpperCAmelCase : str=0 , __UpperCAmelCase : str=2 , __UpperCAmelCase : Union[str, Any]=5_1_2 , __UpperCAmelCase : List[str]="cls" , __UpperCAmelCase : Optional[int]=False , __UpperCAmelCase : str=True , **__UpperCAmelCase : str , ) -> int: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , bos_token_id=__UpperCAmelCase , eos_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = project_dim UpperCAmelCase__ = pooler_fn UpperCAmelCase__ = learn_encoder UpperCAmelCase__ = use_attention_mask class A ( UpperCAmelCase_ ): __UpperCAmelCase : Tuple = [r'pooler', r'logit_scale'] __UpperCAmelCase : int = [r'position_ids', r'predictions.decoder.bias'] __UpperCAmelCase : Any = 'roberta' __UpperCAmelCase : List[str] = RobertaSeriesConfig def __init__(self : Tuple , __UpperCAmelCase : Optional[int] ) -> int: """simple docstring""" super().__init__(__UpperCAmelCase ) UpperCAmelCase__ = XLMRobertaModel(__UpperCAmelCase ) UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = getattr(__UpperCAmelCase , "has_pre_transformation" , __UpperCAmelCase ) if self.has_pre_transformation: UpperCAmelCase__ = nn.Linear(config.hidden_size , config.project_dim ) UpperCAmelCase__ = nn.LayerNorm(config.hidden_size , eps=config.layer_norm_eps ) self.post_init() def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[torch.Tensor] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , __UpperCAmelCase : Optional[bool] = None , ) -> Optional[int]: """simple docstring""" UpperCAmelCase__ = return_dict if return_dict is not None else self.config.use_return_dict UpperCAmelCase__ = self.base_model( input_ids=__UpperCAmelCase , attention_mask=__UpperCAmelCase , token_type_ids=__UpperCAmelCase , position_ids=__UpperCAmelCase , head_mask=__UpperCAmelCase , inputs_embeds=__UpperCAmelCase , encoder_hidden_states=__UpperCAmelCase , encoder_attention_mask=__UpperCAmelCase , output_attentions=__UpperCAmelCase , output_hidden_states=True if self.has_pre_transformation else output_hidden_states , return_dict=__UpperCAmelCase , ) if self.has_pre_transformation: UpperCAmelCase__ = outputs["hidden_states"][-2] UpperCAmelCase__ = self.pre_LN(__UpperCAmelCase ) UpperCAmelCase__ = self.transformation_pre(__UpperCAmelCase ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , ) else: UpperCAmelCase__ = self.transformation(outputs.last_hidden_state ) return TransformationModelOutput( projection_state=__UpperCAmelCase , last_hidden_state=outputs.last_hidden_state , hidden_states=outputs.hidden_states , attentions=outputs.attentions , )
65
1
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface import HuggingFace @pytest.mark.skipif( literal_eval(os.getenv('TEST_SAGEMAKER' , 'False' ) ) is not True , reason='Skipping test because should only be run when releasing minor transformers version' , ) @pytest.mark.usefixtures('sm_env' ) @parameterized_class( [ { 'framework': 'pytorch', 'script': 'run_glue.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_50, 'eval_accuracy': 0.6, 'eval_loss': 0.9}, }, { 'framework': 'tensorflow', 'script': 'run_tf.py', 'model_name_or_path': 'distilbert-base-cased', 'instance_type': 'ml.g4dn.xlarge', 'results': {'train_runtime': 6_00, 'eval_accuracy': 0.3, 'eval_loss': 0.9}, }, ] ) class A ( unittest.TestCase ): def lowercase_ (self : int ) -> Optional[Any]: """simple docstring""" if self.framework == "pytorch": subprocess.run( f"""cp ./examples/pytorch/text-classification/run_glue.py {self.env.test_path}/run_glue.py""".split() , encoding="utf-8" , check=__UpperCAmelCase , ) assert hasattr(self , "env" ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int]=1 ) -> Dict: """simple docstring""" return HuggingFace( entry_point=self.script , source_dir=self.env.test_path , role=self.env.role , image_uri=self.env.image_uri , base_job_name=f"""{self.env.base_job_name}-single""" , instance_count=__UpperCAmelCase , instance_type=self.instance_type , debugger_hook_config=__UpperCAmelCase , hyperparameters={**self.env.hyperparameters, "model_name_or_path": self.model_name_or_path} , metric_definitions=self.env.metric_definitions , py_version="py36" , ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Tuple ) -> Optional[int]: """simple docstring""" TrainingJobAnalytics(__UpperCAmelCase ).export_csv(f"""{self.env.test_path}/{job_name}_metrics.csv""" ) def lowercase_ (self : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.create_estimator() # run training estimator.fit() # result dataframe UpperCAmelCase__ = TrainingJobAnalytics(estimator.latest_training_job.name ).dataframe() # extract kpis UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_accuracy"]["value"] ) UpperCAmelCase__ = list(result_metrics_df[result_metrics_df.metric_name == "eval_loss"]["value"] ) # get train time from SageMaker job, this includes starting, preprocessing, stopping UpperCAmelCase__ = ( Session().describe_training_job(estimator.latest_training_job.name ).get("TrainingTimeInSeconds" , 9_9_9_9_9_9 ) ) # assert kpis assert train_runtime <= self.results["train_runtime"] assert all(t >= self.results["eval_accuracy"] for t in eval_accuracy ) assert all(t <= self.results["eval_loss"] for t in eval_loss ) # dump tests result into json file to share in PR with open(f"""{estimator.latest_training_job.name}.json""" , "w" ) as outfile: json.dump({"train_time": train_runtime, "eval_accuracy": eval_accuracy, "eval_loss": eval_loss} , __UpperCAmelCase )
65
1
import os from shutil import copyfile from typing import Any, Dict, List, Optional, Tuple import sentencepiece as spm from ...tokenization_utils import AddedToken, BatchEncoding, PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = '▁' UpperCamelCase__ = {'vocab_file': 'sentencepiece.bpe.model'} UpperCamelCase__ = { 'vocab_file': { 'facebook/mbart-large-en-ro': ( 'https://huggingface.co/facebook/mbart-large-en-ro/resolve/main/sentencepiece.bpe.model' ), 'facebook/mbart-large-cc25': ( 'https://huggingface.co/facebook/mbart-large-cc25/resolve/main/sentencepiece.bpe.model' ), } } UpperCamelCase__ = { 'facebook/mbart-large-en-ro': 1_0_2_4, 'facebook/mbart-large-cc25': 1_0_2_4, } # fmt: off UpperCamelCase__ = ['ar_AR', 'cs_CZ', 'de_DE', 'en_XX', 'es_XX', 'et_EE', 'fi_FI', 'fr_XX', 'gu_IN', 'hi_IN', 'it_IT', 'ja_XX', 'kk_KZ', 'ko_KR', 'lt_LT', 'lv_LV', 'my_MM', 'ne_NP', 'nl_XX', 'ro_RO', 'ru_RU', 'si_LK', 'tr_TR', 'vi_VN', 'zh_CN'] class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : List[Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Optional[int] = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : str = ['input_ids', 'attention_mask'] __UpperCAmelCase : List[int] = [] __UpperCAmelCase : List[int] = [] def __init__(self : Tuple , __UpperCAmelCase : List[str] , __UpperCAmelCase : Tuple="<s>" , __UpperCAmelCase : List[Any]="</s>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : str="<s>" , __UpperCAmelCase : Any="<unk>" , __UpperCAmelCase : Union[str, Any]="<pad>" , __UpperCAmelCase : Optional[int]="<mask>" , __UpperCAmelCase : List[str]=None , __UpperCAmelCase : Dict=None , __UpperCAmelCase : List[Any]=None , __UpperCAmelCase : Optional[Dict[str, Any]] = None , __UpperCAmelCase : Tuple=None , **__UpperCAmelCase : str , ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = AddedToken(__UpperCAmelCase , lstrip=__UpperCAmelCase , rstrip=__UpperCAmelCase ) if isinstance(__UpperCAmelCase , __UpperCAmelCase ) else mask_token UpperCAmelCase__ = {} if sp_model_kwargs is None else sp_model_kwargs super().__init__( bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , unk_token=__UpperCAmelCase , sep_token=__UpperCAmelCase , cls_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , mask_token=__UpperCAmelCase , tokenizer_file=__UpperCAmelCase , src_lang=__UpperCAmelCase , tgt_lang=__UpperCAmelCase , additional_special_tokens=__UpperCAmelCase , sp_model_kwargs=self.sp_model_kwargs , **__UpperCAmelCase , ) UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.Load(str(__UpperCAmelCase ) ) UpperCAmelCase__ = vocab_file # Original fairseq vocab and spm vocab must be "aligned": # Vocab | 0 | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 # -------- | ------- | ------- | ------ | ------- | --- | --- | --- | ----- | ----- | ---- # fairseq | '<s>' | '<pad>' | '</s>' | '<unk>' | ',' | '.' | '▁' | 's' | '▁de' | '-' # spm | '<unk>' | '<s>' | '</s>' | ',' | '.' | '▁' | 's' | '▁de' | '-' | '▁a' # Mimic fairseq token-to-id alignment for the first 4 token UpperCAmelCase__ = {"<s>": 0, "<pad>": 1, "</s>": 2, "<unk>": 3} # The first "real" token "," has position 4 in the original fairseq vocab and position 3 in the spm vocab UpperCAmelCase__ = 1 UpperCAmelCase__ = len(self.sp_model ) UpperCAmelCase__ = { code: self.sp_model_size + i + self.fairseq_offset for i, code in enumerate(__UpperCAmelCase ) } UpperCAmelCase__ = {v: k for k, v in self.lang_code_to_id.items()} UpperCAmelCase__ = len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset self.fairseq_tokens_to_ids.update(self.lang_code_to_id ) UpperCAmelCase__ = {v: k for k, v in self.fairseq_tokens_to_ids.items()} UpperCAmelCase__ = list(self.lang_code_to_id.keys() ) if additional_special_tokens is not None: # Only add those special tokens if they are not already there. self._additional_special_tokens.extend( [t for t in additional_special_tokens if t not in self._additional_special_tokens] ) UpperCAmelCase__ = src_lang if src_lang is not None else "en_XX" UpperCAmelCase__ = self.lang_code_to_id[self._src_lang] UpperCAmelCase__ = tgt_lang self.set_src_lang_special_tokens(self._src_lang ) def __getstate__(self : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = self.__dict__.copy() UpperCAmelCase__ = None UpperCAmelCase__ = self.sp_model.serialized_model_proto() return state def __setstate__(self : int , __UpperCAmelCase : int ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = d # for backward compatibility if not hasattr(self , "sp_model_kwargs" ): UpperCAmelCase__ = {} UpperCAmelCase__ = spm.SentencePieceProcessor(**self.sp_model_kwargs ) self.sp_model.LoadFromSerializedProto(self.sp_model_proto ) @property def lowercase_ (self : int ) -> Optional[int]: """simple docstring""" return len(self.sp_model ) + len(self.lang_code_to_id ) + self.fairseq_offset + 1 # Plus 1 for the mask token @property def lowercase_ (self : str ) -> str: """simple docstring""" return self._src_lang @src_lang.setter def lowercase_ (self : Any , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = new_src_lang self.set_src_lang_special_tokens(self._src_lang ) def lowercase_ (self : int , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None , __UpperCAmelCase : bool = False ) -> List[int]: """simple docstring""" if already_has_special_tokens: return super().get_special_tokens_mask( token_ids_a=__UpperCAmelCase , token_ids_a=__UpperCAmelCase , already_has_special_tokens=__UpperCAmelCase ) UpperCAmelCase__ = [1] * len(self.prefix_tokens ) UpperCAmelCase__ = [1] * len(self.suffix_tokens ) if token_ids_a is None: return prefix_ones + ([0] * len(__UpperCAmelCase )) + suffix_ones return prefix_ones + ([0] * len(__UpperCAmelCase )) + ([0] * len(__UpperCAmelCase )) + suffix_ones def lowercase_ (self : str , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" if token_ids_a is None: return self.prefix_tokens + token_ids_a + self.suffix_tokens # We don't expect to process pairs, but leave the pair logic for API consistency return self.prefix_tokens + token_ids_a + token_ids_a + self.suffix_tokens def lowercase_ (self : Dict , __UpperCAmelCase : List[int] , __UpperCAmelCase : Optional[List[int]] = None ) -> List[int]: """simple docstring""" UpperCAmelCase__ = [self.sep_token_id] UpperCAmelCase__ = [self.cls_token_id] if token_ids_a is None: return len(cls + token_ids_a + sep ) * [0] return len(cls + token_ids_a + sep + sep + token_ids_a + sep ) * [0] def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] , __UpperCAmelCase : Optional[str] , **__UpperCAmelCase : int ) -> str: """simple docstring""" if src_lang is None or tgt_lang is None: raise ValueError("Translation requires a `src_lang` and a `tgt_lang` for this model" ) UpperCAmelCase__ = src_lang UpperCAmelCase__ = self(__UpperCAmelCase , add_special_tokens=__UpperCAmelCase , return_tensors=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = self.convert_tokens_to_ids(__UpperCAmelCase ) UpperCAmelCase__ = tgt_lang_id return inputs def lowercase_ (self : Dict ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = {self.convert_ids_to_tokens(__UpperCAmelCase ): i for i in range(self.vocab_size )} vocab.update(self.added_tokens_encoder ) return vocab def lowercase_ (self : List[Any] , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" return self.sp_model.encode(__UpperCAmelCase , out_type=__UpperCAmelCase ) def lowercase_ (self : List[str] , __UpperCAmelCase : str ) -> int: """simple docstring""" if token in self.fairseq_tokens_to_ids: return self.fairseq_tokens_to_ids[token] UpperCAmelCase__ = self.sp_model.PieceToId(__UpperCAmelCase ) # Need to return unknown token if the SP model returned 0 return spm_id + self.fairseq_offset if spm_id else self.unk_token_id def lowercase_ (self : str , __UpperCAmelCase : List[Any] ) -> Optional[int]: """simple docstring""" if index in self.fairseq_ids_to_tokens: return self.fairseq_ids_to_tokens[index] return self.sp_model.IdToPiece(index - self.fairseq_offset ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Any ) -> Optional[Any]: """simple docstring""" UpperCAmelCase__ = "".join(__UpperCAmelCase ).replace(__UpperCAmelCase , " " ).strip() return out_string def lowercase_ (self : str , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) if os.path.abspath(self.vocab_file ) != os.path.abspath(__UpperCAmelCase ) and os.path.isfile(self.vocab_file ): copyfile(self.vocab_file , __UpperCAmelCase ) elif not os.path.isfile(self.vocab_file ): with open(__UpperCAmelCase , "wb" ) as fi: UpperCAmelCase__ = self.sp_model.serialized_model_proto() fi.write(__UpperCAmelCase ) return (out_vocab_file,) def lowercase_ (self : List[Any] , __UpperCAmelCase : List[str] , __UpperCAmelCase : str = "en_XX" , __UpperCAmelCase : Optional[List[str]] = None , __UpperCAmelCase : str = "ro_RO" , **__UpperCAmelCase : Union[str, Any] , ) -> BatchEncoding: """simple docstring""" UpperCAmelCase__ = src_lang UpperCAmelCase__ = tgt_lang return super().prepare_seqaseq_batch(__UpperCAmelCase , __UpperCAmelCase , **__UpperCAmelCase ) def lowercase_ (self : Any ) -> List[Any]: """simple docstring""" return self.set_src_lang_special_tokens(self.src_lang ) def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" return self.set_tgt_lang_special_tokens(self.tgt_lang ) def lowercase_ (self : List[Any] , __UpperCAmelCase : Optional[int] ) -> None: """simple docstring""" UpperCAmelCase__ = self.lang_code_to_id[src_lang] UpperCAmelCase__ = [] UpperCAmelCase__ = [self.eos_token_id, self.cur_lang_code] def lowercase_ (self : int , __UpperCAmelCase : str ) -> None: """simple docstring""" UpperCAmelCase__ = self.lang_code_to_id[lang] UpperCAmelCase__ = [] UpperCAmelCase__ = [self.eos_token_id, self.cur_lang_code]
65
import math import random def lowerCAmelCase_ ( __A, __A = False ) -> float: '''simple docstring''' if deriv: return value * (1 - value) return 1 / (1 + math.exp(-value )) # Initial Value UpperCamelCase__ = 0.0_2 def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' UpperCAmelCase__ = float(2 * (random.randint(1, 100 )) - 1 ) for _ in range(__A ): # Forward propagation UpperCAmelCase__ = sigmoid_function(INITIAL_VALUE * weight ) # How much did we miss? UpperCAmelCase__ = (expected / 100) - layer_a # Error delta UpperCAmelCase__ = layer_1_error * sigmoid_function(__A, __A ) # Update weight weight += INITIAL_VALUE * layer_1_delta return layer_a * 100 if __name__ == "__main__": import doctest doctest.testmod() UpperCamelCase__ = int(input('Expected value: ')) UpperCamelCase__ = int(input('Number of propagations: ')) print(forward_propagation(expected, number_propagations))
65
1
import copy import os from typing import Union from ...configuration_utils import PretrainedConfig from ...models.auto.modeling_auto import MODEL_FOR_CAUSAL_LM_MAPPING_NAMES from ...utils import logging from ..auto import CONFIG_MAPPING UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'Salesforce/instruct-blip-flan-t5': 'https://huggingface.co/Salesforce/instruct-blip-flan-t5/resolve/main/config.json', } class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 'instructblip_vision_model' def __init__(self : Union[str, Any] , __UpperCAmelCase : Optional[int]=1_4_0_8 , __UpperCAmelCase : List[Any]=6_1_4_4 , __UpperCAmelCase : Optional[int]=3_9 , __UpperCAmelCase : Optional[Any]=1_6 , __UpperCAmelCase : Optional[Any]=2_2_4 , __UpperCAmelCase : Optional[Any]=1_4 , __UpperCAmelCase : Dict="gelu" , __UpperCAmelCase : Tuple=1E-6 , __UpperCAmelCase : str=0.0 , __UpperCAmelCase : Tuple=1E-10 , __UpperCAmelCase : int=True , **__UpperCAmelCase : Union[str, Any] , ) -> Dict: """simple docstring""" super().__init__(**__UpperCAmelCase ) UpperCAmelCase__ = hidden_size UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = patch_size UpperCAmelCase__ = image_size UpperCAmelCase__ = initializer_range UpperCAmelCase__ = attention_dropout UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = hidden_act UpperCAmelCase__ = qkv_bias @classmethod def lowercase_ (cls : Union[str, Any] , __UpperCAmelCase : Union[str, os.PathLike] , **__UpperCAmelCase : Tuple ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = cls.get_config_dict(__UpperCAmelCase , **__UpperCAmelCase ) # get the vision config dict if we are loading from InstructBlipConfig if config_dict.get("model_type" ) == "instructblip": UpperCAmelCase__ = config_dict["vision_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__UpperCAmelCase , **__UpperCAmelCase ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : List[Any] = 'instructblip_qformer' def __init__(self : str , __UpperCAmelCase : Tuple=3_0_5_2_2 , __UpperCAmelCase : Dict=7_6_8 , __UpperCAmelCase : Any=1_2 , __UpperCAmelCase : List[Any]=1_2 , __UpperCAmelCase : Any=3_0_7_2 , __UpperCAmelCase : List[str]="gelu" , __UpperCAmelCase : Union[str, Any]=0.1 , __UpperCAmelCase : Optional[int]=0.1 , __UpperCAmelCase : Optional[int]=5_1_2 , __UpperCAmelCase : List[str]=0.02 , __UpperCAmelCase : List[str]=1E-12 , __UpperCAmelCase : Union[str, Any]=0 , __UpperCAmelCase : Union[str, Any]="absolute" , __UpperCAmelCase : List[Any]=2 , __UpperCAmelCase : Any=1_4_0_8 , **__UpperCAmelCase : List[Any] , ) -> Dict: """simple docstring""" super().__init__(pad_token_id=__UpperCAmelCase , **__UpperCAmelCase ) UpperCAmelCase__ = vocab_size UpperCAmelCase__ = hidden_size UpperCAmelCase__ = num_hidden_layers UpperCAmelCase__ = num_attention_heads UpperCAmelCase__ = hidden_act UpperCAmelCase__ = intermediate_size UpperCAmelCase__ = hidden_dropout_prob UpperCAmelCase__ = attention_probs_dropout_prob UpperCAmelCase__ = max_position_embeddings UpperCAmelCase__ = initializer_range UpperCAmelCase__ = layer_norm_eps UpperCAmelCase__ = position_embedding_type UpperCAmelCase__ = cross_attention_frequency UpperCAmelCase__ = encoder_hidden_size @classmethod def lowercase_ (cls : Optional[Any] , __UpperCAmelCase : Union[str, os.PathLike] , **__UpperCAmelCase : Dict ) -> "PretrainedConfig": """simple docstring""" cls._set_token_in_kwargs(__UpperCAmelCase ) UpperCAmelCase__ , UpperCAmelCase__ = cls.get_config_dict(__UpperCAmelCase , **__UpperCAmelCase ) # get the qformer config dict if we are loading from InstructBlipConfig if config_dict.get("model_type" ) == "instructblip": UpperCAmelCase__ = config_dict["qformer_config"] if "model_type" in config_dict and hasattr(cls , "model_type" ) and config_dict["model_type"] != cls.model_type: logger.warning( f"""You are using a model of type {config_dict["model_type"]} to instantiate a model of type """ f"""{cls.model_type}. This is not supported for all configurations of models and can yield errors.""" ) return cls.from_dict(__UpperCAmelCase , **__UpperCAmelCase ) class A ( UpperCAmelCase_ ): __UpperCAmelCase : Dict = 'instructblip' __UpperCAmelCase : Optional[int] = True def __init__(self : Tuple , __UpperCAmelCase : Union[str, Any]=None , __UpperCAmelCase : int=None , __UpperCAmelCase : int=None , __UpperCAmelCase : str=3_2 , **__UpperCAmelCase : Optional[Any] ) -> Optional[int]: """simple docstring""" super().__init__(**__UpperCAmelCase ) if vision_config is None: UpperCAmelCase__ = {} logger.info("vision_config is None. initializing the InstructBlipVisionConfig with default values." ) if qformer_config is None: UpperCAmelCase__ = {} logger.info("qformer_config is None. Initializing the InstructBlipQFormerConfig with default values." ) if text_config is None: UpperCAmelCase__ = {} logger.info("text_config is None. Initializing the text config with default values (`OPTConfig`)." ) UpperCAmelCase__ = InstructBlipVisionConfig(**__UpperCAmelCase ) UpperCAmelCase__ = InstructBlipQFormerConfig(**__UpperCAmelCase ) UpperCAmelCase__ = text_config["model_type"] if "model_type" in text_config else "opt" UpperCAmelCase__ = CONFIG_MAPPING[text_model_type](**__UpperCAmelCase ) UpperCAmelCase__ = self.text_config.tie_word_embeddings UpperCAmelCase__ = self.text_config.is_encoder_decoder UpperCAmelCase__ = num_query_tokens UpperCAmelCase__ = self.vision_config.hidden_size UpperCAmelCase__ = self.text_config.model_type in MODEL_FOR_CAUSAL_LM_MAPPING_NAMES UpperCAmelCase__ = 1.0 UpperCAmelCase__ = 0.02 @classmethod def lowercase_ (cls : List[Any] , __UpperCAmelCase : InstructBlipVisionConfig , __UpperCAmelCase : InstructBlipQFormerConfig , __UpperCAmelCase : PretrainedConfig , **__UpperCAmelCase : List[Any] , ) -> Tuple: """simple docstring""" return cls( vision_config=vision_config.to_dict() , qformer_config=qformer_config.to_dict() , text_config=text_config.to_dict() , **__UpperCAmelCase , ) def lowercase_ (self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = copy.deepcopy(self.__dict__ ) UpperCAmelCase__ = self.vision_config.to_dict() UpperCAmelCase__ = self.qformer_config.to_dict() UpperCAmelCase__ = self.text_config.to_dict() UpperCAmelCase__ = self.__class__.model_type return output
65
from __future__ import annotations class A : def __init__(self : Union[str, Any] , __UpperCAmelCase : list[list[int]] ) -> List[str]: """simple docstring""" UpperCAmelCase__ = TypeError( "Matrices must be formed from a list of zero or more lists containing at " "least one and the same number of values, each of which must be of type " "int or float." ) if len(__UpperCAmelCase ) != 0: UpperCAmelCase__ = len(rows[0] ) if cols == 0: raise error for row in rows: if len(__UpperCAmelCase ) != cols: raise error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise error UpperCAmelCase__ = rows else: UpperCAmelCase__ = [] def lowercase_ (self : Any ) -> list[list[int]]: """simple docstring""" return [[row[i] for row in self.rows] for i in range(len(self.rows[0] ) )] @property def lowercase_ (self : Any ) -> int: """simple docstring""" return len(self.rows ) @property def lowercase_ (self : Union[str, Any] ) -> int: """simple docstring""" return len(self.rows[0] ) @property def lowercase_ (self : List[Any] ) -> tuple[int, int]: """simple docstring""" return (self.num_rows, self.num_columns) @property def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return self.order[0] == self.order[1] def lowercase_ (self : Any ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [0 if column_num != row_num else 1 for column_num in range(self.num_rows )] for row_num in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : int ) -> int: """simple docstring""" if not self.is_square: return 0 if self.order == (0, 0): return 1 if self.order == (1, 1): return int(self.rows[0][0] ) if self.order == (2, 2): return int( (self.rows[0][0] * self.rows[1][1]) - (self.rows[0][1] * self.rows[1][0]) ) else: return sum( self.rows[0][column] * self.cofactors().rows[0][column] for column in range(self.num_columns ) ) def lowercase_ (self : Tuple ) -> bool: """simple docstring""" return bool(self.determinant() ) def lowercase_ (self : Dict , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" UpperCAmelCase__ = [ [ self.rows[other_row][other_column] for other_column in range(self.num_columns ) if other_column != column ] for other_row in range(self.num_rows ) if other_row != row ] return Matrix(__UpperCAmelCase ).determinant() def lowercase_ (self : int , __UpperCAmelCase : int , __UpperCAmelCase : int ) -> int: """simple docstring""" if (row + column) % 2 == 0: return self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) return -1 * self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) def lowercase_ (self : Union[str, Any] ) -> Matrix: """simple docstring""" return Matrix( [ [self.get_minor(__UpperCAmelCase , __UpperCAmelCase ) for column in range(self.num_columns )] for row in range(self.num_rows ) ] ) def lowercase_ (self : List[str] ) -> Matrix: """simple docstring""" return Matrix( [ [ self.minors().rows[row][column] if (row + column) % 2 == 0 else self.minors().rows[row][column] * -1 for column in range(self.minors().num_columns ) ] for row in range(self.minors().num_rows ) ] ) def lowercase_ (self : Optional[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = [ [self.cofactors().rows[column][row] for column in range(self.num_columns )] for row in range(self.num_rows ) ] return Matrix(__UpperCAmelCase ) def lowercase_ (self : List[Any] ) -> Matrix: """simple docstring""" UpperCAmelCase__ = self.determinant() if not determinant: raise TypeError("Only matrices with a non-zero determinant have an inverse" ) return self.adjugate() * (1 / determinant) def __repr__(self : Dict ) -> str: """simple docstring""" return str(self.rows ) def __str__(self : Optional[Any] ) -> str: """simple docstring""" if self.num_rows == 0: return "[]" if self.num_rows == 1: return "[[" + ". ".join(str(self.rows[0] ) ) + "]]" return ( "[" + "\n ".join( [ "[" + ". ".join([str(__UpperCAmelCase ) for value in row] ) + ".]" for row in self.rows ] ) + "]" ) def lowercase_ (self : Optional[int] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError("Row must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in row: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_columns: raise ValueError( "Row must be equal in length to the other rows in the matrix" ) if position is None: self.rows.append(__UpperCAmelCase ) else: UpperCAmelCase__ = self.rows[0:position] + [row] + self.rows[position:] def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : list[int] , __UpperCAmelCase : int | None = None ) -> None: """simple docstring""" UpperCAmelCase__ = TypeError( "Column must be a list containing all ints and/or floats" ) if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise type_error for value in column: if not isinstance(__UpperCAmelCase , (int, float) ): raise type_error if len(__UpperCAmelCase ) != self.num_rows: raise ValueError( "Column must be equal in length to the other columns in the matrix" ) if position is None: UpperCAmelCase__ = [self.rows[i] + [column[i]] for i in range(self.num_rows )] else: UpperCAmelCase__ = [ self.rows[i][0:position] + [column[i]] + self.rows[i][position:] for i in range(self.num_rows ) ] def __eq__(self : Any , __UpperCAmelCase : object ) -> bool: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): return NotImplemented return self.rows == other.rows def __ne__(self : int , __UpperCAmelCase : object ) -> bool: """simple docstring""" return not self == other def __neg__(self : Dict ) -> Matrix: """simple docstring""" return self * -1 def __add__(self : Dict , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Addition requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] + other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __sub__(self : Optional[Any] , __UpperCAmelCase : Matrix ) -> Matrix: """simple docstring""" if self.order != other.order: raise ValueError("Subtraction requires matrices of the same order" ) return Matrix( [ [self.rows[i][j] - other.rows[i][j] for j in range(self.num_columns )] for i in range(self.num_rows ) ] ) def __mul__(self : Tuple , __UpperCAmelCase : Matrix | int | float ) -> Matrix: """simple docstring""" if isinstance(__UpperCAmelCase , (int, float) ): return Matrix( [[int(element * other ) for element in row] for row in self.rows] ) elif isinstance(__UpperCAmelCase , __UpperCAmelCase ): if self.num_columns != other.num_rows: raise ValueError( "The number of columns in the first matrix must " "be equal to the number of rows in the second" ) return Matrix( [ [Matrix.dot_product(__UpperCAmelCase , __UpperCAmelCase ) for column in other.columns()] for row in self.rows ] ) else: raise TypeError( "A Matrix can only be multiplied by an int, float, or another matrix" ) def __pow__(self : List[Any] , __UpperCAmelCase : int ) -> Matrix: """simple docstring""" if not isinstance(__UpperCAmelCase , __UpperCAmelCase ): raise TypeError("A Matrix can only be raised to the power of an int" ) if not self.is_square: raise ValueError("Only square matrices can be raised to a power" ) if other == 0: return self.identity() if other < 0: if self.is_invertable(): return self.inverse() ** (-other) raise ValueError( "Only invertable matrices can be raised to a negative power" ) UpperCAmelCase__ = self for _ in range(other - 1 ): result *= self return result @classmethod def lowercase_ (cls : Dict , __UpperCAmelCase : list[int] , __UpperCAmelCase : list[int] ) -> int: """simple docstring""" return sum(row[i] * column[i] for i in range(len(__UpperCAmelCase ) ) ) if __name__ == "__main__": import doctest doctest.testmod()
65
1
import argparse import json from collections import OrderedDict import torch from huggingface_hub import cached_download, hf_hub_url from transformers import AutoImageProcessor, CvtConfig, CvtForImageClassification def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = [] embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.weight""", f"""stage{idx}.patch_embed.proj.weight""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.projection.bias""", f"""stage{idx}.patch_embed.proj.bias""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.weight""", f"""stage{idx}.patch_embed.norm.weight""", ) ) embed.append( ( f"""cvt.encoder.stages.{idx}.embedding.convolution_embeddings.normalization.bias""", f"""stage{idx}.patch_embed.norm.bias""", ) ) return embed def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = [] attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_query.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_q.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_key.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_k.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.convolution.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.conv.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.weight""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.bias""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_mean""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_mean""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.running_var""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.running_var""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.convolution_projection_value.convolution_projection.normalization.num_batches_tracked""", f"""stage{idx}.blocks.{cnt}.attn.conv_proj_v.bn.num_batches_tracked""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_q.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_query.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_q.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_k.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_key.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_k.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj_v.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.attention.projection_value.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj_v.bias""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.weight""", f"""stage{idx}.blocks.{cnt}.attn.proj.weight""", ) ) attention_weights.append( ( f"""cvt.encoder.stages.{idx}.layers.{cnt}.attention.output.dense.bias""", f"""stage{idx}.blocks.{cnt}.attn.proj.bias""", ) ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.weight""", f"""stage{idx}.blocks.{cnt}.mlp.fc1.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.intermediate.dense.bias""", f"""stage{idx}.blocks.{cnt}.mlp.fc1.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.weight""", f"""stage{idx}.blocks.{cnt}.mlp.fc2.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.output.dense.bias""", f"""stage{idx}.blocks.{cnt}.mlp.fc2.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.weight""", f"""stage{idx}.blocks.{cnt}.norm1.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_before.bias""", f"""stage{idx}.blocks.{cnt}.norm1.bias""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.weight""", f"""stage{idx}.blocks.{cnt}.norm2.weight""") ) attention_weights.append( (f"""cvt.encoder.stages.{idx}.layers.{cnt}.layernorm_after.bias""", f"""stage{idx}.blocks.{cnt}.norm2.bias""") ) return attention_weights def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' UpperCAmelCase__ = [] token.append((f"""cvt.encoder.stages.{idx}.cls_token""", "stage2.cls_token") ) return token def lowerCAmelCase_ ( ) -> int: '''simple docstring''' UpperCAmelCase__ = [] head.append(("layernorm.weight", "norm.weight") ) head.append(("layernorm.bias", "norm.bias") ) head.append(("classifier.weight", "head.weight") ) head.append(("classifier.bias", "head.bias") ) return head def lowerCAmelCase_ ( __A, __A, __A, __A ) -> Dict: '''simple docstring''' UpperCAmelCase__ = "imagenet-1k-id2label.json" UpperCAmelCase__ = 1_000 UpperCAmelCase__ = "huggingface/label-files" UpperCAmelCase__ = num_labels UpperCAmelCase__ = json.load(open(cached_download(hf_hub_url(__A, __A, repo_type="dataset" ) ), "r" ) ) UpperCAmelCase__ = {int(__A ): v for k, v in idalabel.items()} UpperCAmelCase__ = idalabel UpperCAmelCase__ = {v: k for k, v in idalabel.items()} UpperCAmelCase__ = UpperCAmelCase__ = CvtConfig(num_labels=__A, idalabel=__A, labelaid=__A ) # For depth size 13 (13 = 1+2+10) if cvt_model.rsplit("/", 1 )[-1][4:6] == "13": UpperCAmelCase__ = [1, 2, 10] # For depth size 21 (21 = 1+4+16) elif cvt_model.rsplit("/", 1 )[-1][4:6] == "21": UpperCAmelCase__ = [1, 4, 16] # For wide cvt (similar to wide-resnet) depth size 24 (w24 = 2 + 2 20) else: UpperCAmelCase__ = [2, 2, 20] UpperCAmelCase__ = [3, 12, 16] UpperCAmelCase__ = [192, 768, 1_024] UpperCAmelCase__ = CvtForImageClassification(__A ) UpperCAmelCase__ = AutoImageProcessor.from_pretrained("facebook/convnext-base-224-22k-1k" ) UpperCAmelCase__ = image_size UpperCAmelCase__ = torch.load(__A, map_location=torch.device("cpu" ) ) UpperCAmelCase__ = OrderedDict() UpperCAmelCase__ = [] for idx in range(len(config.depth ) ): if config.cls_token[idx]: UpperCAmelCase__ = list_of_state_dict + cls_token(__A ) UpperCAmelCase__ = list_of_state_dict + embeddings(__A ) for cnt in range(config.depth[idx] ): UpperCAmelCase__ = list_of_state_dict + attention(__A, __A ) UpperCAmelCase__ = list_of_state_dict + final() for gg in list_of_state_dict: print(__A ) for i in range(len(__A ) ): UpperCAmelCase__ = original_weights[list_of_state_dict[i][1]] model.load_state_dict(__A ) model.save_pretrained(__A ) image_processor.save_pretrained(__A ) # Download the weights from zoo: https://1drv.ms/u/s!AhIXJn_J-blW9RzF3rMW7SsLHa8h?e=blQ0Al if __name__ == "__main__": UpperCamelCase__ = argparse.ArgumentParser() parser.add_argument( '--cvt_model', default='cvt-w24', type=str, help='Name of the cvt model you\'d like to convert.', ) parser.add_argument( '--image_size', default=3_8_4, type=int, help='Input Image Size', ) parser.add_argument( '--cvt_file_name', default=R'cvtmodels\CvT-w24-384x384-IN-22k.pth', type=str, help='Input Image Size', ) parser.add_argument( '--pytorch_dump_folder_path', default=None, type=str, help='Path to the output PyTorch model directory.' ) UpperCamelCase__ = parser.parse_args() convert_cvt_checkpoint(args.cvt_model, args.image_size, args.cvt_file_name, args.pytorch_dump_folder_path)
65
import json import os from typing import Dict, List, Optional, Tuple from ...tokenization_utils import PreTrainedTokenizer from ...utils import logging UpperCamelCase__ = logging.get_logger(__name__) UpperCamelCase__ = { 'vocab_file': 'vocab.json', 'tokenizer_config_file': 'tokenizer_config.json', 'merges_file': 'merges.txt', } UpperCamelCase__ = { 'vocab_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/vocab.json' ), }, 'tokenizer_config_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/tokenizer_config.json' ), }, 'merges_file': { 'facebook/s2t-wav2vec2-large-en-de': ( 'https://huggingface.co/facebook/s2t-wav2vec2-large-en-de/resolve/main/merges.txt' ), }, } UpperCamelCase__ = '</w>' UpperCamelCase__ = '@@ ' def lowerCAmelCase_ ( __A ) -> str: '''simple docstring''' UpperCAmelCase__ = set() UpperCAmelCase__ = word[0] for char in word[1:]: pairs.add((prev_char, char) ) UpperCAmelCase__ = char return pairs # Speech2Text2 has no max input length UpperCamelCase__ = {'facebook/s2t-wav2vec2-large-en-de': 1_0_2_4} class A ( UpperCAmelCase_ ): __UpperCAmelCase : str = VOCAB_FILES_NAMES __UpperCAmelCase : str = PRETRAINED_VOCAB_FILES_MAP __UpperCAmelCase : Union[str, Any] = PRETRAINED_POSITIONAL_EMBEDDINGS_SIZES __UpperCAmelCase : Dict = ['input_ids', 'attention_mask'] def __init__(self : Tuple , __UpperCAmelCase : List[Any] , __UpperCAmelCase : Dict="<s>" , __UpperCAmelCase : Tuple="<pad>" , __UpperCAmelCase : str="</s>" , __UpperCAmelCase : int="<unk>" , __UpperCAmelCase : List[str]=False , __UpperCAmelCase : str=None , **__UpperCAmelCase : Optional[Any] , ) -> Tuple: """simple docstring""" super().__init__( unk_token=__UpperCAmelCase , bos_token=__UpperCAmelCase , eos_token=__UpperCAmelCase , pad_token=__UpperCAmelCase , do_lower_case=__UpperCAmelCase , **__UpperCAmelCase , ) UpperCAmelCase__ = do_lower_case with open(__UpperCAmelCase , encoding="utf-8" ) as vocab_handle: UpperCAmelCase__ = json.load(__UpperCAmelCase ) UpperCAmelCase__ = {v: k for k, v in self.encoder.items()} if merges_file is None: logger.info(f"""No merges files provided. {self.__class__.__name__} can only be used for decoding.""" ) UpperCAmelCase__ = None UpperCAmelCase__ = None else: with open(__UpperCAmelCase , encoding="utf-8" ) as merges_handle: UpperCAmelCase__ = merges_handle.read().split("\n" )[:-1] UpperCAmelCase__ = [tuple(merge.split()[:2] ) for merge in merges] UpperCAmelCase__ = dict(zip(__UpperCAmelCase , range(len(__UpperCAmelCase ) ) ) ) UpperCAmelCase__ = {} @property def lowercase_ (self : List[str] ) -> int: """simple docstring""" return len(self.decoder ) def lowercase_ (self : Union[str, Any] ) -> Dict: """simple docstring""" return dict(self.encoder , **self.added_tokens_encoder ) def lowercase_ (self : Dict , __UpperCAmelCase : Union[str, Any] ) -> str: """simple docstring""" UpperCAmelCase__ = tuple(token[:-1] ) + (token[-1] + BPE_TOKEN_MERGES,) if token in self.cache: return self.cache[token] UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) if not pairs: return token while True: UpperCAmelCase__ = min(__UpperCAmelCase , key=lambda __UpperCAmelCase : self.bpe_ranks.get(__UpperCAmelCase , float("inf" ) ) ) if bigram not in self.bpe_ranks: break UpperCAmelCase__ , UpperCAmelCase__ = bigram UpperCAmelCase__ = [] UpperCAmelCase__ = 0 while i < len(__UpperCAmelCase ): try: UpperCAmelCase__ = word.index(__UpperCAmelCase , __UpperCAmelCase ) except ValueError: new_word.extend(word[i:] ) break else: new_word.extend(word[i:j] ) UpperCAmelCase__ = j if word[i] == first and i < len(__UpperCAmelCase ) - 1 and word[i + 1] == second: new_word.append(first + second ) i += 2 else: new_word.append(word[i] ) i += 1 UpperCAmelCase__ = tuple(__UpperCAmelCase ) UpperCAmelCase__ = new_word if len(__UpperCAmelCase ) == 1: break else: UpperCAmelCase__ = get_pairs(__UpperCAmelCase ) UpperCAmelCase__ = " ".join(__UpperCAmelCase ) if word == "\n " + BPE_TOKEN_MERGES: UpperCAmelCase__ = "\n" + BPE_TOKEN_MERGES if word.endswith(__UpperCAmelCase ): UpperCAmelCase__ = word.replace(__UpperCAmelCase , "" ) UpperCAmelCase__ = word.replace(" " , __UpperCAmelCase ) UpperCAmelCase__ = word return word def lowercase_ (self : Tuple , __UpperCAmelCase : int ) -> Optional[int]: """simple docstring""" if self.bpe_ranks is None: raise ValueError( "This tokenizer was instantiated without a `merges.txt` file, so" " that it can only be used for decoding, not for encoding." "Make sure to provide `merges.txt` file at instantiation to enable " "encoding." ) if self.do_lower_case: UpperCAmelCase__ = text.lower() UpperCAmelCase__ = text.split() UpperCAmelCase__ = [] for token in text: if token: split_tokens.extend(list(self.bpe(__UpperCAmelCase ).split(" " ) ) ) return split_tokens def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str ) -> int: """simple docstring""" return self.encoder.get(__UpperCAmelCase , self.encoder.get(self.unk_token ) ) def lowercase_ (self : Any , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = self.decoder.get(__UpperCAmelCase , self.unk_token ) return result def lowercase_ (self : Dict , __UpperCAmelCase : List[str] ) -> str: """simple docstring""" UpperCAmelCase__ = " ".join(__UpperCAmelCase ) # make sure @@ tokens are concatenated UpperCAmelCase__ = "".join(string.split(__UpperCAmelCase ) ) return string def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : str , __UpperCAmelCase : Optional[str] = None ) -> Tuple[str]: """simple docstring""" if not os.path.isdir(__UpperCAmelCase ): logger.error(f"""Vocabulary path ({save_directory}) should be a directory""" ) return UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["vocab_file"] ) UpperCAmelCase__ = os.path.join( __UpperCAmelCase , (filename_prefix + "-" if filename_prefix else "") + VOCAB_FILES_NAMES["merges_file"] ) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as f: f.write(json.dumps(self.encoder , indent=2 , sort_keys=__UpperCAmelCase , ensure_ascii=__UpperCAmelCase ) + "\n" ) UpperCAmelCase__ = 0 if self.bpe_ranks is None: return (vocab_file,) with open(__UpperCAmelCase , "w" , encoding="utf-8" ) as writer: for bpe_tokens, token_index in sorted(self.bpe_ranks.items() , key=lambda __UpperCAmelCase : kv[1] ): if index != token_index: logger.warning( f"""Saving vocabulary to {merges_file}: BPE merge indices are not consecutive.""" " Please check that the tokenizer is not corrupted!" ) UpperCAmelCase__ = token_index writer.write(" ".join(__UpperCAmelCase ) + "\n" ) index += 1 return (vocab_file, merges_file)
65
1
import gc import unittest import numpy as np import torch from diffusers import DanceDiffusionPipeline, IPNDMScheduler, UNetaDModel from diffusers.utils import slow, torch_device from diffusers.utils.testing_utils import enable_full_determinism, require_torch_gpu, skip_mps from ..pipeline_params import UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS, UNCONDITIONAL_AUDIO_GENERATION_PARAMS from ..test_pipelines_common import PipelineTesterMixin enable_full_determinism() class A ( UpperCAmelCase_ , unittest.TestCase ): __UpperCAmelCase : Optional[int] = DanceDiffusionPipeline __UpperCAmelCase : Optional[Any] = UNCONDITIONAL_AUDIO_GENERATION_PARAMS __UpperCAmelCase : List[Any] = PipelineTesterMixin.required_optional_params - { 'callback', 'latents', 'callback_steps', 'output_type', 'num_images_per_prompt', } __UpperCAmelCase : Optional[Any] = UNCONDITIONAL_AUDIO_GENERATION_BATCH_PARAMS __UpperCAmelCase : str = False __UpperCAmelCase : List[str] = False def lowercase_ (self : Optional[int] ) -> Any: """simple docstring""" torch.manual_seed(0 ) UpperCAmelCase__ = UNetaDModel( block_out_channels=(3_2, 3_2, 6_4) , extra_in_channels=1_6 , sample_size=5_1_2 , sample_rate=1_6_0_0_0 , in_channels=2 , out_channels=2 , flip_sin_to_cos=__UpperCAmelCase , use_timestep_embedding=__UpperCAmelCase , time_embedding_type="fourier" , mid_block_type="UNetMidBlock1D" , down_block_types=("DownBlock1DNoSkip", "DownBlock1D", "AttnDownBlock1D") , up_block_types=("AttnUpBlock1D", "UpBlock1D", "UpBlock1DNoSkip") , ) UpperCAmelCase__ = IPNDMScheduler() UpperCAmelCase__ = { "unet": unet, "scheduler": scheduler, } return components def lowercase_ (self : str , __UpperCAmelCase : Optional[int] , __UpperCAmelCase : Optional[int]=0 ) -> Any: """simple docstring""" if str(__UpperCAmelCase ).startswith("mps" ): UpperCAmelCase__ = torch.manual_seed(__UpperCAmelCase ) else: UpperCAmelCase__ = torch.Generator(device=__UpperCAmelCase ).manual_seed(__UpperCAmelCase ) UpperCAmelCase__ = { "batch_size": 1, "generator": generator, "num_inference_steps": 4, } return inputs def lowercase_ (self : Union[str, Any] ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = "cpu" # ensure determinism for the device-dependent torch.Generator UpperCAmelCase__ = self.get_dummy_components() UpperCAmelCase__ = DanceDiffusionPipeline(**__UpperCAmelCase ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = self.get_dummy_inputs(__UpperCAmelCase ) UpperCAmelCase__ = pipe(**__UpperCAmelCase ) UpperCAmelCase__ = output.audios UpperCAmelCase__ = audio[0, -3:, -3:] assert audio.shape == (1, 2, components["unet"].sample_size) UpperCAmelCase__ = np.array([-0.7265, 1.0000, -0.8388, 0.1175, 0.9498, -1.0000] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 @skip_mps def lowercase_ (self : Tuple ) -> Any: """simple docstring""" return super().test_save_load_local() @skip_mps def lowercase_ (self : Dict ) -> List[str]: """simple docstring""" return super().test_dict_tuple_outputs_equivalent(expected_max_difference=3E-3 ) @skip_mps def lowercase_ (self : Union[str, Any] ) -> Optional[int]: """simple docstring""" return super().test_save_load_optional_components() @skip_mps def lowercase_ (self : Union[str, Any] ) -> str: """simple docstring""" return super().test_attention_slicing_forward_pass() def lowercase_ (self : List[str] ) -> Dict: """simple docstring""" super().test_inference_batch_single_identical(expected_max_diff=3E-3 ) @slow @require_torch_gpu class A ( unittest.TestCase ): def lowercase_ (self : str ) -> int: """simple docstring""" super().tearDown() gc.collect() torch.cuda.empty_cache() def lowercase_ (self : Dict ) -> str: """simple docstring""" UpperCAmelCase__ = torch_device UpperCAmelCase__ = DanceDiffusionPipeline.from_pretrained("harmonai/maestro-150k" ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe(generator=__UpperCAmelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCAmelCase__ = output.audios UpperCAmelCase__ = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCAmelCase__ = np.array([-0.0192, -0.0231, -0.0318, -0.0059, 0.0002, -0.0020] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2 def lowercase_ (self : Union[str, Any] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = torch_device UpperCAmelCase__ = DanceDiffusionPipeline.from_pretrained("harmonai/maestro-150k" , torch_dtype=torch.floataa ) UpperCAmelCase__ = pipe.to(__UpperCAmelCase ) pipe.set_progress_bar_config(disable=__UpperCAmelCase ) UpperCAmelCase__ = torch.manual_seed(0 ) UpperCAmelCase__ = pipe(generator=__UpperCAmelCase , num_inference_steps=1_0_0 , audio_length_in_s=4.096 ) UpperCAmelCase__ = output.audios UpperCAmelCase__ = audio[0, -3:, -3:] assert audio.shape == (1, 2, pipe.unet.sample_size) UpperCAmelCase__ = np.array([-0.0367, -0.0488, -0.0771, -0.0525, -0.0444, -0.0341] ) assert np.abs(audio_slice.flatten() - expected_slice ).max() < 1E-2
65
from dataclasses import dataclass from typing import Optional import numpy as np import torch import torch.nn as nn from ..utils import BaseOutput, is_torch_version, randn_tensor from .attention_processor import SpatialNorm from .unet_ad_blocks import UNetMidBlockaD, get_down_block, get_up_block @dataclass class A ( UpperCAmelCase_ ): __UpperCAmelCase : torch.FloatTensor class A ( nn.Module ): def __init__(self : Union[str, Any] , __UpperCAmelCase : int=3 , __UpperCAmelCase : Dict=3 , __UpperCAmelCase : Optional[Any]=("DownEncoderBlock2D",) , __UpperCAmelCase : int=(6_4,) , __UpperCAmelCase : Union[str, Any]=2 , __UpperCAmelCase : Any=3_2 , __UpperCAmelCase : str="silu" , __UpperCAmelCase : Any=True , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = torch.nn.Convad( __UpperCAmelCase , block_out_channels[0] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) # down UpperCAmelCase__ = block_out_channels[0] for i, down_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_down_block( __UpperCAmelCase , num_layers=self.layers_per_block , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , add_downsample=not is_final_block , resnet_eps=1E-6 , downsample_padding=0 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) self.down_blocks.append(__UpperCAmelCase ) # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # out UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[-1] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = 2 * out_channels if double_z else out_channels UpperCAmelCase__ = nn.Convad(block_out_channels[-1] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : List[Any] , __UpperCAmelCase : int ) -> str: """simple docstring""" UpperCAmelCase__ = x UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : int ): def custom_forward(*__UpperCAmelCase : Optional[Any] ): return module(*__UpperCAmelCase ) return custom_forward # down if is_torch_version(">=" , "1.11.0" ): for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: for down_block in self.down_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase ) # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(self.mid_block ) , __UpperCAmelCase ) else: # down for down_block in self.down_blocks: UpperCAmelCase__ = down_block(__UpperCAmelCase ) # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase ) # post-process UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : List[Any] , __UpperCAmelCase : str=3 , __UpperCAmelCase : Union[str, Any]=3 , __UpperCAmelCase : Optional[int]=("UpDecoderBlock2D",) , __UpperCAmelCase : str=(6_4,) , __UpperCAmelCase : Optional[Any]=2 , __UpperCAmelCase : Tuple=3_2 , __UpperCAmelCase : Any="silu" , __UpperCAmelCase : Any="group" , ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = layers_per_block UpperCAmelCase__ = nn.Convad( __UpperCAmelCase , block_out_channels[-1] , kernel_size=3 , stride=1 , padding=1 , ) UpperCAmelCase__ = None UpperCAmelCase__ = nn.ModuleList([] ) UpperCAmelCase__ = in_channels if norm_type == "spatial" else None # mid UpperCAmelCase__ = UNetMidBlockaD( in_channels=block_out_channels[-1] , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , output_scale_factor=1 , resnet_time_scale_shift="default" if norm_type == "group" else norm_type , attention_head_dim=block_out_channels[-1] , resnet_groups=__UpperCAmelCase , temb_channels=__UpperCAmelCase , ) # up UpperCAmelCase__ = list(reversed(__UpperCAmelCase ) ) UpperCAmelCase__ = reversed_block_out_channels[0] for i, up_block_type in enumerate(__UpperCAmelCase ): UpperCAmelCase__ = output_channel UpperCAmelCase__ = reversed_block_out_channels[i] UpperCAmelCase__ = i == len(__UpperCAmelCase ) - 1 UpperCAmelCase__ = get_up_block( __UpperCAmelCase , num_layers=self.layers_per_block + 1 , in_channels=__UpperCAmelCase , out_channels=__UpperCAmelCase , prev_output_channel=__UpperCAmelCase , add_upsample=not is_final_block , resnet_eps=1E-6 , resnet_act_fn=__UpperCAmelCase , resnet_groups=__UpperCAmelCase , attention_head_dim=__UpperCAmelCase , temb_channels=__UpperCAmelCase , resnet_time_scale_shift=__UpperCAmelCase , ) self.up_blocks.append(__UpperCAmelCase ) UpperCAmelCase__ = output_channel # out if norm_type == "spatial": UpperCAmelCase__ = SpatialNorm(block_out_channels[0] , __UpperCAmelCase ) else: UpperCAmelCase__ = nn.GroupNorm(num_channels=block_out_channels[0] , num_groups=__UpperCAmelCase , eps=1E-6 ) UpperCAmelCase__ = nn.SiLU() UpperCAmelCase__ = nn.Convad(block_out_channels[0] , __UpperCAmelCase , 3 , padding=1 ) UpperCAmelCase__ = False def lowercase_ (self : Optional[int] , __UpperCAmelCase : Tuple , __UpperCAmelCase : Dict=None ) -> List[Any]: """simple docstring""" UpperCAmelCase__ = z UpperCAmelCase__ = self.conv_in(__UpperCAmelCase ) UpperCAmelCase__ = next(iter(self.up_blocks.parameters() ) ).dtype if self.training and self.gradient_checkpointing: def create_custom_forward(__UpperCAmelCase : str ): def custom_forward(*__UpperCAmelCase : List[str] ): return module(*__UpperCAmelCase ) return custom_forward if is_torch_version(">=" , "1.11.0" ): # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase , use_reentrant=__UpperCAmelCase ) else: # middle UpperCAmelCase__ = torch.utils.checkpoint.checkpoint( create_custom_forward(self.mid_block ) , __UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = torch.utils.checkpoint.checkpoint(create_custom_forward(__UpperCAmelCase ) , __UpperCAmelCase , __UpperCAmelCase ) else: # middle UpperCAmelCase__ = self.mid_block(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = sample.to(__UpperCAmelCase ) # up for up_block in self.up_blocks: UpperCAmelCase__ = up_block(__UpperCAmelCase , __UpperCAmelCase ) # post-process if latent_embeds is None: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase ) else: UpperCAmelCase__ = self.conv_norm_out(__UpperCAmelCase , __UpperCAmelCase ) UpperCAmelCase__ = self.conv_act(__UpperCAmelCase ) UpperCAmelCase__ = self.conv_out(__UpperCAmelCase ) return sample class A ( nn.Module ): def __init__(self : Optional[Any] , __UpperCAmelCase : str , __UpperCAmelCase : List[str] , __UpperCAmelCase : List[str] , __UpperCAmelCase : Dict=None , __UpperCAmelCase : Union[str, Any]="random" , __UpperCAmelCase : Dict=False , __UpperCAmelCase : Union[str, Any]=True ) -> Dict: """simple docstring""" super().__init__() UpperCAmelCase__ = n_e UpperCAmelCase__ = vq_embed_dim UpperCAmelCase__ = beta UpperCAmelCase__ = legacy UpperCAmelCase__ = nn.Embedding(self.n_e , self.vq_embed_dim ) self.embedding.weight.data.uniform_(-1.0 / self.n_e , 1.0 / self.n_e ) UpperCAmelCase__ = remap if self.remap is not None: self.register_buffer("used" , torch.tensor(np.load(self.remap ) ) ) UpperCAmelCase__ = self.used.shape[0] UpperCAmelCase__ = unknown_index # "random" or "extra" or integer if self.unknown_index == "extra": UpperCAmelCase__ = self.re_embed UpperCAmelCase__ = self.re_embed + 1 print( f"""Remapping {self.n_e} indices to {self.re_embed} indices. """ f"""Using {self.unknown_index} for unknown indices.""" ) else: UpperCAmelCase__ = n_e UpperCAmelCase__ = sane_index_shape def lowercase_ (self : str , __UpperCAmelCase : str ) -> List[str]: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) UpperCAmelCase__ = (inds[:, :, None] == used[None, None, ...]).long() UpperCAmelCase__ = match.argmax(-1 ) UpperCAmelCase__ = match.sum(2 ) < 1 if self.unknown_index == "random": UpperCAmelCase__ = torch.randint(0 , self.re_embed , size=new[unknown].shape ).to(device=new.device ) else: UpperCAmelCase__ = self.unknown_index return new.reshape(__UpperCAmelCase ) def lowercase_ (self : Tuple , __UpperCAmelCase : Optional[int] ) -> Dict: """simple docstring""" UpperCAmelCase__ = inds.shape assert len(__UpperCAmelCase ) > 1 UpperCAmelCase__ = inds.reshape(ishape[0] , -1 ) UpperCAmelCase__ = self.used.to(__UpperCAmelCase ) if self.re_embed > self.used.shape[0]: # extra token UpperCAmelCase__ = 0 # simply set to zero UpperCAmelCase__ = torch.gather(used[None, :][inds.shape[0] * [0], :] , 1 , __UpperCAmelCase ) return back.reshape(__UpperCAmelCase ) def lowercase_ (self : Optional[Any] , __UpperCAmelCase : Dict ) -> List[str]: """simple docstring""" UpperCAmelCase__ = z.permute(0 , 2 , 3 , 1 ).contiguous() UpperCAmelCase__ = z.view(-1 , self.vq_embed_dim ) # distances from z to embeddings e_j (z - e)^2 = z^2 + e^2 - 2 e * z UpperCAmelCase__ = torch.argmin(torch.cdist(__UpperCAmelCase , self.embedding.weight ) , dim=1 ) UpperCAmelCase__ = self.embedding(__UpperCAmelCase ).view(z.shape ) UpperCAmelCase__ = None UpperCAmelCase__ = None # compute loss for embedding if not self.legacy: UpperCAmelCase__ = self.beta * torch.mean((z_q.detach() - z) ** 2 ) + torch.mean((z_q - z.detach()) ** 2 ) else: UpperCAmelCase__ = torch.mean((z_q.detach() - z) ** 2 ) + self.beta * torch.mean((z_q - z.detach()) ** 2 ) # preserve gradients UpperCAmelCase__ = z + (z_q - z).detach() # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() if self.remap is not None: UpperCAmelCase__ = min_encoding_indices.reshape(z.shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.remap_to_used(__UpperCAmelCase ) UpperCAmelCase__ = min_encoding_indices.reshape(-1 , 1 ) # flatten if self.sane_index_shape: UpperCAmelCase__ = min_encoding_indices.reshape(z_q.shape[0] , z_q.shape[2] , z_q.shape[3] ) return z_q, loss, (perplexity, min_encodings, min_encoding_indices) def lowercase_ (self : Optional[int] , __UpperCAmelCase : int , __UpperCAmelCase : Optional[Any] ) -> Any: """simple docstring""" if self.remap is not None: UpperCAmelCase__ = indices.reshape(shape[0] , -1 ) # add batch axis UpperCAmelCase__ = self.unmap_to_all(__UpperCAmelCase ) UpperCAmelCase__ = indices.reshape(-1 ) # flatten again # get quantized latent vectors UpperCAmelCase__ = self.embedding(__UpperCAmelCase ) if shape is not None: UpperCAmelCase__ = z_q.view(__UpperCAmelCase ) # reshape back to match original input shape UpperCAmelCase__ = z_q.permute(0 , 3 , 1 , 2 ).contiguous() return z_q class A ( UpperCAmelCase_ ): def __init__(self : Any , __UpperCAmelCase : Dict , __UpperCAmelCase : str=False ) -> Tuple: """simple docstring""" UpperCAmelCase__ = parameters UpperCAmelCase__ , UpperCAmelCase__ = torch.chunk(__UpperCAmelCase , 2 , dim=1 ) UpperCAmelCase__ = torch.clamp(self.logvar , -30.0 , 20.0 ) UpperCAmelCase__ = deterministic UpperCAmelCase__ = torch.exp(0.5 * self.logvar ) UpperCAmelCase__ = torch.exp(self.logvar ) if self.deterministic: UpperCAmelCase__ = UpperCAmelCase__ = torch.zeros_like( self.mean , device=self.parameters.device , dtype=self.parameters.dtype ) def lowercase_ (self : Union[str, Any] , __UpperCAmelCase : Optional[torch.Generator] = None ) -> torch.FloatTensor: """simple docstring""" UpperCAmelCase__ = randn_tensor( self.mean.shape , generator=__UpperCAmelCase , device=self.parameters.device , dtype=self.parameters.dtype ) UpperCAmelCase__ = self.mean + self.std * sample return x def lowercase_ (self : str , __UpperCAmelCase : int=None ) -> Any: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) else: if other is None: return 0.5 * torch.sum(torch.pow(self.mean , 2 ) + self.var - 1.0 - self.logvar , dim=[1, 2, 3] ) else: return 0.5 * torch.sum( torch.pow(self.mean - other.mean , 2 ) / other.var + self.var / other.var - 1.0 - self.logvar + other.logvar , dim=[1, 2, 3] , ) def lowercase_ (self : Dict , __UpperCAmelCase : Tuple , __UpperCAmelCase : Any=[1, 2, 3] ) -> Dict: """simple docstring""" if self.deterministic: return torch.Tensor([0.0] ) UpperCAmelCase__ = np.log(2.0 * np.pi ) return 0.5 * torch.sum(logtwopi + self.logvar + torch.pow(sample - self.mean , 2 ) / self.var , dim=__UpperCAmelCase ) def lowercase_ (self : Tuple ) -> Optional[Any]: """simple docstring""" return self.mean
65
1
UpperCamelCase__ = 6_5_5_2_1 def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' UpperCAmelCase__ = 1 UpperCAmelCase__ = 0 for plain_chr in plain_text: UpperCAmelCase__ = (a + ord(__A )) % MOD_ADLER UpperCAmelCase__ = (b + a) % MOD_ADLER return (b << 16) | a
65
import asyncio import os import re import sys import tempfile import unittest from contextlib import contextmanager from copy import deepcopy from distutils.util import strtobool from enum import Enum from importlib.util import find_spec from pathlib import Path from unittest.mock import patch import pyarrow as pa import pytest import requests from packaging import version from datasets import config if config.PY_VERSION < version.parse('3.8'): import importlib_metadata else: import importlib.metadata as importlib_metadata def lowerCAmelCase_ ( __A, __A=False ) -> Any: '''simple docstring''' try: UpperCAmelCase__ = os.environ[key] except KeyError: # KEY isn't set, default to `default`. UpperCAmelCase__ = default else: # KEY is set, convert it to True or False. try: UpperCAmelCase__ = strtobool(__A ) except ValueError: # More values are supported, but let's keep the message simple. raise ValueError(f"""If set, {key} must be yes or no.""" ) return _value UpperCamelCase__ = parse_flag_from_env('RUN_SLOW', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_REMOTE', default=False) UpperCamelCase__ = parse_flag_from_env('RUN_LOCAL', default=True) UpperCamelCase__ = parse_flag_from_env('RUN_PACKAGED', default=True) # Compression UpperCamelCase__ = pytest.mark.skipif(not config.LZ4_AVAILABLE, reason='test requires lz4') UpperCamelCase__ = pytest.mark.skipif(not config.PY7ZR_AVAILABLE, reason='test requires py7zr') UpperCamelCase__ = pytest.mark.skipif(not config.ZSTANDARD_AVAILABLE, reason='test requires zstandard') # Audio UpperCamelCase__ = pytest.mark.skipif( # On Windows and OS X, soundfile installs sndfile find_spec('soundfile') is None or version.parse(importlib_metadata.version('soundfile')) < version.parse('0.12.0'), reason='test requires sndfile>=0.12.1: \'pip install \"soundfile>=0.12.1\"\'; ', ) # Beam UpperCamelCase__ = pytest.mark.skipif( not config.BEAM_AVAILABLE or config.DILL_VERSION >= version.parse('0.3.2'), reason='test requires apache-beam and a compatible dill version', ) # Dill-cloudpickle compatibility UpperCamelCase__ = pytest.mark.skipif( config.DILL_VERSION <= version.parse('0.3.2'), reason='test requires dill>0.3.2 for cloudpickle compatibility', ) # Windows UpperCamelCase__ = pytest.mark.skipif( sys.platform == 'win32', reason='test should not be run on Windows', ) def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' try: import faiss # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires faiss" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import regex # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires regex" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' try: import elasticsearch # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires elasticsearch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' try: import sqlalchemy # noqa except ImportError: UpperCAmelCase__ = unittest.skip("test requires sqlalchemy" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[str]: '''simple docstring''' if not config.TORCH_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires PyTorch" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Union[str, Any]: '''simple docstring''' if not config.TF_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires TensorFlow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not config.JAX_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires JAX" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> int: '''simple docstring''' if not config.PIL_AVAILABLE: UpperCAmelCase__ = unittest.skip("test requires Pillow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import transformers # noqa F401 except ImportError: return unittest.skip("test requires transformers" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Dict: '''simple docstring''' try: import tiktoken # noqa F401 except ImportError: return unittest.skip("test requires tiktoken" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import spacy # noqa F401 except ImportError: return unittest.skip("test requires spacy" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' def _require_spacy_model(__A ): try: import spacy # noqa F401 spacy.load(__A ) except ImportError: return unittest.skip("test requires spacy" )(__A ) except OSError: return unittest.skip("test requires spacy model '{}'".format(__A ) )(__A ) else: return test_case return _require_spacy_model def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' try: import pyspark # noqa F401 except ImportError: return unittest.skip("test requires pyspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Tuple: '''simple docstring''' try: import joblibspark # noqa F401 except ImportError: return unittest.skip("test requires joblibspark" )(__A ) else: return test_case def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' if not _run_slow_tests or _run_slow_tests == 0: UpperCAmelCase__ = unittest.skip("test is slow" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> List[Any]: '''simple docstring''' if not _run_local_tests or _run_local_tests == 0: UpperCAmelCase__ = unittest.skip("test is local" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Optional[Any]: '''simple docstring''' if not _run_packaged_tests or _run_packaged_tests == 0: UpperCAmelCase__ = unittest.skip("test is packaged" )(__A ) return test_case def lowerCAmelCase_ ( __A ) -> Any: '''simple docstring''' if not _run_remote_tests or _run_remote_tests == 0: UpperCAmelCase__ = unittest.skip("test requires remote" )(__A ) return test_case def lowerCAmelCase_ ( *__A ) -> Optional[int]: '''simple docstring''' def decorate(cls ): for name, fn in cls.__dict__.items(): if callable(__A ) and name.startswith("test" ): for decorator in decorators: UpperCAmelCase__ = decorator(__A ) setattr(cls, __A, __A ) return cls return decorate class A ( UpperCAmelCase_ ): pass class A ( UpperCAmelCase_ ): __UpperCAmelCase : Union[str, Any] = 0 __UpperCAmelCase : str = 1 __UpperCAmelCase : int = 2 @contextmanager def lowerCAmelCase_ ( __A=OfflineSimulationMode.CONNECTION_FAILS, __A=1e-16 ) -> List[str]: '''simple docstring''' UpperCAmelCase__ = requests.Session().request def timeout_request(__A, __A, __A, **__A ): # Change the url to an invalid url so that the connection hangs UpperCAmelCase__ = "https://10.255.255.1" if kwargs.get("timeout" ) is None: raise RequestWouldHangIndefinitelyError( f"""Tried a call to {url} in offline mode with no timeout set. Please set a timeout.""" ) UpperCAmelCase__ = timeout try: return online_request(__A, __A, **__A ) except Exception as e: # The following changes in the error are just here to make the offline timeout error prettier UpperCAmelCase__ = url UpperCAmelCase__ = e.args[0] UpperCAmelCase__ = (max_retry_error.args[0].replace("10.255.255.1", f"""OfflineMock[{url}]""" ),) UpperCAmelCase__ = (max_retry_error,) raise def raise_connection_error(__A, __A, **__A ): raise requests.ConnectionError("Offline mode is enabled.", request=__A ) if mode is OfflineSimulationMode.CONNECTION_FAILS: with patch("requests.Session.send", __A ): yield elif mode is OfflineSimulationMode.CONNECTION_TIMES_OUT: # inspired from https://stackoverflow.com/a/904609 with patch("requests.Session.request", __A ): yield elif mode is OfflineSimulationMode.HF_DATASETS_OFFLINE_SET_TO_1: with patch("datasets.config.HF_DATASETS_OFFLINE", __A ): yield else: raise ValueError("Please use a value from the OfflineSimulationMode enum." ) @contextmanager def lowerCAmelCase_ ( *__A, **__A ) -> str: '''simple docstring''' UpperCAmelCase__ = str(Path().resolve() ) with tempfile.TemporaryDirectory(*__A, **__A ) as tmp_dir: try: os.chdir(__A ) yield finally: os.chdir(__A ) @contextmanager def lowerCAmelCase_ ( ) -> Optional[Any]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory > 0, "Arrow memory didn't increase." @contextmanager def lowerCAmelCase_ ( ) -> List[str]: '''simple docstring''' import gc gc.collect() UpperCAmelCase__ = pa.total_allocated_bytes() yield assert pa.total_allocated_bytes() - previous_allocated_memory <= 0, "Arrow memory wasn't expected to increase." def lowerCAmelCase_ ( __A, __A ) -> List[str]: '''simple docstring''' return deepcopy(__A ).integers(0, 100, 10 ).tolist() == deepcopy(__A ).integers(0, 100, 10 ).tolist() def lowerCAmelCase_ ( __A ) -> Optional[int]: '''simple docstring''' import decorator from requests.exceptions import HTTPError def _wrapper(__A, *__A, **__A ): try: return func(*__A, **__A ) except HTTPError as err: if str(__A ).startswith("500" ) or str(__A ).startswith("502" ): pytest.xfail(str(__A ) ) raise err return decorator.decorator(_wrapper, __A ) class A : def __init__(self : Optional[Any] , __UpperCAmelCase : int , __UpperCAmelCase : int , __UpperCAmelCase : List[str] ) -> Union[str, Any]: """simple docstring""" UpperCAmelCase__ = returncode UpperCAmelCase__ = stdout UpperCAmelCase__ = stderr async def lowerCAmelCase_ ( __A, __A ) -> Optional[int]: '''simple docstring''' while True: UpperCAmelCase__ = await stream.readline() if line: callback(__A ) else: break async def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=None, __A=False, __A=False ) -> _RunOutput: '''simple docstring''' if echo: print("\nRunning: ", " ".join(__A ) ) UpperCAmelCase__ = await asyncio.create_subprocess_exec( cmd[0], *cmd[1:], stdin=__A, stdout=asyncio.subprocess.PIPE, stderr=asyncio.subprocess.PIPE, env=__A, ) # note: there is a warning for a possible deadlock when using `wait` with huge amounts of data in the pipe # https://docs.python.org/3/library/asyncio-subprocess.html#asyncio.asyncio.subprocess.Process.wait # # If it starts hanging, will need to switch to the following code. The problem is that no data # will be seen until it's done and if it hangs for example there will be no debug info. # out, err = await p.communicate() # return _RunOutput(p.returncode, out, err) UpperCAmelCase__ = [] UpperCAmelCase__ = [] def tee(__A, __A, __A, __A="" ): UpperCAmelCase__ = line.decode("utf-8" ).rstrip() sink.append(__A ) if not quiet: print(__A, __A, file=__A ) # XXX: the timeout doesn't seem to make any difference here await asyncio.wait( [ _read_stream(p.stdout, lambda __A : tee(__A, __A, sys.stdout, label="stdout:" ) ), _read_stream(p.stderr, lambda __A : tee(__A, __A, sys.stderr, label="stderr:" ) ), ], timeout=__A, ) return _RunOutput(await p.wait(), __A, __A ) def lowerCAmelCase_ ( __A, __A=None, __A=None, __A=180, __A=False, __A=True ) -> _RunOutput: '''simple docstring''' UpperCAmelCase__ = asyncio.get_event_loop() UpperCAmelCase__ = loop.run_until_complete( _stream_subprocess(__A, env=__A, stdin=__A, timeout=__A, quiet=__A, echo=__A ) ) UpperCAmelCase__ = " ".join(__A ) if result.returncode > 0: UpperCAmelCase__ = "\n".join(result.stderr ) raise RuntimeError( f"""'{cmd_str}' failed with returncode {result.returncode}\n\n""" f"""The combined stderr from workers follows:\n{stderr}""" ) # check that the subprocess actually did run and produced some output, should the test rely on # the remote side to do the testing if not result.stdout and not result.stderr: raise RuntimeError(f"""'{cmd_str}' produced no output.""" ) return result def lowerCAmelCase_ ( ) -> Tuple: '''simple docstring''' UpperCAmelCase__ = os.environ.get("PYTEST_XDIST_WORKER", "gw0" ) UpperCAmelCase__ = re.sub(r"^gw", "", __A, 0, re.M ) return int(__A ) def lowerCAmelCase_ ( ) -> List[Any]: '''simple docstring''' UpperCAmelCase__ = 29_500 UpperCAmelCase__ = pytest_xdist_worker_id() return port + uniq_delta
65
1
from __future__ import annotations import typing from collections import Counter def lowerCAmelCase_ ( __A ) -> typing.Counter[int]: '''simple docstring''' UpperCAmelCase__ = Counter() for base in range(1, max_perimeter + 1 ): for perpendicular in range(__A, max_perimeter + 1 ): UpperCAmelCase__ = (base * base + perpendicular * perpendicular) ** 0.5 if hypotenuse == int(__A ): UpperCAmelCase__ = int(base + perpendicular + hypotenuse ) if perimeter > max_perimeter: continue triplets[perimeter] += 1 return triplets def lowerCAmelCase_ ( __A = 1_000 ) -> int: '''simple docstring''' UpperCAmelCase__ = pythagorean_triple(__A ) return triplets.most_common(1 )[0][0] if __name__ == "__main__": print(f'''Perimeter {solution()} has maximum solutions''')
65
def lowerCAmelCase_ ( __A, __A ) -> float: '''simple docstring''' def get_matched_characters(__A, __A ) -> str: UpperCAmelCase__ = [] UpperCAmelCase__ = min(len(_stra ), len(_stra ) ) // 2 for i, l in enumerate(_stra ): UpperCAmelCase__ = int(max(0, i - limit ) ) UpperCAmelCase__ = int(min(i + limit + 1, len(_stra ) ) ) if l in _stra[left:right]: matched.append(__A ) UpperCAmelCase__ = f"""{_stra[0:_stra.index(__A )]} {_stra[_stra.index(__A ) + 1:]}""" return "".join(__A ) # matching characters UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = get_matched_characters(__A, __A ) UpperCAmelCase__ = len(__A ) # transposition UpperCAmelCase__ = ( len([(ca, ca) for ca, ca in zip(__A, __A ) if ca != ca] ) // 2 ) if not match_count: UpperCAmelCase__ = 0.0 else: UpperCAmelCase__ = ( 1 / 3 * ( match_count / len(__A ) + match_count / len(__A ) + (match_count - transpositions) / match_count ) ) # common prefix up to 4 characters UpperCAmelCase__ = 0 for ca, ca in zip(stra[:4], stra[:4] ): if ca == ca: prefix_len += 1 else: break return jaro + 0.1 * prefix_len * (1 - jaro) if __name__ == "__main__": import doctest doctest.testmod() print(jaro_winkler('hello', 'world'))
65
1